Download presentation
Presentation is loading. Please wait.
Published byRodney Hood Modified over 9 years ago
1
Conditional Markov Models: MaxEnt Tagging and MEMMs William W. Cohen CALD
2
Announcements Confused about what to write up? –Mon 2/9: Ratnaparki & Frietag et al –Wed 2/11: Borthwick et al & Mikheev –Mon 2/16: no class (President’s day) –Wed 2/18: Sha & Pereira, Lafferty et al –Mon 2/23: Klein & Manning, Toutanova et al –Wed 2/25: no writeup due –Mon 3/1: no writeup due –Wed 3/3: project proposal due: personnel + 1-2 page –Spring break week, no class
3
Review of review Multinomial HMMs are sequential version of naïve Bayes. One way to drop independence assumption: use a maxent instead of NB, and a conditional model
4
From NB to Maxent
6
Learning: set alpha parameters to maximize this: the ML model of the data, given we’re using the same functional form as NB. Turns out this is the same as maximizing entropy of p(y|x) over all distributions.
7
MaxEnt Comments Functional form same as Naïve Bayes (loglinear model) Numerical issues & smoothing important All methods are iterative Classification performance can be competitive with state-of-art optimizes Pr(y|x), not error rate
8
What is a symbol? Ideally we would like to use many, arbitrary, overlapping features of words. S t-1 S t O t S t+1 O t +1 O t - 1 identity of word ends in “-ski” is capitalized is part of a noun phrase is in a list of city names is under node X in WordNet is in bold font is indented is in hyperlink anchor … … … part of noun phrase is “Wisniewski” ends in “-ski”
9
What is a symbol? S t-1 S t O t S t+1 O t +1 O t - 1 identity of word ends in “-ski” is capitalized is part of a noun phrase is in a list of city names is under node X in WordNet is in bold font is indented is in hyperlink anchor … … … part of noun phrase is “Wisniewski” ends in “-ski” Idea: replace generative model in HMM with a maxent model, where state depends on observations
10
What is a symbol? S t-1 S t O t S t+1 O t +1 O t - 1 identity of word ends in “-ski” is capitalized is part of a noun phrase is in a list of city names is under node X in WordNet is in bold font is indented is in hyperlink anchor … … … part of noun phrase is “Wisniewski” ends in “-ski” Idea: replace generative model in HMM with a maxent model, where state depends on observations and previous state
11
What is a symbol? S t-1 S t O t S t+1 O t +1 O t - 1 identity of word ends in “-ski” is capitalized is part of a noun phrase is in a list of city names is under node X in WordNet is in bold font is indented is in hyperlink anchor … … … part of noun phrase is “Wisniewski” ends in “-ski” Idea: replace generative model in HMM with a maxent model, where state depends on observations and previous state history
12
Ratnaparkhi’s MXPOST Sequential learning problem: predict POS tags of words. Uses MaxEnt model described above. Rich feature set. To smooth, discard features occurring < 10 times.
13
MXPOST
14
MXPOST: learning & inference GIS Feature selection
15
MXPost inference Adwait: consider only extensions suggested by a dictionary
16
MXPost results State of art accuracy (for 1996) Same approach used successfully for several other sequential classification steps of a stochastic parser (also state of art). Same approach used for NER by Borthwick, Malouf, Collins, Manning, and others.
17
Alternative inference
18
Finding the most probable path: the Viterbi algorithm (for HMMs) define to be the probability of the most probable path accounting for the first i characters of x and ending in state k (ending in with tag k) we want to compute, the probability of the most probable path accounting for all of the sequence and ending in the end state can define recursively can use dynamic programming to find efficiently
19
Finding the most probable path: the Viterbi algorithm for HMMs initialization:
20
The Viterbi algorithm for HMMs recursion for emitting states (i =1…L):
21
The Viterbi algorithm for HMMs and Maxent Taggers recursion for emitting states (i =1…L): Previous tag k i-th token
22
MEMMs (Frietag & McCallum) Basic difference from ME tagging: –ME tagging: previous state is feature of MaxEnt classifier –MEMM: build a separate MaxEnt classifier for each state. Can build any HMM architecture you want; eg parallel nested HMM’s, etc. Data is fragmented: examples where previous tag is “proper noun” give no information about learning tags when previous tag is “noun” –Mostly a difference in viewpoint – easier to see parallels to HMMs
23
MEMM task: FAQ parsing
24
MEMM features
25
MEMMs
26
Borthwick et al: MENE system Much like MXPost, with some tricks for NER: –4 tags/field: x_start, x_continue, x_end, x_unique –Features: Section features Tokens in window Lexical features of tokens in window Dictionary features of tokens (is token a firstName?) External system of tokens (is this a NetOwl_company_start? proteus_person_unique?) Smooth by discarding low-count features –No history: viterbi search used to find best consistent tag sequence (e.g. no continue w/o start)
27
Dictionaries in MENE
28
MENE results (dry run)
29
MENE learning curves 96.3 93.3 92.2
30
Largest U.S. Cable Operator Makes Bid for Walt Disney By ANDREW ROSS SORKIN The Comcast Corporation, the largest cable television operator in the United States, made a $54.1 billion unsolicited takeover bid today for The Walt Disney Company, the storied family entertainment colossus. If successful, Comcast's audacious bid would once again reshape the entertainment landscape, creating a new media behemoth that would combine the power of Comcast's powerful distribution channels to some 21 million subscribers in the nation with Disney's vast library of content and production assets. Those include its ABC television network, ESPN and other cable networks, and the Disney and Miramax movie studios. Short names Longer names
31
LTG system Another MUC-7 competitor Handcoded rules for “easy” cases (amounts, etc) Process of repeated tagging and “matching” for hard cases –Sure-fire (high precision) rules for names where type is clear (“Phillip Morris, Inc – The Walt Disney Company”) –Partial matches to sure-fire rule are filtered maxent classifier (candidate filtering) using contextual information, etc –Higher-recall rules, avoiding conflicts with partial-match output “Phillip Morris announced today…. - “Disney’s ….” –Final partial-match & filter step on titles with different learned filter. Exploits discourse/context information
32
LTG Results
33
LTG Identifinder MENE+Proteus Manitoba (NB filtered names) NetOwl Commercial RBS
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.