Download presentation
Presentation is loading. Please wait.
1
Unsupervised Models for Named Entity Classifcation Michael Collins Yoram Singer AT&T Labs, 1999
2
The Task Tag phrases with “person”, “organization” or “location”. For example, R alph Grishman, of NYU, sure is swell.
3
WHY? Unlabeled data Labeled data
4
Spelling Rules The approach uses two kinds of rules –Spelling Simple look up to see “Honduras” is a location! Look for words in string, like “Mr.”
5
Contextual Rules –Contextual Words surrounding the string –A rule that any proper name modified by an appositive whose head is “president” is a person.
6
Two Categories of Rules The key to the method is redundancy in the two kind of rules. …says Mr. Cooper, a vice president of… spelling contextual Unlabeled data gives us these hints! contextualspelling
7
The Experiment 970,000 New York Times sentences were parsed. Sequences of NNP and NNPS were then extracted as named entity examples if they met one of two critereon.
8
Kinds of Noun Phrases 1.There was an appositive modifier to the NP, whose head is a singular noun (tagged NN). –…says Maury Cooper, a vice president… 2. The NP is a compliment to a preposition which is the head of a PP. This PP modifies another NP whose head is a singular noun. –… fraud related to work on a federally funded sewage plant in Georgia.
9
(spelling, context) pairs created …says Maury Cooper, a vice president… –(Maury Cooper, president) … fraud related to work on a federally funded sewage plant in Georgia. –(Georgia, plant_in)
10
Rules Set of rules –Full-string=x (full-string=Maury Cooper) –Contains(x) (contains(Maury)) –Allcap1 IBM –Allcap2 N.Y. –Nonalpha=x A.T.&T. (nonalpha=..&.) –Context = x (context = president) –Context-type = xappos or prep
11
SEED RULES Full-string = New York Full-string = California Full-string = U.S. Contains(Mr.) Contains(Incorporated) Full-string=Microsoft Full-string=I.B.M.
12
The Algorithm 1.Initialize: Set the spelling decision list equal to the set of seed rules. 2.Label the training set using these rules. 3.Use these to get contextual rules. (x = feature, y = label) 4.Label set using contextual rules, and use to get sp. rules. 5.Set spelling rules to seed plus the new rules. 6.If less than threshold new rules, go to 2 and add 15 more. 7.When finished, label the training data with the combined spelling/contextual decision list, then induce a final decision list from the labeled examples where all rules are added to the decision list.
13
Example (IBM, company) –…IBM, the company that makes… (General Electric, company) –..General Electric, a leading company in the area,… (General Electric, employer ) –… joined General Electric, the biggest employer… (NYU, employer) –NYU, the employer of the famous Ralph Grishman,…
14
The Power I.B.M. Mr. Two classifiers both give labels on 49.2% of unlabeled examples Agree on 99.25% of them!
15
Evaluation 88,962 (spelling, context) pairs. –971,746 sentences 1,000 randomly extracted to be test set. Location, person, organization, noise 186, 289, 402, 123 Took out 38 temporal noise. Clean Accuracy: Nc/ 962 Noise Accuracy: Nc/(962-85)
16
Results AlgorithmClean AccuracyNoise Accuracy Baseline45.8%41.8% EM83.1%75.8% Yarowsky 9581.3%74.1% Yarowsky Cautious91.2%83.2% DL-CoTrain91.3%83.3% CoBoost91.1%83.1%
17
QUESTIONS
18
Thank you! www.lightrail.com/ www.cnnfn.com/ pbskids.org/ www.szilagyi.us www.dflt.org
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.