1 Fine-grained and Coarse-grained Word Sense Disambiguation Jinying Chen, Hoa Trang Dang, Martha Palmer August 22, 2003.

Slides:



Advertisements
Similar presentations
Multilinugual PennTools that capture parses and predicate-argument structures, and their use in Applications Martha Palmer, Aravind Joshi, Mitch Marcus,
Advertisements

CL Research ACL Pattern Dictionary of English Prepositions (PDEP) Ken Litkowski CL Research 9208 Gue Road Damascus,
Word Sense Disambiguation for Machine Translation Han-Bin Chen
Towards Parsing Unrestricted Text into PropBank Predicate- Argument Structures ACL4 Project NCLT Seminar Presentation, 7th June 2006 Conor Cafferkey.
Automatic Metaphor Interpretation as a Paraphrasing Task Ekaterina Shutova Computer Lab, University of Cambridge NAACL 2010.
Simple Features for Chinese Word Sense Disambiguation Hoa Trang Dang, Ching-yi Chia, Martha Palmer, Fu- Dong Chiou Computer and Information Science University.
CS Word Sense Disambiguation. 2 Overview A problem for semantic attachment approaches: what happens when a given lexeme has multiple ‘meanings’?
CS 4705 Lecture 19 Word Sense Disambiguation. Overview Selectional restriction based approaches Robust techniques –Machine Learning Supervised Unsupervised.
1 Noun Homograph Disambiguation Using Local Context in Large Text Corpora Marti A. Hearst Presented by: Heng Ji Mar. 29, 2004.
Taking the Kitchen Sink Seriously: An Ensemble Approach to Word Sense Disambiguation from Christopher Manning et al.
Automatic Verb Sense Grouping --- Term Project Proposal for CIS630 Jinying Chen 10/28/2002.
Word Sense Disambiguation CMSC 723: Computational Linguistics I ― Session #11 Jimmy Lin The iSchool University of Maryland Wednesday, November 11, 2009.
CS 4705 Word Sense Disambiguation. Overview Selectional restriction based approaches Robust techniques –Machine Learning Supervised Unsupervised –Dictionary-based.
تمرين شماره 1 درس NLP سيلابس درس NLP در دانشگاه هاي ديگر ___________________________ راحله مکي استاد درس: دکتر عبدالله زاده پاييز 85.
Machine Learning in Natural Language Processing Noriko Tomuro November 16, 2006.
Distributional clustering of English words Authors: Fernando Pereira, Naftali Tishby, Lillian Lee Presenter: Marian Olteanu.
Word sense induction using continuous vector space models
Semantic Role Labeling using Maximum Entropy Model Joon-Ho Lim NLP Lab. Korea Univ.
Word Sense Disambiguation for Automatic Taxonomy Construction from Text-Based Web Corpora 12th International Conference on Web Information System Engineering.
ELN – Natural Language Processing Giuseppe Attardi
Evaluating the Contribution of EuroWordNet and Word Sense Disambiguation to Cross-Language Information Retrieval Paul Clough 1 and Mark Stevenson 2 Department.
Lemmatization Tagging LELA /20 Lemmatization Basic form of annotation involving identification of underlying lemmas (lexemes) of the words in.
Ronan Collobert Jason Weston Leon Bottou Michael Karlen Koray Kavukcouglu Pavel Kuksa.
Empirical Methods in Information Extraction Claire Cardie Appeared in AI Magazine, 18:4, Summarized by Seong-Bae Park.
PropBank, VerbNet & SemLink Edward Loper. PropBank 1M words of WSJ annotated with predicate- argument structures for verbs. –The location & type of each.
“How much context do you need?” An experiment about context size in Interactive Cross-language Question Answering B. Navarro, L. Moreno-Monteagudo, E.
Jiuling Zhang  Why perform query expansion?  WordNet based Word Sense Disambiguation WordNet Word Sense Disambiguation  Conceptual Query.
Distributional Part-of-Speech Tagging Hinrich Schütze CSLI, Ventura Hall Stanford, CA , USA NLP Applications.
Word Sense Disambiguation UIUC - 06/10/2004 Word Sense Disambiguation Another NLP working problem for learning with constraints… Lluís Màrquez TALP, LSI,
Annotating Words using WordNet Semantic Glosses Julian Szymański Department of Computer Systems Architecture, Faculty of Electronics, Telecommunications.
1 Statistical NLP: Lecture 9 Word Sense Disambiguation.
Paper Review by Utsav Sinha August, 2015 Part of assignment in CS 671: Natural Language Processing, IIT Kanpur.
W ORD S ENSE D ISAMBIGUATION By Mahmood Soltani Tehran University 2009/12/24 1.
Improving Subcategorization Acquisition using Word Sense Disambiguation Anna Korhonen and Judith Preiss University of Cambridge, Computer Laboratory 15.
SYMPOSIUM ON SEMANTICS IN SYSTEMS FOR TEXT PROCESSING September 22-24, Venice, Italy Combining Knowledge-based Methods and Supervised Learning for.
A Language Independent Method for Question Classification COLING 2004.
2014 EMNLP Xinxiong Chen, Zhiyuan Liu, Maosong Sun State Key Laboratory of Intelligent Technology and Systems Tsinghua National Laboratory for Information.
AQUAINT Workshop – June 2003 Improved Semantic Role Parsing Kadri Hacioglu, Sameer Pradhan, Valerie Krugler, Steven Bethard, Ashley Thornton, Wayne Ward,
Page 1 SenDiS Sectoral Operational Programme "Increase of Economic Competitiveness" "Investments for your future" Project co-financed by the European Regional.
CS 4705 Lecture 19 Word Sense Disambiguation. Overview Selectional restriction based approaches Robust techniques –Machine Learning Supervised Unsupervised.
Semantic Role Labeling: English PropBank
Modelling Human Thematic Fit Judgments IGK Colloquium 3/2/2005 Ulrike Padó.
Unsupervised Word Sense Disambiguation REU, Summer, 2009.
Opinion Holders in Opinion Text from Online Newspapers Youngho Kim, Yuchul Jung and Sung-Hyon Myaeng Reporter: Chia-Ying Lee Advisor: Prof. Hsin-Hsi Chen.
Combining Lexical Resources: Mapping Between PropBank and VerbNet Edward Loper,Szu-ting Yi, Martha Palmer September 2006.
NUMBER OR NUANCE: Factors Affecting Reliable Word Sense Annotation Susan Windisch Brown, Travis Rood, and Martha Palmer University of Colorado at Boulder.
LING 6520: Comparative Topics in Linguistics (from a computational perspective) Martha Palmer Jan 15,
CSE391 – 2005 NLP 1 Events From KRR lecture. CSE391 – 2005 NLP 2 Ask Jeeves – A Q/A, IR ex. What do you call a successful movie? Tips on Being a Successful.
4. Relationship Extraction Part 4 of Information Extraction Sunita Sarawagi 9/7/2012CS 652, Peter Lindes1.
CIS630 1 Penn Different Sense Granularities Martha Palmer, Olga Babko-Malaya September 20, 2004.
11 Project, Part 3. Outline Basics of supervised learning using Naïve Bayes (using a simpler example) Features for the project 2.
ARDA Visit 1 Penn Lexical Semantics at Penn: Proposition Bank and VerbNet Martha Palmer, Dan Gildea, Paul Kingsbury, Olga Babko-Malaya, Bert Xue, Karin.
1 Gloss-based Semantic Similarity Metrics for Predominant Sense Acquisition Ryu Iida Nara Institute of Science and Technology Diana McCarthy and Rob Koeling.
Using Wikipedia for Hierarchical Finer Categorization of Named Entities Aasish Pappu Language Technologies Institute Carnegie Mellon University PACLIC.
SALSA-WS 09/05 Approximating Textual Entailment with LFG and FrameNet Frames Aljoscha Burchardt, Anette Frank Computational Linguistics Department Saarland.
Multilinugual PennTools that capture parses and predicate-argument structures, for use in Applications Martha Palmer, Aravind Joshi, Mitch Marcus, Mark.
Learning Event Durations from Event Descriptions Feng Pan, Rutu Mulkar, Jerry R. Hobbs University of Southern California ACL ’ 06.
Overview of Statistical NLP IR Group Meeting March 7, 2006.
Second Language Learning From News Websites Word Sense Disambiguation using Word Embeddings.
Word Sense and Subjectivity (Coling/ACL 2006) Janyce Wiebe Rada Mihalcea University of Pittsburgh University of North Texas Acknowledgements: This slide.
Integrating linguistic knowledge in passage retrieval for question answering J¨org Tiedemann Alfa Informatica, University of Groningen HLT/EMNLP 2005.
English Proposition Bank: Status Report
Coarse-grained Word Sense Disambiguation
INAGO Project Automatic Knowledge Base Generation from Text for Interactive Question Answering.
Donna M. Gates Carnegie Mellon University
A method for WSD on Unrestricted Text
Automatic Detection of Causal Relations for Question Answering
CS224N Section 3: Corpora, etc.
CS224N Section 3: Project,Corpora
Progress report on Semantic Role Labeling
Presentation transcript:

1 Fine-grained and Coarse-grained Word Sense Disambiguation Jinying Chen, Hoa Trang Dang, Martha Palmer August 22, 2003

2 Outline Maxent Word Sense Disambiguator Coarse-grained WSD by Decision Tree Future Work

3 Maxent Word Sense Disambiguator (Martha, Hoa, Christiane, 2002 ) A deterministic model producing probability f j (sense,context):binary features : weight of feature j Z(context) : normalizing factor Can combine evidence from different knowledge source Feature weights determined automatically (GIS)

4 Features used in Maxent Model for English WSD Local Contextual Predicates Collocational features, e.g., target verb w, pos of w; pos of words at position –1,+1, w.r.t. w; words at positions –2, - 1, +1, +2, w.r.t. w Systactic features, e.g., active vs. passive, is there a sentential complement, subj, obj or indirect obj etc. Semantic features, e.g., Named Entity tag (PER, ORG, LOC) for proper nouns, and WN synsets and hypernyms for all nouns in above syntactic relation to w Topical Contextual Keywords Select words k with lowest entropy (P(sense|k)), i.e., being most informative, from anywhere in context Maxent Word Sense Disambiguator

5 Fine-grained and coarse-grained WSD Part of the results from (Martha, Hoa, Christiane, 2002 ) Maxent Word Sense Disambiguator VerbWN (corp) Sen WN (corp) Grp MX-fineMX-group call28(14)11(7) develop21(16)9(6) live7(6)4(4) pull18(10)10(5) strike20(16)12(10) Table 1 The Performance of Maxent Word Sense Disambiguator on five verbs

6 Coarse-grained WSD by Decision Tree A simpler model compared with Maxent Model Using Semantic Features from PropBank PropBank Each verb is defined by several framesets All verb instances belonging to the same frameset share a common set of roles Roles can be argn (n=0,1,…) and argM-f Frameset is consistent with Verb Sense Group Frameset tags and roles are semantic features for VSG

7 Building Decision Tree Use c5.0 of DT 3 Feature Sets: SF (Simple Feature set) works best: VOICE: PAS, ACT FRAMESET: 01,02, … ARGn (n=0,1,2 …) : 0(not occur), 1(occur) CoreFrame: 01-ARG0-ARG1, 02-ARG0-ARG2,… ARGM: 0(has not ARGM), 1(has ARGM) ARGM-f(f=DIS, ADV, …): i (occur i times) Automatic Verb Sense GroupingCoarse-grained WSD by Decision Tree

8 Table 2 Error rate of Decision Tree on five verbs Experimental Results Coarse-grained WSD by Decision Tree

9 Discussion Simple feature set and simple DT algorithms works well Potential sparse data problem Complicate DT algorithms (e.g., with boosting) tend to overfit the data Complex features are not utilized by the model Solution: use large corpus, e.g., parsed BNC corpus without frameset annotation Coarse-grained WSD by Decision Tree

10 Future Work Train DT or other models for coarse-grained WSD on large corpus without frameset annotation Unsupervised Frameset Tagging by EM-clustering Clustering nouns automatically instead of using WordNet to group nouns