NUMBER OR NUANCE: Factors Affecting Reliable Word Sense Annotation Susan Windisch Brown, Travis Rood, and Martha Palmer University of Colorado at Boulder.

Slides:



Advertisements
Similar presentations
What is Science? or True False
Advertisements

A Robust Approach to Aligning Heterogeneous Lexical Resources Mohammad Taher Pilehvar Roberto Navigli MultiJEDI ERC
Word Sense Disambiguation MAS.S60 Catherine Havasi Rob Speer.
CL Research ACL Pattern Dictionary of English Prepositions (PDEP) Ken Litkowski CL Research 9208 Gue Road Damascus,
Omega Ontology: Supporting Annotation Eduard Hovy with Andrew Philpot, Jerry Hobbs, Michael Fleischman, and Patrick Pantel USC/ISI.
Word Sense Disambiguation for Machine Translation Han-Bin Chen
Steven Schoonover.  What is VerbNet?  Levin Classification  In-depth look at VerbNet  Evolution of VerbNet  What is FrameNet?  Applications.
Mining Wiki Resources for Multilingual Named Entity Recognition Alexander E. Richman & Patrick Schone Reporter: Chia-Ying Lee Advisor: Prof. Hsin-Hsi Chen.
Cognitive Load Theory Sweller, van Merrienboer, and Paas, 1998 Psych 605 Fall 2009.
1 NSF-ULA Sense tagging and Eventive Nouns Martha Palmer, Miriam Eckert, Jena D. Hwang, Susan Windisch Brown, Dmitriy Dligach, Jinho Choi, Nianwen Xue.
CS Word Sense Disambiguation. 2 Overview A problem for semantic attachment approaches: what happens when a given lexeme has multiple ‘meanings’?
OntoNotes/PropBank Participants: BBN, Penn, Colorado, USC/ISI.
1 Noun Homograph Disambiguation Using Local Context in Large Text Corpora Marti A. Hearst Presented by: Heng Ji Mar. 29, 2004.
OntoNotes project Treebank Syntax Training Data Decoders Propositions Verb Senses and verbal ontology links Noun Senses and targeted nominalizations Coreference.
Chapter 2 – Tools of Positive Analysis
1 Complementarity of Lexical and Simple Syntactic Features: The SyntaLex Approach to S ENSEVAL -3 Saif Mohammad Ted Pedersen University of Toronto, Toronto.
What is Science? or 1.Science is concerned with understanding how nature and the physical world work. 2.Science can prove anything, solve any problem,
ELN – Natural Language Processing Giuseppe Attardi
Evaluating the Contribution of EuroWordNet and Word Sense Disambiguation to Cross-Language Information Retrieval Paul Clough 1 and Mark Stevenson 2 Department.
Printing: This poster is 48” wide by 36” high. It’s designed to be printed on a large-format printer. Customizing the Content: The placeholders in this.
Lemmatization Tagging LELA /20 Lemmatization Basic form of annotation involving identification of underlying lemmas (lexemes) of the words in.
CLEF Ǻrhus Robust – Word Sense Disambiguation exercise UBC: Eneko Agirre, Oier Lopez de Lacalle, Arantxa Otegi, German Rigau UVA & Irion: Piek Vossen.
The CoNLL-2013 Shared Task on Grammatical Error Correction Hwee Tou Ng, Yuanbin Wu, and Christian Hadiwinoto 1 Siew.
Timothy Reeves: Presenter Marisa Orr, Sherrill Biggers Evaluation of the Holistic Method to Size a 3-D Wheel/Soil Model.
Jennie Ning Zheng Linda Melchor Ferhat Omur. Contents Introduction WordNet Application – WordNet Data Structure - WordNet FrameNet Application – FrameNet.
Word Sense Disambiguation UIUC - 06/10/2004 Word Sense Disambiguation Another NLP working problem for learning with constraints… Lluís Màrquez TALP, LSI,
1 Statistical NLP: Lecture 9 Word Sense Disambiguation.
Paper Review by Utsav Sinha August, 2015 Part of assignment in CS 671: Natural Language Processing, IIT Kanpur.
Word Sense Disambiguation Reading: Chap 16-17, Jurafsky & Martin Instructor: Rada Mihalcea.
A resource and tool for Super Sense Tagging of Italian Texts LREC 2010, Malta – 19-21/05/2010 Giuseppe Attardi* Alessandro Lenci* + Stefano Dei Rossi*
Combining terminology resources and statistical methods for entity recognition: an evaluation Angus Roberts, Robert Gaizauskas, Mark Hepple, Yikun Guo.
Grobman, K. H. "Confirmation Bias." Teaching about. Developmentalpsychology.org, Web. 16 Sept Sequence Fits the instructor's Rule? Guess.
Copyright © 2013, 2009, and 2007, Pearson Education, Inc. Chapter 10 Comparing Two Groups Section 10.5 Adjusting for the Effects of Other Variables.
2014 EMNLP Xinxiong Chen, Zhiyuan Liu, Maosong Sun State Key Laboratory of Intelligent Technology and Systems Tsinghua National Laboratory for Information.
Acknowledgements Contact Information Objective An automated annotation tool was developed to assist human annotators in the efficient production of a high.
MASC The Manually Annotated Sub- Corpus of American English Nancy Ide, Collin Baker, Christiane Fellbaum, Charles Fillmore, Rebecca Passonneau.
인공지능 연구실 황명진 FSNLP Introduction. 2 The beginning Linguistic science 의 4 부분 –Cognitive side of how human acquire, produce, and understand.
11 Chapter 19 Lexical Semantics. 2 Lexical Ambiguity Most words in natural languages have multiple possible meanings. –“pen” (noun) The dog is in the.
CS 6998 NLP for the Web Columbia University 04/22/2010 Analyzing Wikipedia and Gold-Standard Corpora for NER Training William Y. Wang Computer Science.
What is Science? or 1.Science is concerned with understanding how nature and the physical world work. 2.Science can prove anything, solve any problem,
Combining Lexical Resources: Mapping Between PropBank and VerbNet Edward Loper,Szu-ting Yi, Martha Palmer September 2006.
Scientific Method Vocabulary Review Compiled By Robert Strawn.
Intelligent Database Systems Lab N.Y.U.S.T. I. M. Word sense disambiguation of WordNet glosses Presenter: Chun-Ping Wu Author: Dan Moldovan, Adrian Novischi.
CLEF Kerkyra Robust – Word Sense Disambiguation exercise UBC: Eneko Agirre, Arantxa Otegi UNIPD: Giorgio Di Nunzio UH: Thomas Mandl.
Information Retrieval using Word Senses: Root Sense Tagging Approach Sang-Bum Kim, Hee-Cheol Seo and Hae-Chang Rim Natural Language Processing Lab., Department.
1 Measuring the Semantic Similarity of Texts Author : Courtney Corley and Rada Mihalcea Source : ACL-2005 Reporter : Yong-Xiang Chen.
1 Gloss-based Semantic Similarity Metrics for Predominant Sense Acquisition Ryu Iida Nara Institute of Science and Technology Diana McCarthy and Rob Koeling.
NLP. Introduction to NLP Last week, Min broke the window with a hammer. The window was broken with a hammer by Min last week With a hammer, Min broke.
1 Fine-grained and Coarse-grained Word Sense Disambiguation Jinying Chen, Hoa Trang Dang, Martha Palmer August 22, 2003.
From Words to Senses: A Case Study of Subjectivity Recognition Author: Fangzhong Su & Katja Markert (University of Leeds, UK) Source: COLING 2008 Reporter:
Cheap and Fast – But is it Good? Evaluating Non-Expert Annotations for Natural Language Tasks EMNLP 2008 Rion Snow CS Stanford Brendan O’Connor Dolores.
Towards Semi-Automated Annotation for Prepositional Phrase Attachment Sara Rosenthal William J. Lipovsky Kathleen McKeown Kapil Thadani Jacob Andreas Columbia.
Overview of Statistical NLP IR Group Meeting March 7, 2006.
Second Language Learning From News Websites Word Sense Disambiguation using Word Embeddings.
Word Sense and Subjectivity (Coling/ACL 2006) Janyce Wiebe Rada Mihalcea University of Pittsburgh University of North Texas Acknowledgements: This slide.
Semantic Evaluation of Machine Translation Billy Wong, City University of Hong Kong 21 st May 2010.
What is Science? or 1.Science is concerned with understanding how nature and the physical world work. 2.Science can prove anything, solve any problem,
English Proposition Bank: Status Report
Coarse-grained Word Sense Disambiguation
DUMMIES RELIABILTY AND VALIDITY FOR By: Jeremy Starkey Lijia Zhang
Key NLP skills to enhance your professional practice
Text Analytics Giuseppe Attardi Università di Pisa
Statistical NLP: Lecture 9
Information Overload and National Culture: Assessing Hofstede’s Model Based on Data from Two Countries Ned Kock, Ph.D. Dept. of MIS and Decision Science.
WordNet WordNet, WSD.
Cognitive Load Theory Sweller, van Merrienboer, and Paas, 1998
Towards an Unequivocal Representation of Actions
Statistical NLP : Lecture 9 Word Sense Disambiguation
Scientific Laws & Theories
Presentation transcript:

NUMBER OR NUANCE: Factors Affecting Reliable Word Sense Annotation Susan Windisch Brown, Travis Rood, and Martha Palmer University of Colorado at Boulder

Annotators in their little nests agree; And ‘tis a shameful sight, When taggers on one project Fall out, and chide, and fight. —[adapted from] Isaac Watts

Automatic word sense disambiguation  Lexical ambiguity is a significant problem in natural language processing (NLP) applications (Agirre & Edmonds, 2006)  Text summarization  Question answering  WSD systems might help  Several studies show benefits for NLP tasks (Sanderson, 2000; Stokoe, 2003; Carpuat and Wu, 2007; Chan, Ng and Chiang, 2007)  But only with higher system accuracy (90%+) 3

Annotation reliability affects system accuracy 4 WSD systemSystem Performance Inter- annotator agreement Sense Inventory SensEval262.5%70%WordNet Chen et al. (2007) 82%89%OntoNotes Palmer (2008) 90%94%PropBank

Senses for the verb control 5 WordNetOntoNotes 1. exercise authoritative control or power over 1. exercise power or influence over; hold within limits 2. control (others or oneself) or influence skillfully 3. handle and cause to function 4. lessen the intensity of; temper 5. check or regulate (a scientific experiment) by conducting a parallel experiment 2. verify something by comparing to a standard 6. verify by using a duplicate register for comparison 7. be careful or certain to do something 8. have a firm understanding of

Possible factors affecting the reliability of word sense annotation 6  Fine-grained senses result in many senses per word, creating a heavy cognitive load on annotators, making accurate and consistent tagging difficult  Fine-grained senses are not distinct enough to reliably discriminate between

Requirements to compare fine-grained and coarse-grained annotation 7  Annotation of the same words on the same corpus instances  Sense inventories differing only in sense granularity  Previous work (Ng et al., 1999; Edmonds & Cotton, 2001; Navigli et al. 2007)

3 experiments 8  40 verbs  Number of senses : 2-26  Sense granularity: WordNet vs. OntoNotes  Exp. 1: confirm difference in reliability between fine- and coarse-grained annotation; vary granularity and number of senses  Exp. 2: hold granularity constant; vary number of senses  Exp. 3: hold number constant; vary granularity

Experiment 1  Compare fine-grained sense inventory to coarse  70 instances for each verb from the ON corpus  Annotated with WN senses by multiple pairs of annotators  Annotated with ON senses by multiple pairs of annotators  Compare the ON ITAs to the WN ITAs 9 Ave. number of senses Granularity OntoNotes6.2Coarse WN14.6Fine

Results 10

Results  Coarse-grained ON annotations had higher ITAs than fine-grained WN annotations  Number of senses No significant effect (t (79) = -1.28, p =.206).  Sense nuance Yes, a significant effect (t (79) = 10.39, p <.0001). With number of senses held constant, coarse-grained annotation is 16.2 percentage points higher than fine- grained. 11

Experiment 2: Number of senses  Hold sense granularity constant; vary # of senses  2 pairs of annotators, using fine-grained WN senses  First pair uses full set of WN senses for a word  Second pair uses a restricted set on instances that we know should fit one of those senses 12 Ave. number of sensesGranularity WN Full set14.6Fine WN Restricted set5.6Fine

13 OntoNotes grouped sense B OntoNotes grouped sense C OntoNotes grouped sense A WN WN 9 10 WN

"Then I just bought plywood, drew the pieces on it and cut them out."                    Full set of WN sensesRestricted set of WN senses

Results 15

Experiment 3  Number of senses controlled; vary sense granularity  Compare the ITAs for the ON tagging with the restricted-set WN tagging 16 Ave. number of senses Granularity OntoNotes6.2Coarse WN Restricted set5.6Fine

Results 17

Conclusion  Number of senses annotators must choose between: never a significant factor  Granularity of the senses: a significant factor, with fine-grained senses leading to lower ITAs  Poor reliability of fine-grained word sense annotation cannot be improved by reducing the cognitive load on annotators.  Annotators cannot reliably discriminate between nuanced sense distinctions. 18

Acknowledgements 19 We gratefully acknowledge the efforts of all of the annotators and the support of the National Science Foundation Grants NSF , Word Sense Disambiguation and CISE-CRI , Towards a Comprehensive Linguistic Annotation and CISE-CRI , as well as a grant from the Defense Advanced Research Projects Agency (DARPA/IPTO) under the GALE program, DARPA/CMO Contract No. HR C-0022, a subcontract from BBN, Inc.

Restricted set annotation 20  Use the adjudicated ON data to determine the ON sense for each instance.  Use instances from experiment1 that were labeled with one selected ON sense (35 instances).  Each restricted-set annotator saw only the WN senses that were clustered to form the appropriate ON sense.  Compare to the full set annotation for those instances.