Word Sense Disambiguation MAS.S60 Catherine Havasi Rob Speer.

Slides:



Advertisements
Similar presentations
Multi-Document Person Name Resolution Michael Ben Fleischman (MIT), Eduard Hovy (USC) From Proceedings of ACL-42 Reference Resolution workshop 2004.
Advertisements

A Robust Approach to Aligning Heterogeneous Lexical Resources Mohammad Taher Pilehvar Roberto Navigli MultiJEDI ERC
How dominant is the commonest sense of a word? Adam Kilgarriff Lexicography MasterClass Univ of Brighton.
TÍTULO GENÉRICO Concept Indexing for Automated Text Categorization Enrique Puertas Sanz Universidad Europea de Madrid.
Sentiment Analysis An Overview of Concepts and Selected Techniques.
A Survey on Text Categorization with Machine Learning Chikayama lab. Dai Saito.
What is Statistical Modeling
Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 28, 2011.
Semi-supervised learning and self-training LING 572 Fei Xia 02/14/06.
CS Word Sense Disambiguation. 2 Overview A problem for semantic attachment approaches: what happens when a given lexeme has multiple ‘meanings’?
CS347 Review Slides (IR Part II) June 6, 2001 ©Prabhakar Raghavan.
CS 4705 Lecture 19 Word Sense Disambiguation. Overview Selectional restriction based approaches Robust techniques –Machine Learning Supervised Unsupervised.
1 Noun Homograph Disambiguation Using Local Context in Large Text Corpora Marti A. Hearst Presented by: Heng Ji Mar. 29, 2004.
I256 Applied Natural Language Processing Fall 2009
Taking the Kitchen Sink Seriously: An Ensemble Approach to Word Sense Disambiguation from Christopher Manning et al.
CS 4705 Word Sense Disambiguation. Overview Selectional restriction based approaches Robust techniques –Machine Learning Supervised Unsupervised –Dictionary-based.
Multi-view Exploratory Learning for AKBC Problems Bhavana Dalvi and William W. Cohen School Of Computer Science, Carnegie Mellon University Motivation.
An Information Theoretic Approach to Bilingual Word Clustering Manaal Faruqui & Chris Dyer Language Technologies Institute SCS, CMU.
Aiding WSD by exploiting hypo/hypernymy relations in a restricted framework MEANING project Experiment 6.H(d) Luis Villarejo and Lluís M à rquez.
Evaluating the Contribution of EuroWordNet and Word Sense Disambiguation to Cross-Language Information Retrieval Paul Clough 1 and Mark Stevenson 2 Department.
COMP423: Intelligent Agent Text Representation. Menu – Bag of words – Phrase – Semantics – Bag of concepts – Semantic distance between two words.
CLEF Ǻrhus Robust – Word Sense Disambiguation exercise UBC: Eneko Agirre, Oier Lopez de Lacalle, Arantxa Otegi, German Rigau UVA & Irion: Piek Vossen.
2007. Software Engineering Laboratory, School of Computer Science S E Towards Answering Opinion Questions: Separating Facts from Opinions and Identifying.
A Simple Unsupervised Query Categorizer for Web Search Engines Prashant Ullegaddi and Vasudeva Varma Search and Information Extraction Lab Language Technologies.
Information Extraction MAS.S60 Catherine Havasi Rob Speer.
Word Sense Disambiguation UIUC - 06/10/2004 Word Sense Disambiguation Another NLP working problem for learning with constraints… Lluís Màrquez TALP, LSI,
1 Statistical NLP: Lecture 9 Word Sense Disambiguation.
This work is supported by the Intelligence Advanced Research Projects Activity (IARPA) via Department of Interior National Business Center contract number.
Paper Review by Utsav Sinha August, 2015 Part of assignment in CS 671: Natural Language Processing, IIT Kanpur.
Word Sense Disambiguation Reading: Chap 16-17, Jurafsky & Martin Instructor: Rada Mihalcea.
W ORD S ENSE D ISAMBIGUATION By Mahmood Soltani Tehran University 2009/12/24 1.
SYMPOSIUM ON SEMANTICS IN SYSTEMS FOR TEXT PROCESSING September 22-24, Venice, Italy Combining Knowledge-based Methods and Supervised Learning for.
An Effective Word Sense Disambiguation Model Using Automatic Sense Tagging Based on Dictionary Information Yong-Gu Lee
2014 EMNLP Xinxiong Chen, Zhiyuan Liu, Maosong Sun State Key Laboratory of Intelligent Technology and Systems Tsinghua National Laboratory for Information.
CS 4705 Lecture 19 Word Sense Disambiguation. Overview Selectional restriction based approaches Robust techniques –Machine Learning Supervised Unsupervised.
Word Sense Disambiguation Kyung-Hee Sung Foundations of Statistical NLP Chapter 7.
CS774. Markov Random Field : Theory and Application Lecture 19 Kyomin Jung KAIST Nov
11 Chapter 19 Lexical Semantics. 2 Lexical Ambiguity Most words in natural languages have multiple possible meanings. –“pen” (noun) The dog is in the.
11 Chapter 14 Part 1 Statistical Parsing Based on slides by Ray Mooney.
Today Ensemble Methods. Recap of the course. Classifier Fusion
1 Bootstrapping without the Boot Jason Eisner Damianos Karakos HLT-EMNLP, October 2005.
Wikipedia as Sense Inventory to Improve Diversity in Web Search Results Celina SantamariaJulio GonzaloJavier Artiles nlp.uned.es UNED,c/Juan del Rosal,
Lecture 21 Computational Lexical Semantics Topics Features in NLTK III Computational Lexical Semantics Semantic Web USCReadings: NLTK book Chapter 10 Text.
1/21 Automatic Discovery of Intentions in Text and its Application to Question Answering (ACL 2005 Student Research Workshop )
Word Translation Disambiguation Using Bilingial Bootsrapping Paper written by Hang Li and Cong Li, Microsoft Research Asia Presented by Sarah Hunter.
Combining labeled and unlabeled data for text categorization with a large number of categories Rayid Ghani KDD Lab Project.
CS460/IT632 Natural Language Processing/Language Technology for the Web Lecture 24 (14/04/06) Prof. Pushpak Bhattacharyya IIT Bombay Word Sense Disambiguation.
NUMBER OR NUANCE: Factors Affecting Reliable Word Sense Annotation Susan Windisch Brown, Travis Rood, and Martha Palmer University of Colorado at Boulder.
1 Masters Thesis Presentation By Debotosh Dey AUTOMATIC CONSTRUCTION OF HASHTAGS HIERARCHIES UNIVERSITAT ROVIRA I VIRGILI Tarragona, June 2015 Supervised.
Number Sense Disambiguation Stuart Moore Supervised by: Anna Korhonen (Computer Lab)‏ Sabine Buchholz (Toshiba CRL)‏
1 Measuring the Semantic Similarity of Texts Author : Courtney Corley and Rada Mihalcea Source : ACL-2005 Reporter : Yong-Xiang Chen.
1 Gloss-based Semantic Similarity Metrics for Predominant Sense Acquisition Ryu Iida Nara Institute of Science and Technology Diana McCarthy and Rob Koeling.
Knowledge Structure Vijay Meena ( ) Gaurav Meena ( )
Cheap and Fast – But is it Good? Evaluating Non-Expert Annotations for Natural Language Tasks EMNLP 2008 Rion Snow CS Stanford Brendan O’Connor Dolores.
Second Language Learning From News Websites Word Sense Disambiguation using Word Embeddings.
Word Sense and Subjectivity (Coling/ACL 2006) Janyce Wiebe Rada Mihalcea University of Pittsburgh University of North Texas Acknowledgements: This slide.
In the name of God Simple past tense Table of contents 1-Positive form 2-Question form 3-Negative form.
BAYESIAN LEARNING. 2 Bayesian Classifiers Bayesian classifiers are statistical classifiers, and are based on Bayes theorem They can calculate the probability.
Concept-Based Analysis of Scientific Literature Chen-Tse Tsai, Gourab Kundu, Dan Roth UIUC.
Graph-based WSD の続き DMLA /7/10 小町守.
Coarse-grained Word Sense Disambiguation
SENSEVAL: Evaluating WSD Systems
Automatic Hedge Detection
Vector-Space (Distributional) Lexical Semantics
Lecture 21 Computational Lexical Semantics
Category-Based Pseudowords
Statistical NLP: Lecture 9
WordNet WordNet, WSD.
Unsupervised Word Sense Disambiguation Using Lesk algorithm
Statistical NLP : Lecture 9 Word Sense Disambiguation
Presentation transcript:

Word Sense Disambiguation MAS.S60 Catherine Havasi Rob Speer

Banks? The edge of a river – “I fished on the bank of the Mississippi.” A financial institution – “Bank of America failed to return my call.” The building that houses the financial institution – “The bank burned down last Thursday.” A “biological repository” – “I gave blood at the blood bank”.

Word Sense Disambiguation Most NLP tasks need WSD “Played a lot of pool last night… my bank shot is improving!” Usually keying to WordNet “I hit the ball with the bat.”

Types “All words” – Guess the WN sysnet Lexical Subset – A small number of pre-defined words Course Word Sense – All words, but more intuitive senses

Types “All words” – Guess the WN sysnet Lexical Subset – A small number of pre-defined words Coarse Word Sense – All words, but more intuitive senses IAA is 75-80% for all words task with WordNet 90% for simple binary tasks

What is a Coarse Word Sense? How many word senses does the word “bag” have in WordNet?

What is a Coarse Word Sense? How many word senses does the word “bag” have in WordNet? – 9 noun senses, 5 verb senses Coarse WSD: 6 nouns, 2 verbs A Coarse WordNet: 6,000 words ( Navigli and Litkowski 2006) These distinctions are hard even for humans (Snyder and Palmer 2004) – Fine Grained IAA: 72.5% – Coarse Grained IAA: 86.4%

“Bag”: Noun 1. A coarse sense containing: – bag (a flexible container with a single opening) – bag, handbag, pocketbook, purse (a container used for carrying money and small personal items or accessories) – bag, bagful (the quantity that a bag will hold) – bag, traveling bag, travelling bag, grip, suitcase (a portable rectangular container for carrying clothes) 2. bag (the quantity of game taken in a particular period) 3. base, bag (a place that the runner must touch before scoring) 4. bag, old bag (an ugly or ill-tempered woman) 5. udder, bag (mammary gland of bovids (cows and sheep and goats)) 6. cup of tea, bag, dish (an activity that you like or at which you are superior)

Frequent Ingredients Open Mind Word Expert WordNet eXtended WordNet (XWN) SemCor 3.0 (“brown1” and “brown2”) ConceptNet

Semcor

No training set, no problem Julia Hockenmaier’s “Psudoword” evaluation Pick two random words – Say, “banana” and “door” Combine them together – “BananaDoor” Replace all instances of either in your corpora with your new pseudoword Evaluate A bit easier…

The “Flip-flop” Method Stephen Brown and Jonathan Rose, 1991 Find a single feature or set of features which disambiguated the words – think the named entity recognizer

An Example

Standard Techniques Naïve Bayes (notice a trend) – Bag of words – Priors are based on word frequencies Unsupervised clustering techniques – Expectation Maximization (EM) – Yarowsky

Yarowsky (slides from Julia Hockenmaier)

Training Yarowsky

Created a blend using a large number of resources Created an ad hoc category for a word and its surroundings in sentence Find which word sense is most similar to category Keep the system machinery as general as possible. Using OMCS

Adding Associations ConceptNet was included in two forms: – Concept vs. feature matrices – Concept-to-concept associations Associations help to represent topic areas If the document mentions computer-related words, expect more computer-related word senses

Constructing the Blend

Calculating the Right Sense “I put my money in the bank”

SemEval Task 7 14 different systems were submitted in 2007 Baseline: Most frequent sense Spoiler!: Our system would have placed 4 th Top three systems: – NUS-PT: parallel corpora with SVM (Chang et al, 2007) – NUS-ML: Bayesian LDA with specialized features (Chai, et al, 2007) – LCC-WSD: multiple methods approach with end- to-end system and corpora (Novichi et al, 2007)

Results

Parallel Corpora IMVHO the “right” way to do it. Different words have different sense in different languages Use parallel corpora to find those instances – Like Euro or UN proceedings

English and Romanian

Gold standards are overrated Rada Mihalcea, 2007: “Using Wikipedia for Automatic Word Sense Disambiguation”

Lab: making a simple supervised WSD classifier Big thanks to some guy with a blog (Jim Plush) Training data: Wikipedia articles surrounding “Apple” (the fruit) and “Apple Inc.” Test data: hand-classified tweets about apples and Apple products Use familiar features + Naïve Bayes to get > 90% accuracy Optional: use it with tweetstream to show only tweets about apples (the fruit)

Slide Thanks James Pustejovsky, Gerard Bakx, Julie Hockenmaier Manning and Schutze