Unsupervised Word Sense Disambiguation Using Lesk algorithm

Slides:



Advertisements
Similar presentations
A Robust Approach to Aligning Heterogeneous Lexical Resources Mohammad Taher Pilehvar Roberto Navigli MultiJEDI ERC
Advertisements

11 Chapter 20 Part 2 Computational Lexical Semantics Acknowledgements: these slides include material from Rada Mihalcea, Ray Mooney, Katrin Erk, and Ani.
How dominant is the commonest sense of a word? Adam Kilgarriff Lexicography MasterClass Univ of Brighton.
Word Sense Disambiguation MAS.S60 Catherine Havasi Rob Speer.
1 Extended Gloss Overlaps as a Measure of Semantic Relatedness Satanjeev Banerjee Ted Pedersen Carnegie Mellon University University of Minnesota Duluth.
LEDIR : An Unsupervised Algorithm for Learning Directionality of Inference Rules Advisor: Hsin-His Chen Reporter: Chi-Hsin Yu Date: From EMNLP.
Encyclopaedic Annotation of Text.  Entity level difficulty  All the entities in a document may not be in reader’s knowledge space  Lexical difficulty.
CS Word Sense Disambiguation. 2 Overview A problem for semantic attachment approaches: what happens when a given lexeme has multiple ‘meanings’?
CS 4705 Lecture 19 Word Sense Disambiguation. Overview Selectional restriction based approaches Robust techniques –Machine Learning Supervised Unsupervised.
1 Noun Homograph Disambiguation Using Local Context in Large Text Corpora Marti A. Hearst Presented by: Heng Ji Mar. 29, 2004.
Taking the Kitchen Sink Seriously: An Ensemble Approach to Word Sense Disambiguation from Christopher Manning et al.
Presented by Zeehasham Rasheed
CS 4705 Word Sense Disambiguation. Overview Selectional restriction based approaches Robust techniques –Machine Learning Supervised Unsupervised –Dictionary-based.
1 Complementarity of Lexical and Simple Syntactic Features: The SyntaLex Approach to S ENSEVAL -3 Saif Mohammad Ted Pedersen University of Toronto, Toronto.
Using Information Content to Evaluate Semantic Similarity in a Taxonomy Presenter: Cosmin Adrian Bejan Philip Resnik Sun Microsystems Laboratories.
Word Sense Disambiguation for Automatic Taxonomy Construction from Text-Based Web Corpora 12th International Conference on Web Information System Engineering.
Evaluating the Contribution of EuroWordNet and Word Sense Disambiguation to Cross-Language Information Retrieval Paul Clough 1 and Mark Stevenson 2 Department.
Personalisation Seminar on Unlocking the Secrets of the Past: Text Mining for Historical Documents Sven Steudter.
Francisco Viveros-Jiménez Alexander Gelbukh Grigori Sidorov.
Empirical Methods in Information Extraction Claire Cardie Appeared in AI Magazine, 18:4, Summarized by Seong-Bae Park.
1 Statistical NLP: Lecture 10 Lexical Acquisition.
Modeling Documents by Combining Semantic Concepts with Unsupervised Statistical Learning Author: Chaitanya Chemudugunta America Holloway Padhraic Smyth.
Jiuling Zhang  Why perform query expansion?  WordNet based Word Sense Disambiguation WordNet Word Sense Disambiguation  Conceptual Query.
Part 3. Knowledge-based Methods for Word Sense Disambiguation.
Exploiting Ontologies for Automatic Image Annotation M. Srikanth, J. Varner, M. Bowden, D. Moldovan Language Computer Corporation
1 Statistical NLP: Lecture 9 Word Sense Disambiguation.
Paper Review by Utsav Sinha August, 2015 Part of assignment in CS 671: Natural Language Processing, IIT Kanpur.
W ORD S ENSE D ISAMBIGUATION By Mahmood Soltani Tehran University 2009/12/24 1.
Improving Subcategorization Acquisition using Word Sense Disambiguation Anna Korhonen and Judith Preiss University of Cambridge, Computer Laboratory 15.
SYMPOSIUM ON SEMANTICS IN SYSTEMS FOR TEXT PROCESSING September 22-24, Venice, Italy Combining Knowledge-based Methods and Supervised Learning for.
10/22/2015ACM WIDM'20051 Semantic Similarity Methods in WordNet and Their Application to Information Retrieval on the Web Giannis Varelas Epimenidis Voutsakis.
An Effective Word Sense Disambiguation Model Using Automatic Sense Tagging Based on Dictionary Information Yong-Gu Lee
2014 EMNLP Xinxiong Chen, Zhiyuan Liu, Maosong Sun State Key Laboratory of Intelligent Technology and Systems Tsinghua National Laboratory for Information.
Page 1 SenDiS Sectoral Operational Programme "Increase of Economic Competitiveness" "Investments for your future" Project co-financed by the European Regional.
CS 4705 Lecture 19 Word Sense Disambiguation. Overview Selectional restriction based approaches Robust techniques –Machine Learning Supervised Unsupervised.
CS774. Markov Random Field : Theory and Application Lecture 19 Kyomin Jung KAIST Nov
HyperLex: lexical cartography for information retrieval Jean Veronis Presented by: Siddhanth Jain( ) Samiulla Shaikh( )
Wikipedia as Sense Inventory to Improve Diversity in Web Search Results Celina SantamariaJulio GonzaloJavier Artiles nlp.uned.es UNED,c/Juan del Rosal,
Lecture 21 Computational Lexical Semantics Topics Features in NLTK III Computational Lexical Semantics Semantic Web USCReadings: NLTK book Chapter 10 Text.
Using Semantic Relatedness for Word Sense Disambiguation
1 Masters Thesis Presentation By Debotosh Dey AUTOMATIC CONSTRUCTION OF HASHTAGS HIERARCHIES UNIVERSITAT ROVIRA I VIRGILI Tarragona, June 2015 Supervised.
Number Sense Disambiguation Stuart Moore Supervised by: Anna Korhonen (Computer Lab)‏ Sabine Buchholz (Toshiba CRL)‏
Information Retrieval using Word Senses: Root Sense Tagging Approach Sang-Bum Kim, Hee-Cheol Seo and Hae-Chang Rim Natural Language Processing Lab., Department.
Multi-level Bootstrapping for Extracting Parallel Sentence from a Quasi-Comparable Corpus Pascale Fung and Percy Cheung Human Language Technology Center,
Improved Video Categorization from Text Metadata and User Comments ACM SIGIR 2011:Research and development in Information Retrieval - Katja Filippova -
Finding document topics for improving topic segmentation Source: ACL2007 Authors: Olivier Ferret (18 route du Panorama, BP6) Reporter:Yong-Xiang Chen.
1 Measuring the Semantic Similarity of Texts Author : Courtney Corley and Rada Mihalcea Source : ACL-2005 Reporter : Yong-Xiang Chen.
1 Gloss-based Semantic Similarity Metrics for Predominant Sense Acquisition Ryu Iida Nara Institute of Science and Technology Diana McCarthy and Rob Koeling.
Using Wikipedia for Hierarchical Finer Categorization of Named Entities Aasish Pappu Language Technologies Institute Carnegie Mellon University PACLIC.
From Words to Senses: A Case Study of Subjectivity Recognition Author: Fangzhong Su & Katja Markert (University of Leeds, UK) Source: COLING 2008 Reporter:
2/10/2016Semantic Similarity1 Semantic Similarity Methods in WordNet and Their Application to Information Retrieval on the Web Giannis Varelas Epimenidis.
Overview of Statistical NLP IR Group Meeting March 7, 2006.
Word Sense and Subjectivity (Coling/ACL 2006) Janyce Wiebe Rada Mihalcea University of Pittsburgh University of North Texas Acknowledgements: This slide.
Finding Predominant Word Senses in Untagged Text Diana McCarthy & Rob Koeling & Julie Weeds & Carroll Department of Indormatics, University of Sussex {dianam,
NTNU Speech Lab 1 Topic Themes for Multi-Document Summarization Sanda Harabagiu and Finley Lacatusu Language Computer Corporation Presented by Yi-Ting.
Word Sense Disambiguation Algorithms in Hindi
Graph-based WSD の続き DMLA /7/10 小町守.
Coarse-grained Word Sense Disambiguation
SENSEVAL: Evaluating WSD Systems
Using lexical chains for keyword extraction
Lecture 21 Computational Lexical Semantics
Statistical NLP: Lecture 9
WordNet WordNet, WSD.
CS 620 Class Presentation Using WordNet to Improve User Modelling in a Web Document Recommender System Using WordNet to Improve User Modelling in a Web.
A method for WSD on Unrestricted Text
Enriching Taxonomies With Functional Domain Knowledge
CS224N Section 3: Corpora, etc.
Giannis Varelas Epimenidis Voutsakis Paraskevi Raftopoulou
Statistical NLP : Lecture 9 Word Sense Disambiguation
Statistical NLP: Lecture 10
Presentation transcript:

Unsupervised Word Sense Disambiguation Using Lesk algorithm Su Weifeng Sep. 22, 2004

The SENSEVAL–3 English Lexical Sample Task Rada Mihalcea, Timothy Chklovski (Senseval-3) Disambiguating Noun Groupings with Respect to WordNet Senses Philip Resnik (1995) Finding Predominant Word Senses in Untagged Text Diana McCarthy, Rob Koeling, Julie Weeds, John Carroll (ACL2004, Best Paper Award) HKUST Human Language Technology Center © 2003

An unsupervised WSD method Intuition: Lesk algorithm assigns a target word the sense that is most related to the senses of its neighboring words. An unsupervised WSD method Intuition: Related word senses will be defined using similar words, and there will be overlaps in their definitions that will indicate their relatedness. We disambiguate a polysemous word by picking the sense of the target word whose definition has the most words in common with the definitions of other words in a given window of context. Key point: Similarity or relatedness between senses Definition of the sense HKUST Human Language Technology Center © 2003

A gloss-based similarity method is proposed for WSD. The SENSEVAL–3 English Lexical Sample Task (Rada Mihalcea, Timothy Chklovski) 66.1 precision & 65.7 recall fine-grained scoring can be get for Senseval-3 English lexical sample task. The baseline precision for this task was 53.5% The second best score of unsupervised method for Senseval-3 English lexical sample task reaches a precision/recall to be 56.3% A gloss-based similarity method is proposed for WSD. HKUST Human Language Technology Center © 2003

A gloss-centered method is proposed. Definition for each sense of the target word is represented as a concatenation of two glosses of the sense: Descriptive glosses: the description of the sense of a word 1. (10) tape -- (a long thin piece of cloth or paper as used for binding or fastening; "he used a piece of tape for a belt"; "he wrapped a tape around the package") 2. (2) tape, tape recording, taping -- (a recording made on magnetic tape; "the several recordings were combined on a master tape") Hypernymy glosses: what hierarchical the sense belong to Vesuvius#n#1 => volcano => mountain, mount => natural elevation, elevation => geological formation => natural object => object, physical object => entity Cosine similarity between words of the context and words of the definition is used as the similarity measure HKUST Human Language Technology Center © 2003

Disambiguating Noun Groupings with Respect to WordNet Senses (Philip Resnik, 1995) Task statement: automatic sense disambiguation of nouns appearing within sets of related nouns. Example: burglars, thief, rob, mugging, stray, robbing, look out, chase, crate, thieves we expect to assign a high score for sense ``lookout, lookout man, sentinel, sentry, watch, scout: a person employed to watch for something to happen.'' and a low score for sense ``an observation tower, or to the activity of watching.'' HKUST Human Language Technology Center © 2003

An information content based semantic similarity measure is used. Information content is a measure of specificity that is assigned to each sense in a hierarchy. A sense with a high information content is very specific to a particular topic Senses with lower information content are associated with more general, less specific senses. carving fork has a higher information content than entity has. Information content of a sense is estimated by counting the frequency of that sense in a large corpus and thereby determining its probability via a maximum likelihood estimate. IC(sense)=-log(P(sense)) lcs: least common subsumer lcs(c1,c2) c1 c2 HKUST Human Language Technology Center © 2003

The related words of the target word are treated as context. The algorithm assign performing well at the task of assigning a high score to the best sense, it does a good job of assigning low scores to senses that are clearly inappropriate. HKUST Human Language Technology Center © 2003

Finding Predominant Word Senses in Untagged Text (Diana McCarthy, Rob Koeling, Julie Weeds, John Carroll 2004) Task statement: Use a thesaurus acquired from raw textual corpora to find predominant noun senses automatically. For each target noun, Input: k nearest neighbors and their distributional similarity score to the target word. WordNet similarity package which give a semantic similarity measure Output: The predominant sense of the target noun HKUST Human Language Technology Center © 2003

Finding Predominant Word Senses in Untagged Text (Diana McCarthy, Rob Koeling, Julie Weeds, John Carroll 2004) Task statement: Use a thesaurus acquired from raw textual corpora to find predominant noun senses automatically. For each target noun, Input: k nearest neighbors and their distributional similarity score to the target word. WordNet similarity package which give a semantic similarity measure Output: The predominant sense of the target noun HKUST Human Language Technology Center © 2003

HKUST Human Language Technology Center © 2003

An adapted Lesk algorithm is performed An information content based semantic similarity measure is used. Jcn: The semantic similarity is multiplied by a distributional similarity HKUST Human Language Technology Center © 2003

The method can effectively discover appropriate predominant senses for words within a corpus. The acquired predominant senses give a precision of 64% on the nouns of the SENSEVAL-2 English all-words task. This is nearly as well the first sense provided by Semcor. The method discovers appropriate predominant senses for words from two domain specific corpora. When performed on SemCor, the method get a precision of 54% while the random selection can only reach a precision of 32%. The first sense in SemCor provides an upperbound for this task of 67%. HKUST Human Language Technology Center © 2003

Conclusions The gloss and hierarchy of a sense provides valuable information for sense disambiguation. Unsupervised WSD method can reach a precision comparable with supervised WSD method. HKUST Human Language Technology Center © 2003