Specialized models and ranking for coreference resolution Pascal Denis ALPAGE Project Team INRIA Rocquencourt F-78153 Le Chesnay, France Jason Baldridge.

Slides:



Advertisements
Similar presentations
CILC2011 A framework for structured knowledge extraction and representation from natural language via deep sentence analysis Stefania Costantini Niva Florio.
Advertisements

A Machine Learning Approach to Coreference Resolution of Noun Phrases By W.M.Soon, H.T.Ng, D.C.Y.Lim Presented by Iman Sen.
QA-LaSIE Components The question document and each candidate answer document pass through all nine components of the QA-LaSIE system in the order shown.
NYU ANLP-00 1 Automatic Discovery of Scenario-Level Patterns for Information Extraction Roman Yangarber Ralph Grishman Pasi Tapanainen Silja Huttunen.
Processing of large document collections Part 6 (Text summarization: discourse- based approaches) Helena Ahonen-Myka Spring 2006.
Jean-Eudes Ranvier 17/05/2015Planet Data - Madrid Trustworthiness assessment (on web pages) Task 3.3.
LEDIR : An Unsupervised Algorithm for Learning Directionality of Inference Rules Advisor: Hsin-His Chen Reporter: Chi-Hsin Yu Date: From EMNLP.
Predicting Text Quality for Scientific Articles AAAI/SIGART-11 Doctoral Consortium Annie Louis : Louis A. and Nenkova A Automatically.
Predicting the Semantic Orientation of Adjective Vasileios Hatzivassiloglou and Kathleen R. McKeown Presented By Yash Satsangi.
CS 4705 Lecture 21 Algorithms for Reference Resolution.
Supervised models for coreference resolution Altaf Rahman and Vincent Ng Human Language Technology Research Institute University of Texas at Dallas 1.
Improving Machine Learning Approaches to Coreference Resolution Vincent Ng and Claire Cardie Cornell Univ. ACL 2002 slides prepared by Ralph Grishman.
NATURAL LANGUAGE TOOLKIT(NLTK) April Corbet. Overview 1. What is NLTK? 2. NLTK Basic Functionalities 3. Part of Speech Tagging 4. Chunking and Trees 5.
Mining and Summarizing Customer Reviews
A Global Relaxation Labeling Approach to Coreference Resolution Coling 2010 Emili Sapena, Llu´ıs Padr´o and Jordi Turmo TALP Research Center Universitat.
A Light-weight Approach to Coreference Resolution for Named Entities in Text Marin Dimitrov Ontotext Lab, Sirma AI Kalina Bontcheva, Hamish Cunningham,
Evaluating the Contribution of EuroWordNet and Word Sense Disambiguation to Cross-Language Information Retrieval Paul Clough 1 and Mark Stevenson 2 Department.
Andreea Bodnari, 1 Peter Szolovits, 1 Ozlem Uzuner 2 1 MIT, CSAIL, Cambridge, MA, USA 2 Department of Information Studies, University at Albany SUNY, Albany,
Thumbs Up or Thumbs Down? Semantic Orientation Applied to Unsupervised Classification on Reviews Peter D. Turney Institute for Information Technology National.
Empirical Methods in Information Extraction Claire Cardie Appeared in AI Magazine, 18:4, Summarized by Seong-Bae Park.
Authors: Ting Wang, Yaoyong Li, Kalina Bontcheva, Hamish Cunningham, Ji Wang Presented by: Khalifeh Al-Jadda Automatic Extraction of Hierarchical Relations.
A Comparison of Features for Automatic Readability Assessment Lijun Feng 1 Matt Huenerfauth 1 Martin Jansche 2 No´emie Elhadad 3 1 City University of New.
Illinois-Coref: The UI System in the CoNLL-2012 Shared Task Kai-Wei Chang, Rajhans Samdani, Alla Rozovskaya, Mark Sammons, and Dan Roth Supported by ARL,
On the Issue of Combining Anaphoricity Determination and Antecedent Identification in Anaphora Resolution Ryu Iida, Kentaro Inui, Yuji Matsumoto Nara Institute.
Incorporating Extra-linguistic Information into Reference Resolution in Collaborative Task Dialogue Ryu Iida Shumpei Kobayashi Takenobu Tokunaga Tokyo.
A semantic based methodology to classify and protect sensitive data in medical records Flora Amato, Valentina Casola, Antonino Mazzeo, Sara Romano Dipartimento.
Arabic Tokenization, Part-of-Speech Tagging and Morphological Disambiguation in One Fell Swoop Nizar Habash and Owen Rambow Center for Computational Learning.
SYMPOSIUM ON SEMANTICS IN SYSTEMS FOR TEXT PROCESSING September 22-24, Venice, Italy Combining Knowledge-based Methods and Supervised Learning for.
A Language Independent Method for Question Classification COLING 2004.
Efficiently Computed Lexical Chains As an Intermediate Representation for Automatic Text Summarization H.G. Silber and K.F. McCoy University of Delaware.
1 Learning Sub-structures of Document Semantic Graphs for Document Summarization 1 Jure Leskovec, 1 Marko Grobelnik, 2 Natasa Milic-Frayling 1 Jozef Stefan.
A Cross-Lingual ILP Solution to Zero Anaphora Resolution Ryu Iida & Massimo Poesio (ACL-HLT 2011)
Opinion Holders in Opinion Text from Online Newspapers Youngho Kim, Yuchul Jung and Sung-Hyon Myaeng Reporter: Chia-Ying Lee Advisor: Prof. Hsin-Hsi Chen.
Exploiting Context Analysis for Combining Multiple Entity Resolution Systems -Ramu Bandaru Zhaoqi Chen Dmitri V.kalashnikov Sharad Mehrotra.
COLING 2012 Extracting and Normalizing Entity-Actions from Users’ comments Swapna Gottipati, Jing Jiang School of Information Systems, Singapore Management.
REFERENTIAL CHOICE AS A PROBABILISTIC MULTI-FACTORIAL PROCESS Andrej A. Kibrik, Grigorij B. Dobrov, Natalia V. Loukachevitch, Dmitrij A. Zalmanov
Processing of large document collections Part 6 (Text summarization: discourse- based approaches) Helena Ahonen-Myka Spring 2005.
Coherence and Coreference Introduction to Discourse and Dialogue CS 359 October 2, 2001.
An Entity-Mention Model for Coreference Resolution with Inductive Logic Programming Xiaofeng Yang 1 Jian Su 1 Jun Lang 2 Chew Lim Tan 3 Ting Liu 2 Sheng.
Using Semantic Relations to Improve Passage Retrieval for Question Answering Tom Morton.
Automatic Evaluation of Linguistic Quality in Multi- Document Summarization Pitler, Louis, Nenkova 2010 Presented by Dan Feblowitz and Jeremy B. Merrill.
Multilingual Opinion Holder Identification Using Author and Authority Viewpoints Yohei Seki, Noriko Kando,Masaki Aono Toyohashi University of Technology.
Multi-level Bootstrapping for Extracting Parallel Sentence from a Quasi-Comparable Corpus Pascale Fung and Percy Cheung Human Language Technology Center,
Evaluation issues in anaphora resolution and beyond Ruslan Mitkov University of Wolverhampton Faro, 27 June 2002.
1 Evaluating High Accuracy Retrieval Techniques Chirag Shah,W. Bruce Croft Center for Intelligent Information Retrieval Department of Computer Science.
Measuring the Influence of Errors Induced by the Presence of Dialogs in Reference Clustering of Narrative Text Alaukik Aggarwal, Department of Computer.
4. Relationship Extraction Part 4 of Information Extraction Sunita Sarawagi 9/7/2012CS 652, Peter Lindes1.
Support Vector Machines and Kernel Methods for Co-Reference Resolution 2007 Summer Workshop on Human Language Technology Center for Language and Speech.
Answer Mining by Combining Extraction Techniques with Abductive Reasoning Sanda Harabagiu, Dan Moldovan, Christine Clark, Mitchell Bowden, Jown Williams.
1 Measuring the Semantic Similarity of Texts Author : Courtney Corley and Rada Mihalcea Source : ACL-2005 Reporter : Yong-Xiang Chen.
1 Gloss-based Semantic Similarity Metrics for Predominant Sense Acquisition Ryu Iida Nara Institute of Science and Technology Diana McCarthy and Rob Koeling.
FILTERED RANKING FOR BOOTSTRAPPING IN EVENT EXTRACTION Shasha Liao Ralph York University.
Using Wikipedia for Hierarchical Finer Categorization of Named Entities Aasish Pappu Language Technologies Institute Carnegie Mellon University PACLIC.
From Words to Senses: A Case Study of Subjectivity Recognition Author: Fangzhong Su & Katja Markert (University of Leeds, UK) Source: COLING 2008 Reporter:
Using Semantic Relations to Improve Information Retrieval
Learning Event Durations from Event Descriptions Feng Pan, Rutu Mulkar, Jerry R. Hobbs University of Southern California ACL ’ 06.
Ontology Engineering and Feature Construction for Predicting Friendship Links in the Live Journal Social Network Author:Vikas Bahirwani 、 Doina Caragea.
Identifying Expressions of Opinion in Context Eric Breck and Yejin Choi and Claire Cardie IJCAI 2007.
Linguistic Graph Similarity for News Sentence Searching
A Deep Memory Network for Chinese Zero Pronoun Resolution
Simone Paolo Ponzetto University of Heidelberg Massimo Poesio
NYU Coreference CSCI-GA.2591 Ralph Grishman.
Natural Language Processing (NLP)
张昊.
Improving a Pipeline Architecture for Shallow Discourse Parsing
Clustering Algorithms for Noun Phrase Coreference Resolution
A Machine Learning Approach to Coreference Resolution of Noun Phrases
Natural Language Processing (NLP)
Extracting Why Text Segment from Web Based on Grammar-gram
Natural Language Processing (NLP)
Presentation transcript:

Specialized models and ranking for coreference resolution Pascal Denis ALPAGE Project Team INRIA Rocquencourt F Le Chesnay, France Jason Baldridge Department of Linguistics University of Texas at Austin In Proceedings of EMNLP Yi-Ting Huang 2010/12/24

Outline 1.Introduction 2.Anaphoric filter 3.Ranking 4.Specialized models 5.Experiment 6.Conclusion 7.Comments 2

1. Introduction Coreference resolution is the task of partitioning a set of entity mentions in a text, where each partition corresponds to some entity in an underlying discourse model. pair-wise classification (c1, m)=Y (c2, m)=N (c3, m)=N (c4, m)=Y (c5, m)=Y (c6, m)=N Combine entity1(c1, c4, c5, m) entity2(c2, c3) entity3(c6) 3 c1….c2……. …c3……c4.. ……c5…….. c6………m..

Research problem 1.Numerous approaches to anaphora and coreference resolution reduce these tasks to a binary classification task, whereby pairs of mentions are classified as coreferential or not. – these approaches make very strong independence assumptions. 2.Only one pair-wise classification to identify any type of noun phrases. 4 pair-wise classification (c1, m)=Y (c2, m)=N (c3, m)=N (c4, m)=Y (c5, m)=Y (c6, m)=N Combine entity1(c1, c4, c5, m) entity2(c2, c3) entity3(c6) c1….c2……. …c3……c4.. ……c5…….. c6………m..

However… (1/2) For many authors, the relation takes the form of a continuum and is often represented in the form of a referential hierarchy, such as: The higher up, the more accessible (or salient) the entity is. This type of hierarchy is validated by corpus studies of the distribution of different types of expressions. 5 M. Ariel Referring and accessibility. Journal of Linguistics, pages 65–87.

However… (2/2) At the extremes are pronouns (these forms typically require a previous mention in the local context) and proper names (these forms are often used without previous mentions of the entity). For instance, pronouns find their antecedents very locally (in a window of 1-2 sentences), while proper names predominantly find theirs at longer distances (Ariel, 1988). 6

Research purpose The use of rankers instead of classifiers The use of linguistically motivated, specialized models for different types of mentions. – third-person pronouns, – speech pronouns (i.e., first and second person pronouns), – proper names, – definite descriptions – other types of nominals (e.g., anaphoric uses of indefinite, quantified, and bare noun phrases). 7

Procedure discourse status classifier Step 1 anaphoric filter m a part of mention a new entity Step 2 3 rd person pronouns Step 3 specialized models and ranker ((c1, c4, c5), m) c1….c2……. …c3……c4.. ……c5…….. c6………m.. ((c2, c3), m) ((c6), m) m proper namesothersSpeech pronounsdefinite descriptions identify m’s type Step 2 identify type 8 (c6, m)

Term definition referential expressions  mention a discourse entity  entity 9 c1….c2……. …c3……c4.. ……c5…….. c6………m.. Combine entity1(c1, c4, c5, m) entity2(c2, c3) entity3(c6)

Procedure discourse status classifier Step 1 anaphoric filter m a part of mention a new entity Step 2 3 rd person pronouns Step 3 specialized models and ranker ((c1, c4, c5), m) c1….c2……. …c3……c4.. ……c5…….. c6………m.. ((c2, c3), m) ((c6), m) m proper names others Speech pronouns definite descriptions identify m’s type Step 2 identify type 10 (c6, m)

2. anaphoric filter (1/2) Not all referential expressions in a given document are anaphors: some expressions introduce a discourse entity, rather than accessing an existing one. It is more troublesome for rankers, which always pick an antecedent from the candidate set. A natural solution is to use a model that specifically predicts the discourse status (discourse-new vs. discourse-old) of each expression: only expressions that are classified as “discourse-old” by this model are considered by rankers. 11

2. anaphoric filter (2/2) We use a similar discourse status classifier to Ng and Cardie’s (Ng and Cardie, 2002a) as a filter on mentions for our rankers. We rely on three main types of information sources: – the form of mention (e.g., type of linguistic expression, number of tokens), – positional features in the text, – comparisons of the given mention to the mentions that precede it in the text. Evaluated on the ACE datasets, the model achieves an overall accuracy score of 80.8%, compared to a baseline of 59.7% when predicting the majority class (“discourse-old”). 12 V. Ng and C. Cardie. 2002a. Identifying anaphoric and non-anaphoric noun phrases to improve coreference resolution. In Proceedings of COLING 2002.

Procedure discourse status classifier Step 1 anaphoric filter m a part of mention a new entity Step 2 c1….c2……. …c3……c4.. ……c5…….. c6………m.. 3 rd person pronouns Step 3 specialized models and ranker ((c1, c4, c5), m) ((c2, c3), m) ((c6), m) m proper namesothersSpeech pronounsdefinite descriptions identify m’s type Step 2 identify type 13 (c6, m)

3. Ranking (1/3) 14 c1….c2……. …c3……c4.. ……c5…….. c6………m..

3. Ranking (2/3) For the training of the different ranking models, we use the following procedure. For each model, instances are created by pairing each anaphor with a set of candidates which contains: – true antecedent – a set of non-antecedents 15

3. Ranking (3/3) True antecedent: The selection of the true antecedent varies depending on the model we are training: – for pronominal forms, the antecedent is selected as the closest preceding mention in the chain; – for non-pronominal forms, we used the closest preceding non-pronominal mention in the chain as the antecedent. Non-antecedent set: we collect all the non-antecedents that appear in a window of two sentences around the antecedent. 16

3. Specialized models (1/2) Our second strategy is to use different, specialized models for different referential expressions. We use separate models for the following types: 1.third person pronouns, 2.speech pronouns, 3.proper names, 4.definite descriptions, and 5.others. 17

3. Specialized models (2/2) Note that our split of referential types only partially cover the referential hierarchies of Ariel (1988). Thus, there is no separate model for demonstrative noun phrases and pronouns: these are very rare in the corpus we used (i.e., the ACE corpus). 18

3.1 features (1/6) The total number of anaphors (i.e., of mentions that are not chain heads) in the data is and 4599 for training and testing, respectively % 11-13% 33-40% 16-17% 19

3.1 features (2/6) Our feature extraction relies on limited linguistic processing: we only made use of a sentence detector, a tokenizer, a POS tagger (as provided by the OpenNLP Toolkit) and the WordNet database. Since we did not use parser, lexical heads for the NP mentions were computed using simple heuristics relying solely on POS sequences. 20

3.1 features (3/6) 21

3.1 features (4/6) Linguistic form: Distance: Morph syntactic Agreement Semantic compatibility 22 we collected the synonym set (or synset) as well as the synset of their direct hypernyms associated with each mention. we used the synset associated with the first sense associated with the mention’s head word. we collected the synonym set (or synset) as well as the synset of their direct hypernyms associated with each mention. we used the synset associated with the first sense associated with the mention’s head word.

3.1 features (5/6) Context: – these features can be seen as approximations of the grammatical roles, as indicators of the salience of the potential candidate. the part of speech tags surrounding the candidate, indicates whether the potential antecedent is the first mention in a sentence (approximating subject-hood), indicating whether the candidate is embedded inside another mention. 23

3.1 features (6/6) String similarity: for PN, Def-NP, Other Acronym: for PN, Def-NP Apposition: for PN 24

5.1 Experiment 1 Purpose: their ability to select a correct antecedent for each anaphor. Accuracy: the ratio of correctly resolved anaphors. 25

5.2 Experiment 2 Metric: MUC, B 3, CEAF DataSet: ACE02 26 Ng and Cardie (2002b) V. Ng and C. Cardie. 2002b. Improving machine learning approaches to coreference resolution. In Proceedings of ACL 2002, pages 104–111.

5.2 error analysis missed anaphors – i.e., an anaphoric mention that fails to be linked to a previous mention spurious anaphors – i.e., an non-anaphoric mention that is linked to a previous mention invalid resolutions – a true anaphor that is linked to a incorrect antecedent 27

5.3 Experiment 3 RANK+DS-ORACLE+SP – uses the specialized rankers in combination with a perfect discourse status classifier. LINK-ORACLE – uses the discourse status classifier with a perfect coreference resolver. 28

6. Conclusion We present and evaluate two straight-forward tactics for improving coreference resolution: – ranking models, and – separate, specialized models for different types of referring expressions. This simple pipeline architecture produces significant improvements over various implementations of the standard, classifier- based coreference system. 29

7. Comments Advantage: – The proposed method by ranker and specialized model, and the result of 3 rd pronouns, proper names and link-oracle shows good performance. 30

Thank you for your attention & Merry Christmas! 31