Algorithms for Reference Resolution

Slides:



Advertisements
Similar presentations
Referring Expressions: Definition Referring expressions are words or phrases, the semantic interpretation of which is a discourse entity (also called referent)
Advertisements

Computational language: week 10 Lexical Knowledge Representation concluded Syntax-based computational language Sentence structure: syntax Context free.
Pronouns and Reference Resolution CS 4705 Julia Hirschberg CS 4705.
Specialized models and ranking for coreference resolution Pascal Denis ALPAGE Project Team INRIA Rocquencourt F Le Chesnay, France Jason Baldridge.
Albert Gatt LIN3022 Natural Language Processing Lecture 11.
Curs 7: Teorii ale discursului Centering Dan Cristea Selecţie de slide-uri prezentate în tutoriale (RANLP-03, Borovits, Sept. 2003; ICON-04, Hyderabad,
Discourse Analysis David M. Cassel Natural Language Processing Villanova University April 21st, 2005 David M. Cassel Natural Language Processing Villanova.
Automatic Essay Scoring Evaluation of text coherence for electronic essay scoring systems (E. Miltsakaki and K. Kukich, 2004) Universität des Saarlandes.
Processing of large document collections Part 6 (Text summarization: discourse- based approaches) Helena Ahonen-Myka Spring 2006.
1 Discourse, coherence and anaphora resolution Lecture 16.
Discourse Martin Hassel KTH NADA Royal Institute of Technology Stockholm
Anaphora Resolution Spring 2010, UCSC – Adrian Brasoveanu [Slides based on various sources, collected over a couple of years and repeatedly modified –
Chapter 18: Discourse Tianjun Fu Ling538 Presentation Nov 30th, 2006.
Discourse: Reference Ling571 Deep Processing Techniques for NLP March 2, 2011.
CS 4705 Algorithms for Reference Resolution. Anaphora resolution Finding in a text all the referring expressions that have one and the same denotation.
Final Review CS4705 Natural Language Processing. Semantics Meaning Representations –Predicate/argument structure and FOPC Thematic roles and selectional.
Basi di dati distribuite Prof. M.T. PAZIENZA a.a
CS 4705 Lecture 21 Algorithms for Reference Resolution.
1 Special Electives of Comp.Linguistics: Processing Anaphoric Expressions Eleni Miltsakaki AUTH Fall 2005-Lecture 4.
Natural Language Generation Martin Hassel KTH CSC Royal Institute of Technology Stockholm
Centering theory and its direct applications
Improving Machine Learning Approaches to Coreference Resolution Vincent Ng and Claire Cardie Cornell Univ. ACL 2002 slides prepared by Ralph Grishman.
1 Pragmatics: Discourse Analysis J&M’s Chapter 21.
Pragmatics I: Reference resolution Ling 571 Fei Xia Week 7: 11/8/05.
CS 4705 Final Review CS4705 Julia Hirschberg. Format and Coverage Covers only material from thru (i.e. beginning with Probabilistic Parsing) Same format.
Pronouns and Reference Resolution CS  HW3 deadline extended to Wednesday, Nov. 25 th at 11:58 pm  Michael Collins will talk Thursday, Dec. 3 rd.
11 CS 388: Natural Language Processing: Syntactic Parsing Raymond J. Mooney University of Texas at Austin.
A Light-weight Approach to Coreference Resolution for Named Entities in Text Marin Dimitrov Ontotext Lab, Sirma AI Kalina Bontcheva, Hamish Cunningham,
Empirical Methods in Information Extraction Claire Cardie Appeared in AI Magazine, 18:4, Summarized by Seong-Bae Park.
1 Natural Language Processing Computational Discourse.
1 Special Electives of Comp.Linguistics: Processing Anaphoric Expressions Eleni Miltsakaki AUTH Fall 2005-Lecture 3.
Differential effects of constraints in the processing of Russian cataphora Kazanina and Phillips 2010.
A multiple knowledge source algorithm for anaphora resolution Allaoua Refoufi Computer Science Department University of Setif, Setif 19000, Algeria .
On the Issue of Combining Anaphoricity Determination and Antecedent Identification in Anaphora Resolution Ryu Iida, Kentaro Inui, Yuji Matsumoto Nara Institute.
Incorporating Extra-linguistic Information into Reference Resolution in Collaborative Task Dialogue Ryu Iida Shumpei Kobayashi Takenobu Tokunaga Tokyo.
1 Exploiting Syntactic Patterns as Clues in Zero- Anaphora Resolution Ryu Iida, Kentaro Inui and Yuji Matsumoto Nara Institute of Science and Technology.
1 Special Electives of Comp.Linguistics: Processing Anaphoric Expressions Eleni Miltsakaki AUTH Fall 2005-Lecture 2.
1 Learning Sub-structures of Document Semantic Graphs for Document Summarization 1 Jure Leskovec, 1 Marko Grobelnik, 2 Natasa Milic-Frayling 1 Jozef Stefan.
A Cross-Lingual ILP Solution to Zero Anaphora Resolution Ryu Iida & Massimo Poesio (ACL-HLT 2011)
Opinion Holders in Opinion Text from Online Newspapers Youngho Kim, Yuchul Jung and Sung-Hyon Myaeng Reporter: Chia-Ying Lee Advisor: Prof. Hsin-Hsi Chen.
REFERENTIAL CHOICE AS A PROBABILISTIC MULTI-FACTORIAL PROCESS Andrej A. Kibrik, Grigorij B. Dobrov, Natalia V. Loukachevitch, Dmitrij A. Zalmanov
Processing of large document collections Part 6 (Text summarization: discourse- based approaches) Helena Ahonen-Myka Spring 2005.
Reference Resolution. Sue bought a cup of coffee and a donut from Jane. She met John as she left. He looked at her enviously as she drank the coffee.
Coherence and Coreference Introduction to Discourse and Dialogue CS 359 October 2, 2001.
1 Natural Language Processing Chapter Outline Reference –Kinds of reference phenomena –Constraints on co-reference –Preferences for co-reference.
Reference Resolution CMSC Discourse and Dialogue September 30, 2004.
Supertagging CMSC Natural Language Processing January 31, 2006.
Dec 11, Human Parsing Do people use probabilities for parsing?! Sentence processing Study of Human Parsing.
Evaluation issues in anaphora resolution and beyond Ruslan Mitkov University of Wolverhampton Faro, 27 June 2002.
Support Vector Machines and Kernel Methods for Co-Reference Resolution 2007 Summer Workshop on Human Language Technology Center for Language and Speech.
◦ Process of describing the structure of phrases and sentences Chapter 8 - Phrases and sentences: grammar1.
1 Special Electives of Comp.Linguistics: Processing Anaphoric Expressions Eleni Miltsakaki AUTH Fall 2005-Lecture 5.
Reference Resolution CMSC Natural Language Processing January 15, 2008.
Discourse Analysis Natural Language Understanding basis
CSC 594 Topics in AI – Natural Language Processing
Simone Paolo Ponzetto University of Heidelberg Massimo Poesio
Statistical NLP: Lecture 3
Entity- & Topic-Based Information Ordering
4 (c) parsing.
Referring Expressions: Definition
Social Knowledge Mining
CS 388: Natural Language Processing: Syntactic Parsing
Lecture 17: Discourse: Anaphora Resolution and Coherence
Anaphora Resolution Spring 2010, UCSC – Adrian Brasoveanu
CSCI 5832 Natural Language Processing
Chunk Parsing CS1573: AI Application Development, Spring 2003
Linguistic Essentials
CS4705 Natural Language Processing
The Winograd Schema Challenge Hector J. Levesque AAAI, 2011
David Kauchak CS159 – Spring 2019
Presentation transcript:

Algorithms for Reference Resolution CS 4705 Algorithms for Reference Resolution CS 4705

Anaphora resolution Finding in a text all the referring expressions that have one and the same denotation Pronominal anaphora resolution Anaphora resolution between named entities Full noun phrase anaphora resolution

Review: What Factors Affect Reference Resolution? Lexical factors Reference type: Inferability, discontinuous set, generics, one anaphora, pronouns,… Discourse factors: Recency Focus/topic structure, digression Repeated mention Syntactic factors: Agreement: gender, number, person, case Parallel construction Grammatical role

Semantic/lexical factors Pragmatic factors Selectional restrictions Semantic/lexical factors Verb semantics, thematic role Pragmatic factors

Reference Resolution Given these types of constraints, can we construct an algorithm that will apply them such that we can identify the correct referents of anaphors and other referring expressions?

Issues Which constraints/features can/should we make use of? How should we order them? I.e. which override which? What should be stored in our discourse model? I.e., what types of information do we need to keep track of? How to evaluate?

Three Algorithms Lappin & Leas ‘94: weighting via recency and syntactic preferences Hobbs ‘78: syntax tree-based referential search Centering (Grosz, Joshi, Weinstein, ‘95 and various): discourse-based search

Lappin & Leass ‘94 Weights candidate antecedents by recency and syntactic preference (86% accuracy) Two major functions to perform: Update the discourse model when an NP that evokes a new entity is found in the text, computing the salience of this entity for future anaphora resolution Find most likely referent for current anaphor by considering possible antecedents and their salience values Partial example for 3P, non-reflexives

Saliency Factor Weights Sentence recency (in current sentence?) 100 Subject emphasis (is it the subject?) 80 Existential emphasis (existential prednom?) 70 Accusative emphasis (is it the dir obj?) 50 Indirect object/oblique comp emphasis 40 Non-adverbial emphasis (not in PP,) 50 Head noun emphasis (is head noun) 80

Implicit ordering of arguments: subj/exist pred/obj/indobj-oblique/dem.advPP On the sofa, the cat was eating bonbons. sofa: 100+80=180 cat: 100+80+50+80=310 bonbons: 100+50+50+80=280 Update: Weights accumulate over time Cut in half after each sentence processed Salience values for subsequent referents accumulate for equivalence class of co-referential items (exceptions, e.g. multiple references in same sentence)

The bonbons were clearly very tasty. sofa: 180/2=90 cat: 310/2=155 bonbons: 280/2 +(100+80+50+80)=450 Additional salience weights for grammatical role parallelism (35) and cataphora (-175) calculated when pronoun to be resolved Additional constraints on gender/number agrmt/syntax They were a gift from an unknown admirer. sofa: 90/2=45 cat: 155/2=77.5 bonbons: 450/2=225 (+35) = 260….

Reference Resolution Collect potential referents (up to four sentences back): {sofa,cat,bonbons} Remove those that don’t agree in number/gender with pronoun {bonbons} Remove those that don’t pass intra-sentential syntactic coreference constraints The cat washed it. (itcat) Add applicable values for role parallelism (+35) or cataphora (-175) to current salience value for each potential antecedent Select referent with highest salience; if tie, select closest referent in string

A Different Aproach: Centering Theory (Grosz et al 1995) examines interactions between local coherence and the choice of referring expressions A pretty woman entered the restaurant. She sat at the table next to mine… A woman entered the restaurant. They like ice cream.

Centering theory: Motivation (Grosz et al 1995) examine interactions between local coherence and the choice of referring expressions Pronouns and definite descriptions are not equivalent with respect to their effect on coherence Different inference demands on the hearer/reader.

Centering theory: Definitions The centers of an utterance are discourse entities serving to link the utterance to other utterances Forward looking centers: a ranked list A backward looking center: the entity currently ‘in focus’ or salient Centers are semantic objects, not words, phrases, or syntactic forms but They are realized by such in an utterance Their realization can give us clues about their likely salience

More Definitions More on discourse centers and utterances Un: an utterance Backward-looking center Cb(Un): current focus after Un interpreted Forward-looking centers Cf(Un): ordered list of potential focii referred to in Un Cb(Un+1) is highest ranked member of Cf(Un) Cf may be ordered subj<exist. Prednom<obj<indobj-oblique<dem. advPP (Brennan et al) Cp(Un): preferred (highest ranked) center of Cf(Un)

Transitions from Un to Un+1

Rules If any element of Cf(Un) is pronominalized in Un+1, then Cb(Un+1) must also be Preference: Continue > Retain > Smooth-Shift > Rough-Shift Algorithm Generate Cb and Cf assignments for all possible reference assignments Filter by constraints (syntactic coreference, selectional restrictions,…) Rank by preference among transition orderings

Example U1:George gave Harry a cookie. U2:He baked the cookie Thursday. U3: He ate the cookie all up. One Cf(U1): {George,cookie,Harry} Cp(U1): George Cb(U1): undefined Two Cf(U2): {George,cookie,Thursday} Cp(U2): George Cb(U2): George Continue (Cp(U2)=Cb(U2); Cb(U1) undefined

Three Or, Three The winner is…..George! Cf(U3): {George?,cookie} Cp(U3): George? Cb(U3): George? Continue (Cp(U3)=Cb(U3); Cb(U3)= Cb(U2) Or, Three Cf(U3): {Harry?,cookie} Cp(U3): Harry? Cb(U3): Harry? Smooth-Shift (Cp(U3)=Cb(U3); Cb(U3)  Cb(U2) The winner is…..George!

Centering Theory vs. Lappin & Leass Centering sometimes prefers an antecedent Lappin and Leass (or Hobbs) would consider to have low salience Always prefers a single pronominalization strategy: prescriptive, assumes discourse coherent Constraints too simple: grammatical role, recency, repeated mention Assumes correct syntactic information available as input

Evaluation Centering only now being specified enough to be tested automatically on real data Specifying the Parameters of Centering Theory: A Corpus-Based Evaluation using Text from Application-Oriented Domains (Poesio et al., ACL 2000) Walker ‘89 manual comparison of Centering vs. Hobbs ‘78 Only 281 examples from 3 genres Assumed correct features given as input to each Centering 77.6% vs. Hobbs 81.8% Lappin and Leass’ 86% accuracy on test set from computer training manuals

Rule-based vs. Statistical Approaches (Kennedy & Boguraev 1996), (Lappin & Leass 1994) vs (Ge, Hale & Charniak 1998) Performed on full syntactic parse vs on shallow syntactic parse (Lap 1994), (Ge 1998) vs (Ken 1996) Type of text used for the evaluation (Lap 1994) computer manual texts (86% accuracy) (Ge 1998) WSJ articles (83% accuracy) (Ken 1996) different genres (75% accuracy)