Reference Resolution CMSC 35900-1 Discourse and Dialogue October 12, 2004.

Slides:



Advertisements
Similar presentations
School of something FACULTY OF OTHER School of Computing FACULTY OF ENGINEERING Chunking: Shallow Parsing Eric Atwell, Language Research Group.
Advertisements

Referring Expressions: Definition Referring expressions are words or phrases, the semantic interpretation of which is a discourse entity (also called referent)
Specialized models and ranking for coreference resolution Pascal Denis ALPAGE Project Team INRIA Rocquencourt F Le Chesnay, France Jason Baldridge.
A Machine Learning Approach to Coreference Resolution of Noun Phrases By W.M.Soon, H.T.Ng, D.C.Y.Lim Presented by Iman Sen.
1 Discourse, coherence and anaphora resolution Lecture 16.
Pragmatics II: Discourse structure Ling 571 Fei Xia Week 7: 11/10/05.
Discourse Martin Hassel KTH NADA Royal Institute of Technology Stockholm
1 Spoken Dialogue Systems Dialogue and Conversational Agents (Part IV) Chapter 19: Draft of May 18, 2005 Speech and Language Processing: An Introduction.
INTERNATIONAL CONFERENCE ON NATURAL LANGUAGE PROCESSING NLP-AI IIIT-Hyderabad CIIL, Mysore ICON DECEMBER, 2003.
Predicting Text Quality for Scientific Articles Annie Louis University of Pennsylvania Advisor: Ani Nenkova.
Discourse: Reference Ling571 Deep Processing Techniques for NLP March 2, 2011.
NaLIX: A Generic Natural Language Search Environment for XML Data Presented by: Erik Mathisen 02/12/2008.
CS 4705 Discourse Structure and Text Coherence. What makes a text/dialogue coherent? Incoherent? “Consider, for example, the difference between passages.
Discourse Structure Grosz and Sidner. Why bother? Leads to an account of discourse meaning Constrains how utterances are related Useful for explaining.
Final Review CS4705 Natural Language Processing. Semantics Meaning Representations –Predicate/argument structure and FOPC Thematic roles and selectional.
CS 4705 Lecture 21 Algorithms for Reference Resolution.
Supervised models for coreference resolution Altaf Rahman and Vincent Ng Human Language Technology Research Institute University of Texas at Dallas 1.
Weakness of Structural linguistics Functionalism
A Light-weight Approach to Coreference Resolution for Named Entities in Text Marin Dimitrov Ontotext Lab, Sirma AI Kalina Bontcheva, Hamish Cunningham,
AQUAINT Kickoff Meeting – December 2001 Integrating Robust Semantics, Event Detection, Information Fusion, and Summarization for Multimedia Question Answering.
Challenges in Information Retrieval and Language Modeling Michael Shepherd Dalhousie University Halifax, NS Canada.
Richard Socher Cliff Chiung-Yu Lin Andrew Y. Ng Christopher D. Manning
Models of Discourse Structure II Discourse & Dialogue CMSC October 14, 2004.
Empirical Methods in Information Extraction Claire Cardie Appeared in AI Magazine, 18:4, Summarized by Seong-Bae Park.
Lecture 12: 22/6/1435 Natural language processing Lecturer/ Kawther Abas 363CS – Artificial Intelligence.
Interactive Dialogue Systems Professor Diane Litman Computer Science Department & Learning Research and Development Center University of Pittsburgh Pittsburgh,
Discourse Markers Discourse & Dialogue CS November 25, 2006.
Discourse Topics, Linguistics, and Language Teaching Richard Watson Todd King Mongkut’s University of Technology Thonburi arts.kmutt.ac.th/crs/research/
Illinois-Coref: The UI System in the CoNLL-2012 Shared Task Kai-Wei Chang, Rajhans Samdani, Alla Rozovskaya, Mark Sammons, and Dan Roth Supported by ARL,
Scott Duvall, Brett South, Stéphane Meystre A Hands-on Introduction to Natural Language Processing in Healthcare Annotation as a Central Task for Development.
Ling 570 Day 17: Named Entity Recognition Chunking.
1 Statistical NLP: Lecture 9 Word Sense Disambiguation.
Psycholinguistic Theory
Reference Resolution Natural Language Processing January 22, 2008.
1 Special Electives of Comp.Linguistics: Processing Anaphoric Expressions Eleni Miltsakaki AUTH Fall 2005-Lecture 2.
1 Learning Sub-structures of Document Semantic Graphs for Document Summarization 1 Jure Leskovec, 1 Marko Grobelnik, 2 Natasa Milic-Frayling 1 Jozef Stefan.
Indirect Supervision Protocols for Learning in Natural Language Processing II. Learning by Inventing Binary Labels This work is supported by DARPA funding.
Coherence and Coreference Introduction to Discourse and Dialogue CS 359 October 2, 2001.
A Scalable Machine Learning Approach for Semi-Structured Named Entity Recognition Utku Irmak(Yahoo! Labs) Reiner Kraft(Yahoo! Inc.) WWW 2010(Information.
Recognizing Discourse Structure: Speech Discourse & Dialogue CMSC October 11, 2006.
1/21 Automatic Discovery of Intentions in Text and its Application to Question Answering (ACL 2005 Student Research Workshop )
ACE TESOL Diploma Program – London Language Institute OBJECTIVES You will understand: 1. The terminology and concepts of semantics, pragmatics and discourse.
Reference Resolution- Extension CMSC Discourse and Dialogue October 2, 2006.
Dialog Models September 18, 2003 Thomas Harris.
Inference Protocols for Coreference Resolution Kai-Wei Chang, Rajhans Samdani, Alla Rozovskaya, Nick Rizzolo, Mark Sammons, and Dan Roth This research.
Discourse & Dialogue CS 359 November 13, 2001
Reference Resolution CMSC Discourse and Dialogue September 30, 2004.
Tuning in to children’s thinking and learning
Evaluation issues in anaphora resolution and beyond Ruslan Mitkov University of Wolverhampton Faro, 27 June 2002.
Measuring the Influence of Errors Induced by the Presence of Dialogs in Reference Clustering of Narrative Text Alaukik Aggarwal, Department of Computer.
Reference Resolution CMSC Natural Language Processing January 15, 2008.
An evolutionary approach for improving the quality of automatic summaries Constantin Orasan Research Group in Computational Linguistics School of Humanities,
Intention & Cooperation Discourse and Dialogue CS 359 October 18, 2001.
Discourse: Structure and Coherence Kathy McKeown Thanks to Dan Jurafsky, Diane Litman, Andy Kehler, Jim Martin.
Instance Discovery and Schema Matching With Applications to Biological Deep Web Data Integration Tantan Liu, Fan Wang, Gagan Agrawal {liut, wangfa,
Agent-Based Dialogue Management Discourse & Dialogue CMSC November 10, 2006.
AQUAINT Mid-Year PI Meeting – June 2002 Integrating Robust Semantics, Event Detection, Information Fusion, and Summarization for Multimedia Question Answering.
Relation Extraction (RE) via Supervised Classification See: Jurafsky & Martin SLP book, Chapter 22 Exploring Various Knowledge in Relation Extraction.
Simone Paolo Ponzetto University of Heidelberg Massimo Poesio
NYU Coreference CSCI-GA.2591 Ralph Grishman.
Natural Language Processing (NLP)
Improving a Pipeline Architecture for Shallow Discourse Parsing
Clustering Algorithms for Noun Phrase Coreference Resolution
Statistical NLP: Lecture 9
Discourse Structure in Generation
A Machine Learning Approach to Coreference Resolution of Noun Phrases
Natural Language Processing (NLP)
CS4705 Natural Language Processing
Information Retrieval
Natural Language Processing (NLP)
Presentation transcript:

Reference Resolution CMSC Discourse and Dialogue October 12, 2004

Agenda Coherence: Holding discourse together –Coherence types and relations Reference resolution –Knowledge-rich, deep analysis approaches –Lappin&Leass, Centering, Hobbs –Knowledge-based, shallow analysis: CogNIAC (‘95) –Learning approaches: Fully, Weakly Supervised –Cardie&Ng ’02,’03,’04

Data-driven Reference Resolution Prior approaches –Knowledge-based, hand-crafted Data-driven machine learning approach –Cast coreference as classification problem For each pair NPi,NPj, do they corefer? Cluster to form equivalence classes

NP Coreference Examples Link all NPs that refer to same entity Queen Elizabeth set about transforming her husband, King George VI, into a viable monarch. Logue, a renowned speech therapist, was summoned to help the King overcome his speech impediment...

Training Instances 25 features per instance: 2NPs, features, class –lexical (3) string matching for pronouns, proper names, common nouns –grammatical (18) pronoun_1, pronoun_2, demonstrative_2, indefinite_2, … number, gender, animacy appositive, predicate nominative binding constraints, simple contra-indexing constraints, … span, maximalnp, … –semantic (2) same WordNet class alias –positional (1) distance between the NPs in terms of # of sentences –knowledge-based (1) naïve pronoun resolution algorithm

Classification & Clustering Classifiers: –C4.5 (Decision Trees), RIPPER Cluster: Best-first, single link clustering –Each NP in own class –Test preceding NPs –Select highest confidence coref, merge classes Tune: Training sample skew: class, type

Classifier for MUC-6 Data Set

Problem 1 NP3NP4NP5NP6NP7NP8NP9NP2NP1 farthest antecedent Coreference is a rare relation –skewed class distributions (2% positive instances) –remove some negative instances

Problem 2 Coreference is a discourse-level problem –different solutions for different types of NPs proper names: string matching and aliasing –inclusion of “hard” positive training instances –positive example selection: selects easy positive training instances (cf. Harabagiu et al. (2001)) Queen Elizabeth set about transforming her husband, King George VI, into a viable monarch. Logue, the renowned speech therapist, was summoned to help the King overcome his speech impediment...

Problem 3 Coreference is an equivalence relation –loss of transitivity –need to tighten the connection between classification and clustering –prune learned rules w.r.t. the clustering-level coreference scoring function [Queen Elizabeth] set about transforming [her] [husband],... coref ? not coref ?

Weakly Supervised Learning Exploit small pool of labeled training data –Larger pool unlabeled Single-View Multi-Learner Co-training –2 different learning algorithms, same feature set –each classifier labels unlabeled instances for the other classifier –data pool is flushed after each iteration

Effectiveness Supervised learning approaches –Comparable performance to knowledge-based Weakly supervised approaches –Decent effectiveness, still lags supervised –Dramatically less labeled training data 1K vs 500K

Reference Resolution: Extensions Cross-document co-reference (Baldwin & Bagga 1998) –Break “the document boundary” –Question: “John Smith” in A = “John Smith” in B? –Approach: Integrate: –Within-document co-reference with –Vector Space Model similarity

Cross-document Co-reference Run within-document co-reference (CAMP) –Produce chains of all terms used to refer to entity Extract all sentences with reference to entity –Pseudo per-entity summary for each document Use Vector Space Model (VSM) distance to compute similarity between summaries

Cross-document Co-reference Experiments: –197 NYT articles referring to “John Smith” 35 different people, 24: 1 article each With CAMP: Precision 92%; Recall 78% Without CAMP: Precision 90%; Recall 76% Pure Named Entity: Precision 23%; Recall 100%

Conclusions Co-reference establishes coherence Reference resolution depends on coherence Variety of approaches: –Syntactic constraints, Recency, Frequency,Role Similar effectiveness - different requirements Co-reference can enable summarization within and across documents (and languages!)

Challenges Alternative approaches to reference resolution –Different constraints, rankings, combination Different types of referent –Speech acts, propositions, actions, events –“Inferrables” - e.g. car -> door, hood, trunk,.. –Discontinuous sets –Generics –Time

Discourse Structure Theories Discourse & Dialogue CMSC October 12, 2004

Roadmap Goals of Discourse Structure Models –Limitations of early approaches Models of Discourse Structure –Attention & Intentions (Grosz & Sidner 86) –Rhetorical Structure Theory (Mann & Thompson 87) Contrasts, Constraints & Conclusions

Why Model Discourse Structure? (Theoretical) Discourse: not just constituent utterances –Create joint meaning –Context guides interpretation of constituents –How???? –What are the units? –How do they combine to establish meaning? How can we derive structure from surface forms? –What makes discourse coherent vs not? –How do they influence reference resolution?

Why Model Discourse Structure? (Applied) Design better summarization, understanding Improve speech synthesis –Influenced by structure Develop approach for generation of discourse Design dialogue agents for task interaction Guide reference resolution

Early Discourse Models Schemas & Plans (McKeown, Reichman, Litman & Allen) –Task/Situation model = discourse model Specific->General: “restaurant” -> AI planning Topic/Focus Theories ( Grosz 76, Sidner 76 ) –Reference structure = discourse structure Speech Act – single utt intentions vs extended discourse

Discourse Models: Common Features Hierarchical, Sequential structure applied to subunits –Discourse “segments” – Need to detect, interpret Referring expressions provide coherence –Explain and link Meaning of discourse more than that of component utterances Meaning of units depends on context

Earlier Models Issues: –Conflate different aspects of discourse Task plan, discourse plan –Ignore aspects of discourse Goals & intentions vs focus –Overspecific Fixed plan, schema, relation inventory

Attention, Intentions and the Structure of Discourse Grosz&Sidner (1986) Goals: –Integrate approaches for focus (reference res.), plan/task structure, discourse structure, goals Three part model: –Linguistic structure (utterances) –Attentional structure (focus, reference) –Intentional structure (plans, purposes)

Linguistic Structure Utterances group into discourse segments –Hierarchical, not ne, cessarily contiguous –Not strictly decompositional 2-way interactions –Utterances define structure; Cue phrases mark segment boundaries –But, okay, fine, incidentally –Structure guides interpretation –Reference

Intentional Structure Discourse & participants: overall purpose –Discourse segments have purposes (DP/DSP) Contribute to overall Main DP/DSP intended to be recognized

Intentional Structure: Relations Two relations between purposes –Dominance DSP1 dominates DSP2 if doing DSP2 contributes to achieving DSP1 –Satisfaction-Precedence DSP1 must be satisfied before DSP2 Purposes: –Intend that someone know something, do something, believe something, etc –Open-ended

Attentional State Captures focus of attention in discourse –Incremental –Focus Spaces Include entities salient/evoked in discourse Include a current DSP Stack-structured: –higher->more salient, lower still accessible –Push:segment contributes to previous DSP –Pop: segment to contributes to more dominant DSP »Tied to intentional structure

Attentional State cntd. Focusing structure depends on the intentional structure: the relationships between DSPs determine pushes and pops from the stack Focusing structure coordinates the linguistic and intentional structures during processing Like the other 2 structures, focusing structure evolves as discourse proceeds

Discourse examples Essay Task-oriented dialog –Intentional structure is neither identical nor isomorphic to the general plan

The "movies" are so attractive to the great American public, especially to young people, that it is time to take careful thought about their effect on mind and morals. Ought any parent to permit his children to attend a moving picture show often or without being quite certain of the show he permits them to see? No one can deny, of course, that great educational and ethical gains may be made through the movies because of their astonishing vividness. But the important fact to be determined is the total result of continuous and indiscriminate attendance on shows of this kind. Can it other than harmful? In the first place the character of the plays is seldom of the best. One has only to read the ever-present "movie" billboard to see how cheap, melodramatic and vulgar most of the photoplays are. Even the best plays, moreover, are bound to be exciting and over-emotional. Without spoken words, facial expression and gesture must carry the meaning: but only strong emotion or buffoonery can be represented through facial expression and gesture. The more reasonable and quiet aspects of life are necessarily neglected. How can our young people drink in through their eyes a continuous spectacle of intense and strained activity and feeling without harmful effects? Parents and teachers will do well to guard the young against overindulgence in the taste for the "movie"

H:1. First you have to remove the flywheel. R:2. How do I remove the flywheel? H:3. First, loosen the screw, then pull it off. R:4. OK.5. The tool I have is awkward. Is there another tool that I could use instead? H:6. Show me the tool you are using. R:7. OK. H:8. Are you sure you are using the right size key? R:9. I’ll try some others. 10. I found an angle I can get at it. 11. The screw is loose, but I’m having trouble getting the flywheel off. H:12. Use the wheelpuller. Do you know how to use it ? R:13. No. H:14. Do you know what it looks like? R:15. Yes. H:16. Show it to me please. R:17. OK. H:18. Good. Loosen the screw in the center and place the jaws around the hub of the flywheel, then tighten the screw onto the center of the shaft. The flywheel should slide off.

Processing issues Intention recognition –What info can be used to recognize an intention –At what point does this info become available Overall processing module has to be able to operate on partial information It must allow for incrementally constraining the range of possibilities on the basis of new info that becomes available as the segment progresses

Info constraining DSP: –Specific linguistic markers –Utterance-level intentions –General knowledge about actions and objects in the domain of discourse Applications of the theory: –Interruptions Weak – not linked to immediate DSP Strong - not linked to any DSP –Cue words

Interruption John came by and left the groceries Stop that you kids And I put them away after he left John, groceries DSP1 kids DSP2 John, groceries DSP1

Conclusions Generalizes approaches to task-oriented dialogue –Goal: Domain-independence –Broad, general, abstract model Accounts for interesting phenomena –Interruptions, returns, cue phrases

More conclusions Asks more questions than it answers. How do we implement these aspects of dialog? –Is it remotely feasible????