1 Computational Discourse Chapter 21 November 2012 Lecture #15.

Slides:



Advertisements
Similar presentations
Discourse Structure and Discourse Coherence
Advertisements

Referring Expressions: Definition Referring expressions are words or phrases, the semantic interpretation of which is a discourse entity (also called referent)
Specialized models and ranking for coreference resolution Pascal Denis ALPAGE Project Team INRIA Rocquencourt F Le Chesnay, France Jason Baldridge.
A Machine Learning Approach to Coreference Resolution of Noun Phrases By W.M.Soon, H.T.Ng, D.C.Y.Lim Presented by Iman Sen.
Discourse Analysis David M. Cassel Natural Language Processing Villanova University April 21st, 2005 David M. Cassel Natural Language Processing Villanova.
Lecture 11: Binding and Reflexivity.  Pronouns differ from nouns in that their reference is determined in context  The reference of the word dog is.
Processing of large document collections Part 6 (Text summarization: discourse- based approaches) Helena Ahonen-Myka Spring 2006.
1 Discourse, coherence and anaphora resolution Lecture 16.
Pragmatics II: Discourse structure Ling 571 Fei Xia Week 7: 11/10/05.
Discourse Martin Hassel KTH NADA Royal Institute of Technology Stockholm
Natural Language Processing
Chapter 18: Discourse Tianjun Fu Ling538 Presentation Nov 30th, 2006.
Reference Resolution #1 CSCI-GA.2590 Ralph Grishman NYU.
Lexical chains for summarization a summary of Silber & McCoy’s work by Keith Trnka.
 Christel Kemke 2007/08 COMP 4060 Natural Language Processing Discourse and Dialogue.
ZERO PRONOUN RESOLUTION IN JAPANESE Jeffrey Shu Ling 575 Discourse and Dialogue.
Discourse: Reference Ling571 Deep Processing Techniques for NLP March 2, 2011.
Discourse: Structure Ling571 Deep Processing Techniques for NLP March 7, 2011.
Corpus 06 Discourse Characteristics. Reasons why discourse studies are not corpus-based: 1. Many discourse features cannot be identified automatically.
Final Review CS4705 Natural Language Processing. Semantics Meaning Representations –Predicate/argument structure and FOPC Thematic roles and selectional.
CS 4705 Lecture 21 Algorithms for Reference Resolution.
Natural Language Generation Martin Hassel KTH CSC Royal Institute of Technology Stockholm
1 Pragmatics: Discourse Analysis J&M’s Chapter 21.
Pragmatics I: Reference resolution Ling 571 Fei Xia Week 7: 11/8/05.
A Light-weight Approach to Coreference Resolution for Named Entities in Text Marin Dimitrov Ontotext Lab, Sirma AI Kalina Bontcheva, Hamish Cunningham,
Language Objectives. Planning Teachers should write both content and language objectives Content objectives are drawn from the subject area standards.
Empirical Methods in Information Extraction Claire Cardie Appeared in AI Magazine, 18:4, Summarized by Seong-Bae Park.
1 Natural Language Processing Computational Discourse.
1 Statistical Parsing Chapter 14 October 2012 Lecture #9.
Illinois-Coref: The UI System in the CoNLL-2012 Shared Task Kai-Wei Chang, Rajhans Samdani, Alla Rozovskaya, Mark Sammons, and Dan Roth Supported by ARL,
Binding Theory Describing Relationships between Nouns.
Introduction  Information Extraction (IE)  A limited form of “complete text comprehension”  Document 로부터 entity, relationship 을 추출 
1 CS3730/ISP3120 Discourse Processing and Pragmatics Lecture Notes Jan 10, 12.
Discourse Analysis Radhika Mamidi. Coherence and cohesion What makes a text coherent? What are the coherent devices? Discourses have to have connectivity.
Department of English Introduction To Linguistics Level Four Dr. Mohamed Younis.
UNIT 7 DEIXIS AND DEFINITENESS
1 Special Electives of Comp.Linguistics: Processing Anaphoric Expressions Eleni Miltsakaki AUTH Fall 2005-Lecture 2.
Efficiently Computed Lexical Chains As an Intermediate Representation for Automatic Text Summarization H.G. Silber and K.F. McCoy University of Delaware.
1 LIN 1310B Introduction to Linguistics Prof: Nikolay Slavkov TA: Qinghua Tang CLASS 24, April 3, 2007.
1 Learning Sub-structures of Document Semantic Graphs for Document Summarization 1 Jure Leskovec, 1 Marko Grobelnik, 2 Natasa Milic-Frayling 1 Jozef Stefan.
Processing of large document collections Part 6 (Text summarization: discourse- based approaches) Helena Ahonen-Myka Spring 2005.
Reference Resolution. Sue bought a cup of coffee and a donut from Jane. She met John as she left. He looked at her enviously as she drank the coffee.
Coherence and Coreference Introduction to Discourse and Dialogue CS 359 October 2, 2001.
1 Natural Language Processing Chapter Outline Reference –Kinds of reference phenomena –Constraints on co-reference –Preferences for co-reference.
Lecture 21 Computational Lexical Semantics Topics Features in NLTK III Computational Lexical Semantics Semantic Web USCReadings: NLTK book Chapter 10 Text.
ACE TESOL Diploma Program – London Language Institute OBJECTIVES You will understand: 1. The terminology and concepts of semantics, pragmatics and discourse.
Unit 4: REFERRING EXPRESSIONS
Topic and the Representation of Discourse Content
Reference Resolution CMSC Discourse and Dialogue September 30, 2004.
UWMS Data Mining Workshop Content Analysis: Automated Summarizing Prof. Marti Hearst SIMS 202, Lecture 16.
Foundation year WEEK TWO Lecturer: Ola Ahmed Refaat Academic year 2015 / English Language English Language Reading - ENGL 101.
Yule: “Words themselves do not refer to anything, people refer” Reference and inference Pragmatics: Reference and inference.
Grammatical and lexical coherence in writing group Done by: O`rinboyeva M. Checked by : RasulovaS.
An evolutionary approach for improving the quality of automatic summaries Constantin Orasan Research Group in Computational Linguistics School of Humanities,
Discourse: Structure and Coherence Kathy McKeown Thanks to Dan Jurafsky, Diane Litman, Andy Kehler, Jim Martin.
Discourse & Natural Language Generation Martin Hassel KTH NADA Royal Institute of Technology Stockholm
Natural Language Processing COMPSCI 423/723 Rohit Kate.
Informative Paragraph Writing 101
Discourse: Structure and Coherence Kathy McKeown Thanks to Dan Jurafsky, Diane Litman, Andy Kehler, Jim Martin.
Reference. “John went to the candy store to shop for chocolate.” “He bought some.”
Discourse Analysis Natural Language Understanding basis
Grammar.
Referring Expressions: Definition
Lecture 17: Discourse: Anaphora Resolution and Coherence
A Machine Learning Approach to Coreference Resolution of Noun Phrases
Introduction Task: extracting relational facts from text
A Machine Learning Approach to Coreference Resolution of Noun Phrases
CS4705 Natural Language Processing
Pragmatics: Reference and inference
Presentation transcript:

1 Computational Discourse Chapter 21 November 2012 Lecture #15

Discourse Consists of collocated, structured, coherent groups of sentences –What makes something a discourse as opposed to a set of unrelated sentences? –How can text be structured (related)? * Monologue: a speaker (writer) and hearer (reader) with communication flow in one direction only Dialogue: each participant takes turn being the speaker and the hearer (so 2-way participation) –Human-human dialogue –Human-computer dialogue (conversational agent) 2

Discourse Phenomina: Coreference Resolution The Tin Woodman went to the Emerald City to see the Wizard of Oz and ask for a heart. After he asked for it, the Woodman waited for the Wizard’s response. 3

Discourse Phenomina: Coreference Resolution The Tin Woodman went to the Emerald City to see the Wizard of Oz and ask for a heart. After he asked for it, the Woodman waited for the Wizard’s response. What do we need to resolve? Why is it important? –Information extraction, summarization, conversational agents 4

Coherence Relations: Coreference First Union Corp is continuing to wrestle with severe problems. According to industry insiders at Pain Webber, their president, John R. Georgius, is planning to announce his retirement tomorrow. 5

Coherence Relations: Coreference First Union Corp is continuing to wrestle with severe problems. According to industry insiders at Pain Webber, their president, John R. Georgius, is planning to announce his retirement tomorrow. First Union Corp is continuing to wrestle with severe problems. According to industry insiders at Pain Webber, their president, John R. Georgius, believes Pain Webber can be instrumental in solving most of First Union’s problems. 6

Coherence Relations (Discourse Structure) First Union Corp is continuing to wrestle with severe problems. According to industry insiders at Pain Webber, their president, John R. Georgius, is planning to announce his retirement tomorrow. Reasonable summary: First Union President John R. Georgius is planning to announce his retirement tomorrow. What you need to know: coherence relations between text segment – the first sentence is providing background for the more important 2 nd sentence. 7

Coherence (relation based) John hid Bill’s car keys. He was drunk. ?? John hid Bill’s car keys. He likes spinach. Coherence Relations – relations such as EXPLANATION or CAUSE that exists between two coherent sentences. Connections between utterances. 8

More Coherence (entity based) a)John went to his favorite music store to buy a piano. b)He had frequented the store for many years. c)He was excited that he could finally buy a piano. d)He arrived just as the store was closing for the day. e)John went to his favorite music store to buy a piano. f)It was a store John had frequented for many years. g)He was excited that he could finally buy a piano. h)It was closing just as John arrived. 9

More Coherence (entity based) a)John went to his favorite music store to buy a piano. b)He had frequented the store for many years. c)He was excited that he could finally buy a piano. d)He arrived just as the store was closing for the day. e)John went to his favorite music store to buy a piano. f)It was a store John had frequented for many years. g)He was excited that he could finally buy a piano. h)It was closing just as John arrived. 10

Discourse Segmentation We want to separate a document into a linear sequence of subtopics Unsupervised Discourse Segmentation: Marti Hearst’s TextTiling (done in early 90’s) 11

Consider a 23 paragraph article broken into segments (subtopics): 1-2 Intro to Magellan space probe 3-4 Intro to Venus 5-7 Lack of craters 8-11 Evidence of volcanic action River Styx Crustal spreading Recent volcanism Future of Magellan Wants to do this in an unsupervised fashion – how? Text Cohesion 12

Text Cohesion Halliday and Hasan (1976): “The use of certain linguistic devices to link or tie together textual units” Lexical cohesion: Indicated by relations between words in the two units (identical word, synonym, hypernym) –Before winter I built a chimney, and shingled the sides of my house.. –I thus have a tight shingled and plastered house. Non-lexical cohesion like anaphora –Peel, core and slice the pears and the apples. –Add the fruit to the skillet. 13

Intuition to a Cohesion-based approach to segmentation Sentences or paragraphs in a subtopic are cohesive with each other, but not with paragraphs in a neighboring subtopic. 14

From Hearst

TextTiling (Hearst, 1997) 1.Tokenization – convert words to lower case, remove stop words, stem words, group into pseudo- sentences 2.Lexical Score Determination – check scores between each pair of sentences = average similarity of the words in the pseudo-sentences before the gap to the pseudo-sentences after the gap 3.Boundary Identification – assign a cut-off distance to identify a new segment. 16

Copyright ©2009 by Pearson Education, Inc. Upper Saddle River, New Jersey All rights reserved. Speech and Language Processing, Second Edition Daniel Jurafsky and James H. Martin Figure 21.1

Supervised Discourse Segmentation To be used when it is relatively easy to acquire boundary-labeled training data –News stories from TV broadcasts –Paragraph segmentation Lots of different classifiers have been used –Feature set; generally a superset of those used for unsupervised segmentation –+ discourse markers and cue words Discourse Markers generally domain specific 18

Supervised Discourse Segmentation Supervised machine learning –Label segment boundaries in training and test set –Extract features in training –Learn a classifier –In testing, apply features to predict boundaries Evaluation – usual measures of precision, recall, and F-measure don’t work – need to be sensitive to near- misses. 19

Copyright ©2009 by Pearson Education, Inc. Upper Saddle River, New Jersey All rights reserved. Speech and Language Processing, Second Edition Daniel Jurafsky and James H. Martin Figure 21.2

What makes a text coherent? Appropriate use of coherence relations between subparts of the discourse --rhetorical structure Appropriate sequencing of subparts of the discourse - -discourse/topic structure Appropriate use of referring expressions 21

Coherence Relations Possible connections between utterances in a discourse. Such as in Hobbs Result: Infer that the state or event asserted by S 0 causes or could cause the state or event asserted in S 1. –The Tin Woodman was caught in the rain. His joints rusted. Explanation: Infer that the state or event asserted by S 1 causes or could cause the state or event asserted by S 0. –John hid Bill’s car keys. He was drunk. 22

Coherence Relations Parallel: Infer p(a 1, a 2,…) from the assertion of S 0 and p(b 1, b 2,…) from the assertion of S 1, where a i and b i are similar, for all i. –The scarecrow wanted some brains. The Tin Woodman wanted a heart. Elaboration: Infer the same proposition P from the assertions of S 0 and S 1. –Dorothy was from Kansas. She lived in the midst of the great Kansas prairies. 23

Coherence Relations Occasion: A change of state can be inferred from the assertion of S 0, whose final state can be inferred from S 1, or a change of state can be inferred from the assertion of S 1, whose initial state can be inferred from S 0. –Dorothy picked up the oil-can. She oiled the Tin Woodman’s joints. 24

Hierarchical structures (S1) John went to the bank to deposit his paycheck. (S2) He then took a train to Bill’s car dealership. (S3) He needed to buy a car. (S4) The company he works for now isn’t near any public transportation. (S5) He also wanted to talk to Bill about their softball league. 25

Rhetorical Structure Theory See old slides See old slides on referring and discourse models 26

5 Types of Referring Expressions Indefinite Noun Phrases: Introduces into discourse context entities that are new to the hearer. –A man, some walnuts, this new computer Definite Noun Phrases: refers to an entity that is identifiable to the hearer (e.g., been mentioned previously or well known, in set of beliefs about the world). –“a big dog…. the dog…”, the sun 27

5 Types of Referring Expressions Pronouns: another form of definite reference, generally stronger constraints on use than standard definite reference. –He, she, him, it, they… Demonstratives: demonstrative pronouns (this, that) can be alone or as determiners. Names: Common method of referring including people, organizations, and locations. 28

Features for Filtering Potential Referents Number Agreement: pronoun and referent must agree in number (single, plural) Person Agreement: 1 st, 2 nd, 3 rd Gender Agreement: male, female, nonpersonal (it) Binding Theory Constraints: constraints by syntactic relationships between a referential expression and a possible antecedent noun phrase in the same sentence. –John bought himself a new Ford. [himself = John] –John bought him a new Ford. [him ≠ John] –He said that he bought John a new Ford. [He≠John; he≠John] 29

Preferences in Pronoun Interpretation Recency – entities from most recent utterances more likely –The doctor found an old map in the captian’s chest. Jim found an even older map hidden on the shelf. It described an island. Grammatical Role – salience hierarchy of entities that is ordered by the grammatical position of the expressions that denote them. [subject, object,…] –Billy Bones went to the bar with Jim Hawkins. He called for a glass of rum. [He = Billy Bones] –Jim Hawkins went to the bar with Billy Bones. He called for a glass of rum. [He =Jim Hawkins] 30

Preferences (cont.) Repeated Mention – keep talking about the same thing. Parallelism – subject to subject; object to object. –Long John Silber went with Jim to the Old Parrot. Billy Bones went with him to the Old Anchor Inn. [him = Jim] Verb Semantics – some verbs seem to place emphasis on one of their argument positions. –John telephoned Bill. He lost the laptop. –John criticized Bill. He lost the laptop. Selectional Restrictions – other semantic knowledge playing a role. –John parked his car in the garage after driving it around for hours. [it = garage??] 31

Algorithms for co-reference resolution Hobbs Algorithm Centering Log-Linear Model (Learning Model) 32

Log-Linear Model for Pronominal Anaphora Resolution Simple supervised machine learning approach Train classifier on a hand-labeled corpus in which are marked –Positive examples – antecedents marked with each pronoun –Negative examples (derived) – pairing pronouns with non- antecendent NPs Train on set of features Given a pro-antecedent pair predict 1 if they co-refer and 0 otherwise. 33

Features for Pronominal Anaphora Resolution Strict number [true or false] Compatible number [true or false] Strict gender [true or false] Compatible gender [true or false] Sentence distance [0, 1, 2, 3,…] Hobbs distance [0, 1, 2, 3,…] (noun groups) Grammatical role [subject, object, PP] – taken by potential antecedent Linguistic form [proper, definite, indefinite, pronoun] – of the pronoun 34

Copyright ©2009 by Pearson Education, Inc. Upper Saddle River, New Jersey All rights reserved. Speech and Language Processing, Second Edition Daniel Jurafsky and James H. Martin Figure 21.8 (U1) John saw a beautiful 1961 Ford Falcon at the used car dealership. (U2) He showed it to Bob. (U3) He bought it.

Coreference Resolution Harder – must decide if any 2 noun phrases co-refer. New Features: Anaphor edit distance Antecedent edit distance Alias [true or false] – use named entity tagger Appositive [true or false] Linguistic form [proper, definite, indefinite, pronoun] 36

Evaluation Look at coreference chains as forming a set We represent the fact that A, B, and C corefer by having a class with A, B, and C in it. Reference Chain – True Chain – correct or true coreference chain an entity occurs in. Hypothesis chain – chain/class assigned to the entity by a coreference algorithm. Precision (weighted sum of correct # in hypothesis chain/# in hyp chain) and recall (# correct in hyp chain/# of elements in reference chain) 37