Discourse: Structure and Coherence Kathy McKeown Thanks to Dan Jurafsky, Diane Litman, Andy Kehler, Jim Martin.

Slides:



Advertisements
Similar presentations
Discourse Structure and Discourse Coherence
Advertisements

Specialized models and ranking for coreference resolution Pascal Denis ALPAGE Project Team INRIA Rocquencourt F Le Chesnay, France Jason Baldridge.
ADGEN USC/ISI ADGEN: Advanced Generation for Question Answering Kevin Knight and Daniel Marcu USC/Information Sciences Institute.
CS 4705 Discourse Structure and Text Coherence What makes a text/dialogue coherent? Incoherent? “Consider, for example, the difference between passages.
1 Discourse, coherence and anaphora resolution Lecture 16.
Pragmatics II: Discourse structure Ling 571 Fei Xia Week 7: 11/10/05.
Discourse Martin Hassel KTH NADA Royal Institute of Technology Stockholm
Natural Language Processing
INTERNATIONAL CONFERENCE ON NATURAL LANGUAGE PROCESSING NLP-AI IIIT-Hyderabad CIIL, Mysore ICON DECEMBER, 2003.
For Friday No reading Homework –Chapter 23, exercises 1, 13, 14, 19 –Not as bad as it sounds –Do them IN ORDER – do not read ahead here.
Lexical chains for summarization a summary of Silber & McCoy’s work by Keith Trnka.
 Christel Kemke 2007/08 COMP 4060 Natural Language Processing Discourse and Dialogue.
Introduction to RST Rhetorical Structure Theory Maite Taboada and Manfred Stede Simon Fraser University / Universität Potsdam Contact:
Discourse Structure Ling575 Discourse & Dialogue April 13, 2011.
Predicting Text Quality for Scientific Articles Annie Louis University of Pennsylvania Advisor: Ani Nenkova.
Predicting Text Quality for Scientific Articles AAAI/SIGART-11 Doctoral Consortium Annie Louis : Louis A. and Nenkova A Automatically.
Discourse: Structure Ling571 Deep Processing Techniques for NLP March 7, 2011.
CS 4705 Discourse Structure and Text Coherence. What makes a text/dialogue coherent? Incoherent? “Consider, for example, the difference between passages.
Natural Language Generation Martin Hassel KTH CSC Royal Institute of Technology Stockholm
Automatic Classification of Semantic Relations between Facts and Opinions Koji Murakami, Eric Nichols, Junta Mizuno, Yotaro Watanabe, Hayato Goto, Megumi.
Introduction to CL Session 1: 7/08/2011. What is computational linguistics? Processing natural language text by computers  for practical applications.
Approaches to automatic summarization Lecture 5. Types of summaries Extracts – Sentences from the original document are displayed together to form a summary.
Discourse and intertextual issues in translation.
Extracting Opinions, Opinion Holders, and Topics Expressed in Online News Media Text Soo-Min Kim and Eduard Hovy USC Information Sciences Institute 4676.
Query session guided multi- document summarization THESIS PRESENTATION BY TAL BAUMEL ADVISOR: PROF. MICHAEL ELHADAD.
1 Computational Discourse Chapter 21 November 2012 Lecture #15.
Lecture 12: 22/6/1435 Natural language processing Lecturer/ Kawther Abas 363CS – Artificial Intelligence.
Ljubljana, November 18JOTA - Faculty of Arts1 Topic Segmentation Gaël Dias and Elsa Alves Human Language Technology Interest Group Department of Computer.
Interactive Dialogue Systems Professor Diane Litman Computer Science Department & Learning Research and Development Center University of Pittsburgh Pittsburgh,
Intonation and Information Discourse and Dialogue CS359 October 16, 2001.
Discourse Topics, Linguistics, and Language Teaching Richard Watson Todd King Mongkut’s University of Technology Thonburi arts.kmutt.ac.th/crs/research/
Scott Duvall, Brett South, Stéphane Meystre A Hands-on Introduction to Natural Language Processing in Healthcare Annotation as a Central Task for Development.
Developing Reading Skills. Key Reading Skills 1.Selecting what is relevant for the current purpose; 2.Using all the features of the text e.g. headings,
Reading: Comprehension, strategies and activities.
Efficiently Computed Lexical Chains As an Intermediate Representation for Automatic Text Summarization H.G. Silber and K.F. McCoy University of Delaware.
CS774. Markov Random Field : Theory and Application Lecture 19 Kyomin Jung KAIST Nov
1 Learning Sub-structures of Document Semantic Graphs for Document Summarization 1 Jure Leskovec, 1 Marko Grobelnik, 2 Natasa Milic-Frayling 1 Jozef Stefan.
1 Cohesion + Coherence Lecture 9 MODULE 2 Meaning and discourse in English.
ERIKA LUSKY JULIE RAINS Collaborative Dialogue in the Classroom
Indirect Supervision Protocols for Learning in Natural Language Processing II. Learning by Inventing Binary Labels This work is supported by DARPA funding.
PSEUDO-RELEVANCE FEEDBACK FOR MULTIMEDIA RETRIEVAL Seo Seok Jun.
1 Natural Language Processing Chapter Outline Reference –Kinds of reference phenomena –Constraints on co-reference –Preferences for co-reference.
Chapter 23: Probabilistic Language Models April 13, 2004.
1/21 Automatic Discovery of Intentions in Text and its Application to Question Answering (ACL 2005 Student Research Workshop )
Methods for Automatic Evaluation of Sentence Extract Summaries * G.Ravindra +, N.Balakrishnan +, K.R.Ramakrishnan * Supercomputer Education & Research.
For Monday Read chapter 24, sections 1-3 Homework: –Chapter 23, exercise 8.
For Monday Read chapter 26 Last Homework –Chapter 23, exercise 7.
Automatic recognition of discourse relations Lecture 3.
UWMS Data Mining Workshop Content Analysis: Automated Summarizing Prof. Marti Hearst SIMS 202, Lecture 16.
Transitions Bridges between ideas and supporting points.
Finding document topics for improving topic segmentation Source: ACL2007 Authors: Olivier Ferret (18 route du Panorama, BP6) Reporter:Yong-Xiang Chen.
Event-Based Extractive Summarization E. Filatova and V. Hatzivassiloglou Department of Computer Science Columbia University (ACL 2004)
Grammatical and lexical coherence in writing group Done by: O`rinboyeva M. Checked by : RasulovaS.
An evolutionary approach for improving the quality of automatic summaries Constantin Orasan Research Group in Computational Linguistics School of Humanities,
For Monday Read chapter 26 Homework: –Chapter 23, exercises 8 and 9.
Overview of Statistical NLP IR Group Meeting March 7, 2006.
Recognizing Discourse Structure: Text Discourse & Dialogue CMSC October 16, 2006.
Natural Language Processing COMPSCI 423/723 Rohit Kate.
AQUAINT Mid-Year PI Meeting – June 2002 Integrating Robust Semantics, Event Detection, Information Fusion, and Summarization for Multimedia Question Answering.
Discourse: Structure and Coherence Kathy McKeown Thanks to Dan Jurafsky, Diane Litman, Andy Kehler, Jim Martin.
Academic Writing Fatima AlShaikh. A duty that you are assigned to perform or a task that is assigned or undertaken. For example: Research papers (most.
Introduction to RST (Rhetorical Structure Theory)
Content Selection: Supervision & Discourse
Discourse Structure and Text Coherence
Web News Sentence Searching Using Linguistic Graph Similarity
Improving a Pipeline Architecture for Shallow Discourse Parsing
Lecture 17: Discourse: Anaphora Resolution and Coherence
CSCI 5832 Natural Language Processing
CSCI 5832 Natural Language Processing
Text Mining & Natural Language Processing
Presentation transcript:

Discourse: Structure and Coherence Kathy McKeown Thanks to Dan Jurafsky, Diane Litman, Andy Kehler, Jim Martin

 Units for pyramid analysis  Summary length

 What areas would you like to review?  Semantic interpretation?  Probabilistic context free parsing?  Earley Algorithm?  Learning?  Information extraction?  Pronoun resolution?  Machine translation?

 Which are more useful where?  Discourse structure: subtopics  Discourse coherence: relations between sentences  Discourse structure: rhetorical relations Generation vs. Interpretation?

 Discourse Structure ◦ Textiling  Coherence ◦ Hobbs coherence relations ◦ Rhetorical Structure Theory

 Conventional structures for different genres ◦ Academic articles:  Abstract, Introduction, Methodology, Results, Conclusion ◦ Newspaper story:  inverted pyramid structure (lead followed by expansion)

 Simpler task ◦ Discourse segmentation  Separating document into linear sequence of subtopics

 Hearst (1997): 21-pgraph science news article called “Stargazers”  Goal: produce the following subtopic segments:

 Information retrieval:  automatically segmenting a TV news broadcast or a long news story into sequence of stories  Text summarization: ?  Information extraction:  Extract info from inside a single discourse segment  Question Answering?

 Halliday and Hasan (1976): “The use of certain linguistic devices to link or tie together textual units”  Lexical cohesion: ◦ Indicated by relations between words in the two units (identical word, synonym, hypernym)  Before winter I built a chimney, and shingled the sides of my house. I thus have a tight shingled and plastered house.  Peel, core and slice the pears and the apples. Add the fruit to the skillet.

 Non-lexical: anaphora ◦ The Woodhouses were first in consequence there. All looked up to them.  Cohesion chain: ◦ Peel, core and slice the pears and the apples. Add the fruit to the skillet. When they are soft…

 Sentences or paragraphs in a subtopic are cohesive with each other  But not with paragraphs in a neighboring subtopic  Thus if we measured the cohesion between every neighboring sentences ◦ We might expect a ‘dip’ in cohesion at subtopic boundaries.

1. Tokenization ◦ Each space-deliminated word ◦ Converted to lower case ◦ Throw out stop list words ◦ Stem the rest ◦ Group into pseudo-sentences of length w=20 2. Lexical Score Determination: cohesion score 1.Three part score including ◦ Average similarity (cosine measure) between gaps 3. Boundary Identification

 Discourse markers or cue words ◦ Broadcast news  Good evening, I’m  …coming up…. ◦ Science articles  “First,….”  “The next topic….”

 Supervised machine learning ◦ Label segment boundaries in training and test set ◦ Extract features in training ◦ Learn a classifier ◦ In testing, apply features to predict boundaries

 Evaluation: WindowDiff (Pevzner and Hearst 2000) assign partial credit

 Which are more useful where?  Discourse structure: subtopics  Discourse coherence: relations between sentences  Discourse structure: rhetorical relations Generation vs. Interpretation?

What makes a discourse coherent? The reason is that these utterances, when juxtaposed, will not exhibit coherence. Almost certainly not. Do you have a discourse? Assume that you have collected an arbitrary set of well-formed and independently interpretable utterances, for instance, by randomly selecting one sentence from each of the previous chapters of this book.

Assume that you have collected an arbitrary set of well- formed and independently interpretable utterances, for instance, by randomly selecting one sentence from each of the previous chapters of this book. Do you have a discourse? Almost certainly not. The reason is that these utterances, when juxtaposed, will not exhibit coherence.

 John hid Bill’s car keys. He was drunk.  ??John hid Bill’s car keys. He likes spinach.

 Appropriate use of coherence relations between subparts of the discourse -- rhetorical structure  Appropriate sequencing of subparts of the discourse -- discourse/topic structure  Appropriate use of referring expressions

 “ Result ” :  Infer that the state or event asserted by S0 causes or could cause the state or event asserted by S1. ◦ The Tin Woodman was caught in the rain. His joints rusted.

 Infer that the state or event asserted by S1 causes or could cause the state or event asserted by S0. ◦ John hid Bill’s car keys. He was drunk.

 Infer p(a1, a2..) from the assertion of S0 and p(b1,b2…) from the assertion of S1, where ai and bi are similar, for all I. ◦ The Scarecrow wanted some brains. The Tin Woodman wanted a heart.

 Infer the same proposition P from the assertions of S0 and S1. ◦ Dorothy was from Kansas. She lived in the midst of the great Kansas prairies.

 Which are more useful where?  Discourse structure: subtopics  Discourse coherence: relations between sentences  Discourse structure: rhetorical relations Generation vs. Interpretation?

 Another theory of discourse structure, based on identifying relations between segments of the text ◦ Nucleus/satellite notion encodes asymmetry  Nucleus is thing that if you deleted it, text wouldn’t make sense. ◦ Some rhetorical relations:  Elaboration: (set/member, class/instance, whole/part…)  Contrast: multinuclear  Condition: Sat presents precondition for N  Purpose: Sat presents goal of the activity in N

 A sample definition ◦ Relation: Evidence ◦ Constraints on N: H might not believe N as much as S think s/he should ◦ Constraints on Sat: H already believes or will believe Sat ◦ Effect: H’s belief in N is increased  An example: Kevin must be here. His car is parked outside. Nucleus Satellite

 Supervised machine learning ◦ Get a group of annotators to assign a set of RST relations to a text ◦ Extract a set of surface features from the text that might signal the presence of the rhetorical relations in that text ◦ Train a supervised ML system based on the training set

 Explicit markers: because, however, therefore, then, etc.  Tendency of certain syntactic structures to signal certain relations: Infinitives are often used to signal purpose relations: Use rm to delete files.  Ordering  Tense/aspect  Intonation

 How many Rhetorical Relations are there?  How can we use RST in dialogue as well as monologue?  RST does not model overall structure of the discourse.  Difficult to get annotators to agree on labeling the same texts

 Which are more useful where?  Discourse structure: subtopics  Discourse coherence: relations between sentences  Discourse structure: rhetorical relations