CSCI 5832 Natural Language Processing

Slides:



Advertisements
Similar presentations
Discourse Structure and Discourse Coherence
Advertisements

Referring Expressions: Definition Referring expressions are words or phrases, the semantic interpretation of which is a discourse entity (also called referent)
Specialized models and ranking for coreference resolution Pascal Denis ALPAGE Project Team INRIA Rocquencourt F Le Chesnay, France Jason Baldridge.
Unit 6 Predicates, Referring Expressions, and Universe of Discourse Part 1: Practices 1-7.
Computational Models of Discourse Analysis Carolyn Penstein Rosé Language Technologies Institute/ Human-Computer Interaction Institute.
CS 4705 Discourse Structure and Text Coherence What makes a text/dialogue coherent? Incoherent? “Consider, for example, the difference between passages.
1 Discourse, coherence and anaphora resolution Lecture 16.
Discourse Martin Hassel KTH NADA Royal Institute of Technology Stockholm
Chapter 18: Discourse Tianjun Fu Ling538 Presentation Nov 30th, 2006.
INTERNATIONAL CONFERENCE ON NATURAL LANGUAGE PROCESSING NLP-AI IIIT-Hyderabad CIIL, Mysore ICON DECEMBER, 2003.
For Friday No reading Homework –Chapter 23, exercises 1, 13, 14, 19 –Not as bad as it sounds –Do them IN ORDER – do not read ahead here.
CIS630 Spring 2013 Lecture 2 Affect analysis in text and speech.
Chapter 20: Natural Language Generation Presented by: Anastasia Gorbunova LING538: Computational Linguistics, Fall 2006 Speech and Language Processing.
Generation of Referring Expressions: Modeling Partner Effects Surabhi Gupta Advisor: Amanda Stent Department of Computer Science.
CS 4705 Discourse Structure and Text Coherence. What makes a text/dialogue coherent? Incoherent? “Consider, for example, the difference between passages.
Natural Language Generation Martin Hassel KTH CSC Royal Institute of Technology Stockholm
Language, Mind, and Brain by Ewa Dabrowska Chapter 2: Language processing: speed and flexibility.
1 Introduction to Computational Linguistics Eleni Miltsakaki AUTH Spring 2006-Lecture 7.
The classification problem (Recap from LING570) LING 572 Fei Xia, Dan Jinguji Week 1: 1/10/08 1.
LIN1180/LIN5082 Semantics Lecture 3
What you should be able to do upon successful completion of ENGL 101: Understand a writer’s thesis. Assess the validity of a writer’s central claims. Understand.
Lecture 12: 22/6/1435 Natural language processing Lecturer/ Kawther Abas 363CS – Artificial Intelligence.
Scott Duvall, Brett South, Stéphane Meystre A Hands-on Introduction to Natural Language Processing in Healthcare Annotation as a Central Task for Development.
1 Statistical NLP: Lecture 9 Word Sense Disambiguation.
1 CS546: Machine Learning and Natural Language Discriminative vs Generative Classifiers This lecture is based on (Ng & Jordan, 02) paper and some slides.
Reading. This part of the TOEFL tests your ability to read academic English. This is important if you want to succeed at an English college or university.
CS774. Markov Random Field : Theory and Application Lecture 19 Kyomin Jung KAIST Nov
CSA2050 Introduction to Computational Linguistics Lecture 1 Overview.
1 Natural Language Processing Chapter Outline Reference –Kinds of reference phenomena –Constraints on co-reference –Preferences for co-reference.
Programming Languages and Design Lecture 3 Semantic Specifications of Programming Languages Instructor: Li Ma Department of Computer Science Texas Southern.
For Monday Read chapter 24, sections 1-3 Homework: –Chapter 23, exercise 8.
For Friday Finish chapter 24 No written homework.
For Monday Read chapter 26 Last Homework –Chapter 23, exercise 7.
Recognizing Stances in Online Debates Unsupervised opinion analysis method for debate-side classification. Mine the web to learn associations that are.
For Friday Finish chapter 23 Homework –Chapter 23, exercise 15.
SYNTAX.
SEMANTICS Referring Expression.
An evolutionary approach for improving the quality of automatic summaries Constantin Orasan Research Group in Computational Linguistics School of Humanities,
For Monday Read chapter 26 Homework: –Chapter 23, exercises 8 and 9.
Discourse: Structure and Coherence Kathy McKeown Thanks to Dan Jurafsky, Diane Litman, Andy Kehler, Jim Martin.
Discourse: Structure and Coherence Kathy McKeown Thanks to Dan Jurafsky, Diane Litman, Andy Kehler, Jim Martin.
Computational Models of Discourse Analysis Carolyn Penstein Rosé Language Technologies Institute/ Human-Computer Interaction Institute.
Intro to NLP - J. Eisner1 Splitting Words a.k.a. “Word Sense Disambiguation”
Step 1: Specify a null hypothesis
Discourse Structure and Text Coherence
Discourse Analysis Natural Language Understanding basis
Lecture 3: Functional Phrases
Erasmus University Rotterdam
Improving a Pipeline Architecture for Shallow Discourse Parsing
Part I: Basics and Constituency
Referring Expressions: Definition
CSCI 5832 Natural Language Processing
Social Knowledge Mining
Clustering Algorithms for Noun Phrase Coreference Resolution
Lecture 17: Discourse: Anaphora Resolution and Coherence
Statistical NLP: Lecture 9
CSCI 5832 Natural Language Processing
CSCI 5832 Natural Language Processing
CSCI 5832 Natural Language Processing
CSCI 5832 Natural Language Processing
CSCI 5832 Natural Language Processing
Text Mining & Natural Language Processing
CSCI 5832 Natural Language Processing
Pragmatics: Reference and inference
Introduction to Computational Linguistics
Narrative Writing Mr. Blaszak.
Information Retrieval
Statistical NLP : Lecture 9 Word Sense Disambiguation
Presentation transcript:

CSCI 5832 Natural Language Processing Lecture 24 Jim Martin 1/2/2019 CSCI 5832 Spring 2007

Today: 4/19 Discourse structure 1/2/2019 CSCI 5832 Spring 2007

What Makes a Discourse Coherent? The reason is that these utterances, when juxtaposed, will not exhibit coherence. Almost certainly not. Do you have a discourse? Assume that you have collected an arbitrary set of well-formed and independently interpretable utterances, for instance, by randomly selecting one sentence from each of the previous chapters of this book. 1/2/2019 CSCI 5832 Spring 2007

Better? Assume that you have collected an arbitrary set of well-formed and independently interpretable utterances, for instance, by randomly selecting one sentence from each of the previous chapters of this book. Do you have a discourse? Almost certainly not. The reason is that these utterances, when juxtaposed, will not exhibit coherence. 1/2/2019 CSCI 5832 Spring 2007

What makes a text coherent? Appropriate use of coherence relations between subparts of the discourse -- rhetorical structure Appropriate sequencing of subparts of the discourse -- discourse/topic structure Appropriate use of referring expressions 1/2/2019 CSCI 5832 Spring 2007

Rhetorical Structure Theory One theory of discourse structure, based on identifying relations between segments of the text Nucleus/satellite notion encodes asymmetry Some rhetorical relations: Elaboration (set/member, class/instance, whole/part…) Contrast: multinuclear Condition: Sat presents precondition for N Purpose: Sat presents goal of the activity in N 1/2/2019 CSCI 5832 Spring 2007

Relations A sample definition An example: Relation: evidence Constraints on N: H might not believe N as much as S think s/he should Constraints on Sat: H already believes or will believe Sat An example: The governor supports big business. He is sure to veto House Bill 1711. 1/2/2019 CSCI 5832 Spring 2007

Automatic Rhetorical Structure Labeling Same old story by now… Get a group of annotators to assign a set of RST relations to a text Extract a set of surface features from the text that might signal the presence of the rhetorical relations in that text Train a supervised ML system based on the training set 1/2/2019 CSCI 5832 Spring 2007

Features Explicit markers: because, however, therefore, then, etc. Tendency of certain syntactic structures to signal certain relations: Infinitives are often used to signal purpose relations: Use rm to delete files. Ordering Tense/aspect Intonation 1/2/2019 CSCI 5832 Spring 2007

Some Problems with RST How many Rhetorical Relations are there? How can we use RST in dialogue as well as monologue? RST does not model overall structure of the discourse. Difficult to get annotators to agree on labeling the same texts 1/2/2019 CSCI 5832 Spring 2007

Break Final project format. Full deliverable varies based on your project But all projects will need to produce an 8-page conference style paper. I’ll provide LaTex and MS Word templates for you to use from the ACM. This paper should be self-contained. As with a program committee, everything needed to assess the project should be in the paper. 1/2/2019 CSCI 5832 Spring 2007

Referring Expressions Referring expressions provide an additional kind of glue that makes texts cohere. 1/2/2019 CSCI 5832 Spring 2007

Referring Expressions: Definition Referring expressions are words or phrases, the semantic interpretation of which is a discourse entity (also called referent) Discourse entities are semantic objects and they can have multiple syntactic realizations within a text 1/2/2019 CSCI 5832 Spring 2007

NY Times Example A college student accused of faking her own kidnapping last month was charged Wednesday with lying to police in what they suggested was a desperate attempt to get her boyfriend's attention. Audrey Seiler, a 20-year-old sophomore at the University of Wisconsin, was charged with two misdemeanor counts of obstructing officers. Each charge carries up to nine months in jail and a $10,000 fine. Seiler disappeared from her off-campus apartment March 27 without her coat or purse. She was discovered curled in a fetal position in a marsh four days later, and told police that a man had abducted her at knifepoint. But police concluded Seiler made up the story after obtaining a store videotape that showed her buying the knife, duct tape, rope and cold medicine she claimed her abductor used to restrain her. Seiler confessed after she was confronted with the tape, according to authorities. 1/2/2019 CSCI 5832 Spring 2007

Referring Expressions: Example A pretty woman entered the restaurant. She sat at the table next to mine and only then I recognized her. This was Amy Garcia, my next door neighbor from 10 years ago. The woman has totally changed! Amy was at the time shy… 1/2/2019 CSCI 5832 Spring 2007

Pronouns vs. Full NPs A pretty woman entered the restaurant. She sat at the table next to mine and only then I recognized her. This was Amy Garcia, my next door neighbor from 10 years ago. The woman has totally changed! Amy was at the time shy… 1/2/2019 CSCI 5832 Spring 2007

Definite vs. Indefinite NPs A pretty woman entered the restaurant. She sat at the table next to mine and only then I recognized her. This was Amy Garcia, my next door neighbor from 10 years ago. The woman has totally changed! Amy was at the time shy… 1/2/2019 CSCI 5832 Spring 2007

Common Noun vs. Proper Noun A pretty woman entered the restaurant. She sat at the table next to mine and only then I recognized her. This was Amy Garcia, my next door neighbor from 10 years ago. The woman has totally changed! Amy was at the time shy… 1/2/2019 CSCI 5832 Spring 2007

Modified vs. Bare head NP A pretty woman entered the restaurant. She sat at the table next to mine and only then I recognized her. This was Amy Garcia, my next door neighbor from 10 years ago. The woman has totally changed! Amy was at the time shy… 1/2/2019 CSCI 5832 Spring 2007

Premodified vs. postmodified A pretty woman entered the restaurant. She sat at the table next to mine and only then I recognized her. This was Amy Garcia, my next door neighbor from 10 years ago. The woman has totally changed! Amy was at the time shy… 1/2/2019 CSCI 5832 Spring 2007

More NP types Inferrables Discontinuous sets Sally bought a used car. The tires need to be replaced. Discontinuous sets John has known Bill for many years now. They often go hiking together. 1/2/2019 CSCI 5832 Spring 2007

Anaphora resolution Finding in a text all the referring expressions that have one and the same denotation Pronominal anaphora resolution Anaphora resolution between named entities Full noun phrase anaphora resolution Zero anaphora detection/resolution 1/2/2019 CSCI 5832 Spring 2007

Automatic Anaphora Resolution Resolving the referents of pronouns Hand built rule systems vs. trained systems Statistical systems vs rules Systems that require a full syntactic parse vs. systems that assume a weaker chunking segmentation Systems that deal with text vs. systems that deal with spoken discourse 1/2/2019 CSCI 5832 Spring 2007

Full Co-reference Analysis Locating all the entity mentions in a text and then clustering them into groups that represent the actual entities. 1/2/2019 CSCI 5832 Spring 2007

NY Times Example A college student accused of faking her own kidnapping last month was charged Wednesday with lying to police in what they suggested was a desperate attempt to get her boyfriend's attention. Audrey Seiler, a 20-year-old sophomore at the University of Wisconsin, was charged with two misdemeanor counts of obstructing officers. Each charge carries up to nine months in jail and a $10,000 fine. Seiler disappeared from her off-campus apartment March 27 without her coat or purse. She was discovered curled in a fetal position in a marsh four days later, and told police that a man had abducted her at knifepoint. But police concluded Seiler made up the story after obtaining a store videotape that showed her buying the knife, duct tape, rope and cold medicine she claimed her abductor used to restrain her. Seiler confessed after she was confronted with the tape, according to authorities. 1/2/2019 CSCI 5832 Spring 2007

NY Times Example A college student accused of faking her own kidnapping last month was charged Wednesday with lying to police in what they suggested was a desperate attempt to get her boyfriend's attention. Audrey Seiler, a 20-year-old sophomore at the University of Wisconsin, was charged with two misdemeanor counts of obstructing officers. Each charge carries up to nine months in jail and a $10,000 fine. Seiler disappeared from her off-campus apartment March 27 without her coat or purse. She was discovered curled in a fetal position in a marsh four days later, and told police that a man had abducted her at knifepoint. But police concluded Seiler made up the story after obtaining a store videotape that showed her buying the knife, duct tape, rope and cold medicine she claimed her abductor used to restrain her. Seiler confessed after she was confronted with the tape, according to authorities. 1/2/2019 CSCI 5832 Spring 2007

Example There are 33 entity mentions 16 actual entities Seiler accounts for 12 of the 33 mentions, with the police in second at 6 There are 11 entities with 1 mention each 1/2/2019 CSCI 5832 Spring 2007

Basic ML Approaches Scan the text from the top. When a new entity mention is encountered apply a binary classifier to each previous mention (refers or not). Result is a set of chains of references Extract all entity mentions and represent them as feature vectors. Apply an unsupervised partition-based clustering technique to all the mentions. Result is a set of (soft) clusters that partition the mentions. 1/2/2019 CSCI 5832 Spring 2007

Cross-Document Co-Reference Take a set of documents (possibly from the Web), extract all the named referents and resolve them into unique entities. Ie. Google “Michael Jordan” and resolve the results into classes based on the real entities. 1/2/2019 CSCI 5832 Spring 2007

Easy Money 1/2/2019 CSCI 5832 Spring 2007

Easy Money 100,000 files provided for development Set of target names of interest Ground truth provided with a F-measure based scoring metric Precision/Recall computed on a pairwise document basis. Assumes that target names are unique within each document (ie. All the Michael Jordan mentions in a single document are the same person). 1/2/2019 CSCI 5832 Spring 2007

Next Time Machine Translation Read Chapter 24 1/2/2019 CSCI 5832 Spring 2007