Discourse & Natural Language Generation Martin Hassel KTH NADA Royal Institute of Technology 100 44 Stockholm +46-8-790 66 34

Slides:



Advertisements
Similar presentations
Kees van Deemter Matthew Stone Formal Issues in Natural Language Generation Lecture 4 Shieber 1993; van Deemter 2002.
Advertisements

Referring Expressions: Definition Referring expressions are words or phrases, the semantic interpretation of which is a discourse entity (also called referent)
Academic Writing Writing an Abstract.
Specialized models and ranking for coreference resolution Pascal Denis ALPAGE Project Team INRIA Rocquencourt F Le Chesnay, France Jason Baldridge.
Natural Language Understanding Difficulties: Large amount of human knowledge assumed – Context is key. Language is pattern-based. Patterns can restrict.
December 2003CSA3050: Natural Language Generation 1 What is Natural Language Generation? When is NLG an Appropriate Technology? NLG System Architectures.
Statistical NLP: Lecture 3
1 Discourse, coherence and anaphora resolution Lecture 16.
Pragmatics II: Discourse structure Ling 571 Fei Xia Week 7: 11/10/05.
Discourse Martin Hassel KTH NADA Royal Institute of Technology Stockholm
Natural Language Processing
Chapter 18: Discourse Tianjun Fu Ling538 Presentation Nov 30th, 2006.
INTERNATIONAL CONFERENCE ON NATURAL LANGUAGE PROCESSING NLP-AI IIIT-Hyderabad CIIL, Mysore ICON DECEMBER, 2003.
Natural Language Generation: Discourse Planning
Chapter 20: Natural Language Generation Presented by: Anastasia Gorbunova LING538: Computational Linguistics, Fall 2006 Speech and Language Processing.
Natural Language Generation Research Presentation Presenter Shamima Mithun.
 Christel Kemke 2007/08 COMP 4060 Natural Language Processing Discourse and Dialogue.
A Pipelined Architecture Document Planning Microplanning Surface Realisation Document Plan Text Specification.
CS 4705 Algorithms for Reference Resolution. Anaphora resolution Finding in a text all the referring expressions that have one and the same denotation.
Final Review CS4705 Natural Language Processing. Semantics Meaning Representations –Predicate/argument structure and FOPC Thematic roles and selectional.
Natural Language Generation Martin Hassel KTH CSC Royal Institute of Technology Stockholm
PSY 369: Psycholinguistics Some basic linguistic theory part3.
Natural Language Generation Ling 571 Fei Xia Week 8: 11/17/05.
Semi-Automatic Learning of Transfer Rules for Machine Translation of Low-Density Languages Katharina Probst April 5, 2002.
PRAGMATICS. 3- Pragmatics is the study of how more gets communicated than is said. It explores how a great deal of what is unsaid is recognized. 4.
1 Pragmatics: Discourse Analysis J&M’s Chapter 21.
Pragmatics I: Reference resolution Ling 571 Fei Xia Week 7: 11/8/05.
1. Introduction Which rules to describe Form and Function Type versus Token 2 Discourse Grammar Appreciation.
Statistical Natural Language Processing. What is NLP?  Natural Language Processing (NLP), or Computational Linguistics, is concerned with theoretical.
Artificial Intelligence Research Centre Program Systems Institute Russian Academy of Science Pereslavl-Zalessky Russia.
Lecture 1, 7/21/2005Natural Language Processing1 CS60057 Speech &Natural Language Processing Autumn 2005 Lecture 1 21 July 2005.
Introduction to Natural Language Generation
9/8/20151 Natural Language Processing Lecture Notes 1.
Empirical Methods in Information Extraction Claire Cardie Appeared in AI Magazine, 18:4, Summarized by Seong-Bae Park.
Lecture 12: 22/6/1435 Natural language processing Lecturer/ Kawther Abas 363CS – Artificial Intelligence.
THE BIG PICTURE Basic Assumptions Linguistics is the empirical science that studies language (or linguistic behavior) Linguistics proposes theories (models)
Psycholinguistic Theory
Discourse Analysis Radhika Mamidi. Coherence and cohesion What makes a text coherent? What are the coherent devices? Discourses have to have connectivity.
An Intelligent Analyzer and Understander of English Yorick Wilks 1975, ACM.
1 Special Electives of Comp.Linguistics: Processing Anaphoric Expressions Eleni Miltsakaki AUTH Fall 2005-Lecture 2.
1 Cohesion + Coherence Lecture 9 MODULE 2 Meaning and discourse in English.
Artificial Intelligence: Natural Language
Ideas for 100K Word Data Set for Human and Machine Learning Lori Levin Alon Lavie Jaime Carbonell Language Technologies Institute Carnegie Mellon University.
Linguistic Essentials
CSA2050 Introduction to Computational Linguistics Lecture 1 Overview.
Artificial Intelligence Research Center Pereslavl-Zalessky, Russia Program Systems Institute, RAS.
Coherence and Coreference Introduction to Discourse and Dialogue CS 359 October 2, 2001.
LECTURE 2: SEMANTICS IN LINGUISTICS
Introduction to Computational Linguistics
Rules, Movement, Ambiguity
Artificial Intelligence: Natural Language
Introduction to Linguistics Class # 1. What is Linguistics? Linguistics is NOT: Linguistics is NOT:  learning to speak many languages  evaluating different.
Jan 2004CSA3050: NLG21 CSA3050: Natural Language Generation 2 Surface Realisation Systemic Grammar Functional Unification Grammar see J&M Chapter 20.3.
Natural Language Generation Martin Hassel KTH NADA Royal Institute of Technology Stockholm
NLP. Introduction to NLP (U)nderstanding and (G)eneration Language Computer (U) Language (G)
NATURAL LANGUAGE PROCESSING
Discourse: Structure and Coherence Kathy McKeown Thanks to Dan Jurafsky, Diane Litman, Andy Kehler, Jim Martin.
Natural Language Processing COMPSCI 423/723 Rohit Kate.
Discourse Analysis Natural Language Understanding basis
Statistical NLP: Lecture 3
Language, Logic, and Meaning
Referring Expressions: Definition
Advanced English 6 November 1-2, 2017
Algorithms for Reference Resolution
Linguistic Essentials
CS4705 Natural Language Processing
Artificial Intelligence 2004 Speech & Natural Language Processing
Presentation transcript:

Discourse & Natural Language Generation Martin Hassel KTH NADA Royal Institute of Technology Stockholm

2Martin Hassel What is a discourse? The linguistic term for a contextually related group of sentences or utterances Basic discourse types Monologue Dialogue HCI turn taking / ”dialogue”

3Martin Hassel Cohesion and Coherence Cohesion The bond that ties sentences to one another on a textual level Coherence The application of cohesion in order to form a discourse

4Martin Hassel Reference Phenomena 1 Indefinite noun phrases an apple, some lazy people Definite noun phrases the fastest computer Demonstratives this, that One-anaphora

5Martin Hassel Reference Phenomena 2 Inferrables car  engine, door Discontinous sets they, them Generics they

6Martin Hassel Referential Constraints Agreement Number Person and case Gender Syntactic constraints Selectional restrictions

7Martin Hassel Coreferential Expressions Coreference Expressions denoting the same discourse entity corefer Anaphors Refer backwards in the discourse The referent is called the antecedent Cataphors Refer forwards in the discourse Although he loved fishing, Paul went skating with Mary.

8Martin Hassel Pronouns Seldom refer more than two sentences back Requires a salient referent as antecedent Antecedent Indicators: Recency Grammatical role Parallellism Repeated mention Verb semantics

9Martin Hassel Text Coherence Coherence relations Result Explanation Parallel Elaboration Occasion

10Martin Hassel A Discourse Tree S2 (e2) Explanation (e2) Occasion (e1;e2) S1 (e1) Parallel (e3;e5) S3 (e3) S5 (e5) Explanation (e3) S4 (e4)

11Martin Hassel Discourse Structure John went to the bank to deposit his paycheck (S1) He then took a train to Bill’s car dealership (S2) He needed to buy a car (S3) The company he works for now isn’t near any public tranportation (S4) He also wanted to talk to Bill about their softball league (S5)

12Martin Hassel Inference 1 Rule:If it rains the ground gets wet Observation:It rains Conclusion:The ground gets wet Deduction: rule + observation → conclusion (modus ponens) Induktion: observation +conclusion → rule (modus tollens) (Abduktion: rule + conclusion - (?!) → obeservation)

13Martin Hassel Inference 2 John hid Bill’s car keys. He was drunk. 1.  John usually does stupid things when drunk 2.  Bill often drives when drunk Bill was drunk. John hid his car keys. 1.  Bill tends to ”borrow” cars when drunk 2.  Bill often drives his car when drunk

14Martin Hassel Background Knowledge The problem of encoding inference is usually said to AI-complete AI-completeness indicates that the problem requires all of the knowledge – and utilities to utilize it – that humas possess

15Martin Hassel Different Levels Syntax Rules for constructing grammatical sentences Semantics Rules for assigning meaning to statements Pragmatics Rules (of thumb) for applying contextual constraints on the semantics of a statement

16Martin Hassel Pragmatics The study of meaning contained by utterences in situations (Leech, 1983) Relates the content of a clause (semantics) with the content of an utterance of that clause (pragmatics) Pragmatic rules often rules of thumb Dialogues – Cooperative Principles

17Martin Hassel Grice’ Cooperative Principle Quantity – Don’t say more that necessary Quality – Don’t say anything you do not believe in or have proof of Relevance – A response should be an answer to the question Form – Be clear / avoid ambiguity – Be consice – Be methodical

18Martin Hassel Discourse, what for? Information Retrieval Summarization Pronoun Resolution … Natural Language Generation

19Martin Hassel What Is Natural Language Generation? A process of constructing a natural language output from non linguistic inputs that maps meaning to text.

20Martin Hassel Related Simple Text Generation Canned text Ouputs predefined text Template filling Outputs predefined text with predefined variable words/phrases

21Martin Hassel Areas of Use NLG techniques can be used to: generate textual weather forecasts from representations of graphical weather maps summarize statistical data extracted from a database or spreadsheet explain medical info in a patient-friendly way describe a chain of reasoning carried out by an expert system paraphrase information in a diagram for inexperienced users

22Martin Hassel Goals of a NLG System To supply text that is: correct and relevant information non redundant suiting the needs of the user in an understandable form in a correct form

23Martin Hassel Choices for NLG Content selection Lexical selection Sentence structure Aggregation Referring expressions Orthographic realisation Discourse structure

24Martin Hassel Example Architecture Discourse Specification Knowledge Base Communicative Goal Natural Language Output Discourse Planner Surface Realizer

25Martin Hassel Discourse Planner 1. Text shemata Use consistent patterns of discourse structure Used for manuals and descriptive texts 2. Rhetorical Relations Uses the Rhetorical Structure Theory Used for varied generation tasks

26Martin Hassel Discourse Planner – Rhetorical Relations Rhetorical Structure Theory (Mann & Thompson 1988) Nucleus Multi-nuclear Satellite

27Martin Hassel Discourse Planner Rhetorical Relations 23 rhetorical relations, among these: Cause Circumstance Condition Contrast Elaboration Explanation List Occasion Parallel Purpose Result Sequence

28Martin Hassel Surface Realisation 1. Systemic Grammar Using functional categorization Represents sentences as collections of functions Directed, acyclic and/or graph 2. Functional Unification Grammar Using functional categorization Unifies generation grammar with a feature structure

29Martin Hassel Surface Realisation – Systemic Grammar Emphasises the functional organisation of language Surface forms are viewed as the consequences of selecting a set of abstract functional features Choices correspond to minimal grammatical alternatives The interpolation of an intermediate abstract representation allows the specification of the text to accumulate gradually

30Martin Hassel Surface Realisation – Systemic Grammar Present-Participle Past-Participle Infinitive Polar Wh- Imperative Indicative Declarative Interrogative Major Minor Mood … Bound Relative

31Martin Hassel Surface Realisation – Functional Unification Grammar Basic idea: Input specification in the form of a FUNCTIONAL DESCRIPTION, a recursive matrix The grammar is a large functional description with alternations representing choice points Realisation is achieved by unifying the input FD with the grammar FD

32Martin Hassel Surface Realisation – Functional Unification Grammar ((cat clause) (process ((type composite) (relation possessive) (lex ‘hand’))) (participants ((agent ((cat pers_pro) (gender feminine))) ((affected  ((cat np) (lex ‘editor’))) ((possessor  )) ((possessed ((cat np) (lex ‘draft’))))) She hands the draft to the editor.

33Martin Hassel Microplanning 1 Lexical selection Referring expression generation Morphological realization Syntactic realization Orthographic realization

34Martin Hassel Microplanning 3 Aggregation Some possibilities: Simple conjunction Ellipsis Set introduction

35Martin Hassel Aggregation Without aggregation: It has a snack bar. It has a restaurant car. With set introduction : It has {a snack bar, a restaurant car}. It has a snack bar and a restaurant car. Caution! Need to avoid changing the meaning: John bought a TV. Bill bought a TV. ≠John and Bill bought a TV.

36Martin Hassel Further Reading Siggen Allen 1995: Natural Language Understanding cl/Allen95/al1995co.htm