Kees van Deemter Matthew Stone Formal Issues in Natural Language Generation Lecture 5 Stone, Doran,Webber, Bleam & Palmer.

Slides:



Advertisements
Similar presentations
Analysis of Algorithms
Advertisements

Kees van Deemter Matthew Stone Formal Issues in Natural Language Generation Lecture 4 Shieber 1993; van Deemter 2002.
Generation of Referring Expressions: Managing Structural Ambiguities I.H. KhanG. Ritchie K. van Deemter University of Aberdeen, UK.
Some common assumptions behind Computational Generation of Referring Expressions (GRE) (Introductory remarks at the start of the workshop)
Generation of Referring Expressions (GRE) Reading: Dale & Reiter (1995) (key paper in this area)
Conceptual coherence in the generation of referring expressions Albert Gatt & Kees van Deemter University of Aberdeen {agatt,
Generation of Referring Expressions: the State of the Art SELLC Summer School, Harbin 2010 Kees van Deemter Computing Science University of Aberdeen.
Kees van Deemter Matthew Stone Formal Issues in Natural Language Generation Lecture 1 Overview + Reiter & Dale 1997.
Heuristic Search techniques
Lecture 15. Graph Algorithms
CS4018 Formal Models of Computation weeks Computability and Complexity Kees van Deemter (partly based on lecture notes by Dirk Nikodem)
Computer Science CPSC 322 Lecture 25 Top Down Proof Procedure (Ch 5.2.2)
ARCHITECTURES FOR ARTIFICIAL INTELLIGENCE SYSTEMS
Problems and Their Classes
NP-Hard Nattee Niparnan.
1 Programming Languages (CS 550) Mini Language Interpreter Jeremy R. Johnson.
CLASSICAL PLANNING What is planning ?  Planning is an AI approach to control  It is deliberation about actions  Key ideas  We have a model of the.
Natural Language Understanding Difficulties: Large amount of human knowledge assumed – Context is key. Language is pattern-based. Patterns can restrict.
C O N T E X T - F R E E LANGUAGES ( use a grammar to describe a language) 1.
Logic Use mathematical deduction to derive new knowledge.
CSE 425: Semantic Analysis Semantic Analysis Allows rigorous specification of a program’s meaning –Lets (parts of) programming languages be proven correct.
LTAG Semantics on the Derivation Tree Presented by Maria I. Tchalakova.
Junction Trees: Motivation Standard algorithms (e.g., variable elimination) are inefficient if the undirected graph underlying the Bayes Net contains cycles.
CPSC 322, Lecture 19Slide 1 Propositional Logic Intro, Syntax Computer Science cpsc322, Lecture 19 (Textbook Chpt ) February, 23, 2009.
1 Introduction to Computability Theory Lecture12: Decidable Languages Prof. Amos Israeli.
Xyleme A Dynamic Warehouse for XML Data of the Web.
Constraint Logic Programming Ryan Kinworthy. Overview Introduction Logic Programming LP as a constraint programming language Constraint Logic Programming.
LEARNING FROM OBSERVATIONS Yılmaz KILIÇASLAN. Definition Learning takes place as the agent observes its interactions with the world and its own decision-making.
Analysis of Algorithms CS 477/677
Semantics with Applications Mooly Sagiv Schrirber html:// Textbooks:Winskel The.
Chapter 5 Outline Formal definition of CSP CSP Examples
Lecture 1, 7/21/2005Natural Language Processing1 CS60057 Speech &Natural Language Processing Autumn 2005 Lecture 1 21 July 2005.
Lecture 12: 22/6/1435 Natural language processing Lecturer/ Kawther Abas 363CS – Artificial Intelligence.
EECS 203: It’s the end of the class and I feel fine. Graphs.
Logics for Data and Knowledge Representation
Nattee Niparnan. Easy & Hard Problem What is “difficulty” of problem? Difficult for computer scientist to derive algorithm for the problem? Difficult.
THE BIG PICTURE Basic Assumptions Linguistics is the empirical science that studies language (or linguistic behavior) Linguistics proposes theories (models)
RELATIONAL FAULT TOLERANT INTERFACE TO HETEROGENEOUS DISTRIBUTED DATABASES Prof. Osama Abulnaja Afraa Khalifah
CSE 326: Data Structures NP Completeness Ben Lerner Summer 2007.
Formal Semantics Chapter Twenty-ThreeModern Programming Languages, 2nd ed.1.
CSE332: Data Abstractions Lecture 24.5: Interlude on Intractability Dan Grossman Spring 2012.
Politeness & Speaking Style Discourse & Dialogue CS 359 November 15, 2001.
Semantic Construction lecture 2. Semantic Construction Is there a systematic way of constructing semantic representation from a sentence of English? This.
Albert Gatt LIN3021 Formal Semantics Lecture 4. In this lecture Compositionality in Natural Langauge revisited: The role of types The typed lambda calculus.
LECTURE 2: SEMANTICS IN LINGUISTICS
Artificial Intelligence: Natural Language
XML Access Control Koukis Dimitris Padeleris Pashalis.
Automated Reasoning Early AI explored how to automated several reasoning tasks – these were solved by what we might call weak problem solving methods as.
SPANNING TREES Lecture 21 CS2110 – Fall Nate Foster is out of town. NO 3-4pm office hours today!
CS6045: Advanced Algorithms NP Completeness. NP-Completeness Some problems are intractable: as they grow large, we are unable to solve them in reasonable.
LDK R Logics for Data and Knowledge Representation Propositional Logic Originally by Alessandro Agostini and Fausto Giunchiglia Modified by Fausto Giunchiglia,
JavaScript Introduction and Background. 2 Web languages Three formal languages HTML JavaScript CSS Three different tasks Document description Client-side.
EEL 5937 Content languages EEL 5937 Multi Agent Systems Lecture 10, Feb. 6, 2003 Lotzi Bölöni.
Understanding Naturally Conveyed Explanations of Device Behavior Michael Oltmans and Randall Davis MIT Artificial Intelligence Lab.
NATURAL LANGUAGE PROCESSING
Ch3 /Lecture #4 Brute Force and Exhaustive Search 1.
CMSC 104, L041 Algorithms, Part 1 of 3 Topics Definition of an Algorithm Example: The Euclidean Algorithm Syntax versus Semantics Reading Sections 3.1.
Artificial Intelligence Logical Agents Chapter 7.
Business Process Modeling What is a process model? – A formal way of representing how a business system operates. – Illustrates the activities that are.
SYNTAX.
P & NP.
Algorithms, Part 1 of 3 The First step in the programming process
CSPs: Search and Arc Consistency Computer Science cpsc322, Lecture 12
CSPs: Search and Arc Consistency Computer Science cpsc322, Lecture 12
Language, Logic, and Meaning
2008/09/17: Lecture 4 CMSC 104, Section 0101 John Y. Park
Lecture 7: Introduction to Parsing (Syntax Analysis)
Semantic Markup for Semantic Web Tools:
Logics for Data and Knowledge Representation
Presentation transcript:

Kees van Deemter Matthew Stone Formal Issues in Natural Language Generation Lecture 5 Stone, Doran,Webber, Bleam & Palmer

GRE and surface realization Arguably, GRE uses a grammar. –Parameters such as the preference order on properties reflect knowledge of how to communicate effectively. –Decisions about usefulness or completeness of a referring expression reflect beliefs about utterance interpretation. Maybe this is a good idea for NLG generally.

GRE and surface realization But weve thought GRE outputs semantics: referent: furniture886 type: desk status: definite color: brown origin: sweden

GRE and surface realization We also need to link this up with surface form: the brown Swedish desk Note: not ?the Swedish brown desk

Todays initial observations Its hard to do realization on its own mapping from semantics to surface structure. Its easy to combine GRE and realization because GRE is grammatical reasoning! if you have a good representation for syntax.

Why its hard to do realization A pathological grammar of adjective order: NP the N(w). N(w) w N(w ) if w is an adjective and wRw. N(w) w if w is a noun.

Syntax with this grammar Derivation of example: the brown Swedish desk NP N(brown) N(Swedish) N(desk) Requires: brown R Swedish, Swedish R desk

Realization, formally You start with k properties. Each property can be realized lexically. assume: one noun, many adjectives (not that its easy to enforce this) Realization solution: NP which realizes each property exactly once.

Quick formal analysis View problem graph-theoretically: k words, corresponding to vertices in a graph R is a graph on the k words Surface structure is a Hamiltonian path (which visits each vertex exactly once) through R. This is a famous NP complete problem So surface realization itself is intractable!

Moral of the example Semantics underdetermines syntactic relations. Here, semantics underdetermines syntactic relations of adjectives to one another and to the head. Searching for the correspondence is hard. See also Brew 92, Koller and Striegnitz 02.

Todays initial observations Its hard to do realization on its own mapping from semantics to surface structure. Its easy to combine GRE and realization because GRE is grammatical reasoning! if you have a good representation for syntax.

Syntactic processing for GRE Lexicalization Steps of grammatical derivation correspond to meaningful choices in NLG. E.g., steps of grammar are synched with steps of adding a property to a description.

Syntactic processing for GRE Key ideas: lexicalization, plus Flat dependency structure (adjs modify noun) Hierarchical representation of word-order NP N(color) N(origin) N(size) N(material) thedesk

Syntactic processing for GRE Other syntactic lexical entries Adj N(origin) Swedish N(color) Adj brown

Describing syntactic combination Operation of combination 1: Substitution NP + = NP N(color) N(origin) N(size) N(material) thedesk NP N(color) N(origin) N(size) N(material) thedesk

Describing syntactic combination Operation of combination 2: Sister adjunction + = NP N(color) N(origin) N(size) N(material) thedesk NP N(color) N(origin) N(size) N(material) thedesk N(color) Adj brown Adj brown

Abstracting syntax Tree rewriting: Each lexical item is associated with a structure. You have a starting structure. You have ways of combining two structures together.

Abstracting syntax Derivation tree records elements and how they are combined the desk brown color) Swedish origin)

An extended incremental algorithm r = individual to be described P = lexicon of entries, in preference order P is an individual entry sem(P) is a property or set of entries from the context syn(P) is a syntactic element L = surface syntax of description

Extended incremental algorithm L := NP C := Domain For each P P do: If r sem(P) & C sem(P) Then do L := add(syn(P), L) C := C sem(P) If C = {r} then return L Return failure

Observations Why use tree-rewriting - not,e.g. CFG derivation? NP the N(w). N(w) w N(w ) if w is an adjective and wRw. N(w) w if w is a noun. CFG derivation forces you to select properties in the surface word-order.

Observations Tree-rewriting frees word-order from choice- order. NP N(color) N(origin) N(size) N(material) the desk NP N(color) N(origin) N(size) N(material) the desk Adj brown NP N(color) N(origin) N(size) N(material) the desk Adj brown Adj Swedish

Observations Tree-rewriting frees word-order from choice- order. NP N(color) N(origin) N(size) N(material) the desk NP N(color) N(size) the NP N(color) N(origin) N(size) N(material) the desk Adj brown Adj Swedish N(origin) N(material) desk Adj Swedish

This is reflected in derivation tree Derivation tree records elements and how they are combined the desk brown color) Swedish origin)

Formal results Logical completeness. If theres a flat derivation tree for an NP that identifies referent r, Then the incremental algorithm finds it. But Sensible combinations of properties may not yield surface NPs. Hierarchical derivation trees may require lookahead in usefulness check.

Formal results Computational complexity Nothing changes – we just add properties, one after another…

Now, though, were choosing specific lexical entries NP N(departure) N(destination) N(stops) the express Adj 3:35 N Trenton vs NP N(departure) N(destination) N(stops) the express Adj 15:35 N Trenton maybe these lexical items express the same property…

Use in 12-hour time context Use in 24-hour time context What motivates these choices? N(departure) Adj 3:35 N(departure) Adj 15:35

P = lexicon of entries, in preference order P is an individual entry sem(P) is a property or set of entries from the context syn(P) is a syntactic element prags(P) is a test which the context must satisfy for the entry to be appropriate Need to extend grammar again

For example: syn: sem: departure(x, 1535) prags: twentyfourhourtime Need to extend grammar again N(departure) Adj 15:35

Extended incremental algorithm L := NP C := Domain For each P P do: If r sem(P) & C sem(P) & prags(P) is true Then do L := add(syn(P), L) C := C sem(P) If C = {r} then return L Return failure

Discussion: What does this entry do? syn: sem: thing(x) prags: in-focus(x) NP it

Suggestion: find best value Given: –A set of entries that combine syntactically with L in the same way –Related by semantic generality and pragmatic specificity. –Current distractors Take entries that remove the most distractors Of those, take the most semantically general Of those, take the most pragmatically specific

Extended incremental algorithm L := NP C := Domain Repeat Choices := { P : add(syn(P), L) at next node & r sem(P) & prags(P) is true } P := find best value(Choices) L := add(syn(P), L) C := C sem(P) If C = {r} then return L Return failure

What is generation anyway? Generation is intentional (or rational) action thats why Grices maxims apply, for example. You have a goal You build a plan to achieve it (& achieve it economically in a recognizable way) You carry out the plan

In GRE… The goal is for hearer to know the identity of r (in general g) The plan will be to utter some NP U such that the interpretation of U identifies { r } (in general c u c g) Carrying out the plan means realizing this utterance.

In other words GRE amounts to a process of deliberation. Adding a property to L incrementally is like committing to an action. These commitments are called intentions. Incrementality is characteristic of intentions – though in general intentions are open to revision. Note: this connects with belief-desire- intention models of bounded rationality.

GRE as (BDI) rational agency L := NP // Initial plan C := Domain// Interpretation while (P := FindBest(P, C, L)) { // Deliberation L := add(syn(P), L)// Adopt new intention C := C sem(P)// Update interpretation if C = { r } return L// Goal satisfied } fail

NLG as (BDI) rational agency L := X C := Initial Interpretation while (P := FindBest(P, C, L)) { L := AddSyntax(syn(P), L) C := AddInterpretation(sem(P), C) if GoalSatisfied(C) return L } fail

Conclusions for NLG researchers Its worth asking (and answering) formal questions about NLG. Questions of logical completeness – can a generator express everything it ought to? Questions of computational complexity – is the cost of a generation algorithm worth the results?

Conclusions for linguists NLG offers a precise perspective on questions of language use. For example, whats the best way of communicating some message? NLG – as opposed to other perspectives – gives more complete, smaller-scale models.

Conclusions for AI in general NLG does force us to characterize and implement representations & inference for practical interactive systems Good motivation for computational semantics. Meaty problems like logical form equivalence. Many connections and possibilities for implementation (graphs, CSPs, circuit optimization, data mining,…)

Open Problems Sets and salience in REs. Generating parallel REs. Theoretical and empirical measures of quality/utility for REs. Avoiding ambiguity in REs. Any problem in RE generalizes to one in NLG.

Followup information Course web page: Kees.van.Deemter/esslli-notes.html –downloadable papers –final lecture notes –papers weve talked about –links (recent/upcoming events, siggen, sigsem) by Monday August 26.