Generation of Referring Expressions: Managing Structural Ambiguities I.H. KhanG. Ritchie K. van Deemter University of Aberdeen, UK.

Slides:



Advertisements
Similar presentations
Kees van Deemter Matthew Stone Formal Issues in Natural Language Generation Lecture 4 Shieber 1993; van Deemter 2002.
Advertisements

Some common assumptions behind Computational Generation of Referring Expressions (GRE) (Introductory remarks at the start of the workshop)
Kees van Deemter Matthew Stone Formal Issues in Natural Language Generation Lecture 5 Stone, Doran,Webber, Bleam & Palmer.
Conceptual coherence in the generation of referring expressions Albert Gatt & Kees van Deemter University of Aberdeen {agatt,
Generation of Referring Expressions: the State of the Art SELLC Summer School, Harbin 2010 Kees van Deemter Computing Science University of Aberdeen.
Charting the Potential of Description Logic for the Generation of Referring Expression SELLC, Guangzhou, Dec Yuan Ren, Kees van Deemter and Jeff.
Generation of Referring Expressions: the State of the Art SELLC Winter School, Guangzhou 2010 Kees van Deemter Computing Science University of Aberdeen.
A small taste of inferential statistics
Microplanning (Sentence planning) Part 1 Kees van Deemter.
School of something FACULTY OF OTHER School of Computing FACULTY OF ENGINEERING Chunking: Shallow Parsing Eric Atwell, Language Research Group.
CS4018 Formal Models of Computation weeks Computability and Complexity Kees van Deemter (partly based on lecture notes by Dirk Nikodem)
Literacy Test Preparation
Level 1 Recall Recall of a fact, information, or procedure. Level 2 Skill/Concept Use information or conceptual knowledge, two or more steps, etc. Level.
Statistical Machine Translation Part II: Word Alignments and EM Alexander Fraser ICL, U. Heidelberg CIS, LMU München Statistical Machine Translation.
Dr. Ehud Reiter, Computing Science, University of Aberdeen1 NLG Shared Tasks: Lets try it and see what happens Ehud Reiter (Univ of Aberdeen)
Psych 156A/ Ling 150: Acquisition of Language II Lecture 12 Poverty of the Stimulus I.
Describing Process Specifications and Structured Decisions Systems Analysis and Design, 7e Kendall & Kendall 9 © 2008 Pearson Prentice Hall.
Using the Crosscutting Concepts As conceptual tools when meeting an unfamiliar problem or phenomenon.
Chapter 1 What is Science
Thinking Maps for Reading Comprehension
Chapter 10 Algorithmic Thinking. Copyright © 2013 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Learning Objectives List the five essential.
Automatic Metaphor Interpretation as a Paraphrasing Task Ekaterina Shutova Computer Lab, University of Cambridge NAACL 2010.
ACH: Analysis of Competing Hypotheses Presented by: Jingshan Huang.
The Game of Algebra or The Other Side of Arithmetic The Game of Algebra or The Other Side of Arithmetic © 2007 Herbert I. Gross By Herbert I. Gross & Richard.
Generation of Referring Expressions: Modeling Partner Effects Surabhi Gupta Advisor: Amanda Stent Department of Computer Science.
Bayesian Learning Rong Jin. Outline MAP learning vs. ML learning Minimum description length principle Bayes optimal classifier Bagging.
Lecture 2 Research Questions: Defining and Justifying Problems; Defining Hypotheses.
Chapter 2: Algorithm Discovery and Design
Phonetics, Phonology, Morphology and Syntax
Outline What is a collocation? Automatic approaches 1: frequency-based methods Automatic approaches 2: ruling out the null hypothesis, t-test Automatic.
Albert Gatt LIN 3098 Corpus Linguistics. In this lecture Some more on corpora and grammar Construction Grammar as a theoretical framework Collostructional.
Broca’s Aphasia Paul Broca Language and the Brain First connections drawn:
Empirical Methods in Information Extraction Claire Cardie Appeared in AI Magazine, 18:4, Summarized by Seong-Bae Park.
1 Statistical NLP: Lecture 10 Lexical Acquisition.
Chapter 2: Algorithm Discovery and Design Invitation to Computer Science, C++ Version, Third Edition.
Writing Tips: The Word “Prove” Generally, practicing scientists refrain from using the word prove and its variations (proof, proven, etc) –“Prove” is avoided.
Summary-Response Essay Responding to Reading. Reading Critically Not about finding fault with author Rather engaging author in a discussion by asking.
LOGIC AND ONTOLOGY Both logic and ontology are important areas of philosophy covering large, diverse, and active research projects. These two areas overlap.
Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Developing and Evaluating Theories of Behavior.
Noun-Phrase Analysis in Unrestricted Text for Information Retrieval David A. Evans, Chengxiang Zhai Laboratory for Computational Linguistics, CMU 34 th.
© 2008 The McGraw-Hill Companies, Inc. Chapter 8: Cognition and Language.
1 Statistical NLP: Lecture 7 Collocations. 2 Introduction 4 Collocations are characterized by limited compositionality. 4 Large overlap between the concepts.
1 The Theoretical Framework. A theoretical framework is similar to the frame of the house. Just as the foundation supports a house, a theoretical framework.
Uncertainty Management in Rule-based Expert Systems
Basic Parsing Algorithms: Earley Parser and Left Corner Parsing
Scientific Methods and Terminology. Scientific methods are The most reliable means to ensure that experiments produce reliable information in response.
1 Universidad de Buenos Aires Maestría en Data Mining y Knowledge Discovery Aprendizaje Automático 5-Inducción de árboles de decisión (2/2) Eduardo Poggi.
Creating Subjective and Objective Sentence Classifier from Unannotated Texts Janyce Wiebe and Ellen Riloff Department of Computer Science University of.
Corpus-based evaluation of Referring Expression Generation Albert Gatt Ielka van der Sluis Kees van Deemter Department of Computing Science University.
Jette Viethen 20 April 2007NLGeval07 Automatic Evaluation of Referring Expression Generation is Possible.
BY DR. HAMZA ABDULGHANI MBBS,DPHC,ABFM,FRCGP (UK), Diploma MedED(UK) Associate Professor DEPT. OF MEDICAL EDUCATION COLLEGE OF MEDICINE June 2012 Writing.
BLHC4032 CRITICAL AND CREATIVE THINKING SIX STEPS OF CRITICAL THINKING.
2 nd Grade Language Arts World Class Outcome Create meaning strategically in: Reading Writing Speaking Listening Evaluate how authors are strategic in.
An evolutionary approach for improving the quality of automatic summaries Constantin Orasan Research Group in Computational Linguistics School of Humanities,
Copyright © 2011 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 1 Research: An Overview.
Depth of Knowledge: Elementary ELA Smarter Balanced Professional Development for Washington High-need Schools University of Washington Tacoma Belinda Louie,
Chapter 2: Algorithm Discovery and Design Invitation to Computer Science.
King Faisal University جامعة الملك فيصل Deanship of E-Learning and Distance Education عمادة التعلم الإلكتروني والتعليم عن بعد [ ] 1 جامعة الملك فيصل عمادة.
Features of science revision
Writing a sound proposal
Learning Target.
Purpose of Research Research may be broadly classified into two areas; basic and applied research. The primary purpose of basic research (as opposed to.
BBI 3212 ENGLISH SYNTAX AND MORPHOLOGY
Automatic Detection of Causal Relations for Question Answering
Command Terms
Danyun Xu, Gong Cheng*, Yuzhong Qu
Research Design and Methods
Chapter 8 VALIDITY AND RELIABILITY
Writing Tips: The Word “Prove”
TECHNICAL REPORTS WRITING
Presentation transcript:

Generation of Referring Expressions: Managing Structural Ambiguities I.H. KhanG. Ritchie K. van Deemter University of Aberdeen, UK

A natural language generator should avoid generating those phrases, which are too ambiguous to understand. But, how the generator can know whether a phrase is too ambiguous or not? We use corpus-based heuristics, backed by empirical evidence, that estimate the likelihood of different readings of a phrase, and guide the generator to choose an optimal phrase from the available alternatives.

Process of generating text in natural language (e.g., English) from some non-linguistic data (Reiter & Dale, 2000) Example NLG system –Pollen Forecast: generates reports from pollen forecast data Natural Language Generation (NLG) Grass pollen levels for Tuesday have decreased from the high levels of yesterday with values of around 4 to 5 across most parts of the country. However, in South Eastern areas, pollen levels will be high with values of 6. [courtesy E. Reiter]

Referring Expression = Noun Phrase –e.g., the black cat; the black cats and dogs (etc.) A key component in most NLG systems Task of GRE: –Given a set of intended referents, compute the properties of these referents that distinguish them from distractors in a KB Generation of Referring Expressions (GRE)

GRE: An Example ObjectsProperties Object 1 (Type, Sheep), (Color, Black) Object 2 (Type, Sheep), (Color, Brown) Object 3 (Type, Sheep), (Color, Black) Object 4 (Type, Goat), (Color, Black) Object 5 (Type, Goat), (Color, Yellow) Object 6 (Type, Goat), (Color, Black) Object 7 (Type, Goat), (Color, Brown) Object 8 (Type, Cow), (Color, Black) Output: Distinguishing Description (DD) – (Black Sheep) (Black Goat) KB Input: KB, Intended Referents R Task: find properties that distinguish R from distractors

The Problem NP 1 : The black sheep and the black goats NP 2 : The black sheep and goats (Black Sheep) (Black Goat) = {Object 1,Object 3,Object 4,Object 6 } (Black Sheep) Goat = {Object 1,Object 3,Object 4,Object 5,Object 6,Object 7 } NP 1 unambiguous and long; NP 2 ambiguous and brief Question: How the generator might chose between NP 1 and NP 2 ? Linguistic ambiguities can arise when DDs are realised

Our Approach Psycholinguistic evidence –Avoidance of all ambiguity is not feasible (Abney, 1996) Avoid only distractor interpretations –An interpretation is distractor if it is more likely or almost as likely as the intended one. Question –How to make distractor interpretation precise? Our solution –Getting likelihood using word sketches (cf. Chantree et el., 2004) –Word sketches provide detailed information about word relationships, based on corpus frequencies –Relationships are grammatical

Hypothesis 1 –If Adj modifies N 1 more often than N 2, then a narrow-scope reading is likely (no matter how frequently N1 and N2 co- occur). bearded men and women handsome men and women Hypothesis 2 –If Adj does not modify N 1 more often than N 2, then a wide- scope reading is likely (no matter how frequently N1 and N2 co-occur).. old men and women tall men and trees Pattern: the Adj N 1 and N 2

Experiment 1 Please, remove the roaring lions and horses.

Experiment 1: Results Hypothesis 2 (i.e., predictions for WS reading) is confirmed Hypothesis 1 (i.e., predictions for NS reading) is not confirmed –Tendency for WS (even though results are not stat. sig.) Tentative conclusion –An intrinsic bias in favour of WS reading BUT: The use of *unusual* features may have made peoples judgements unreliable

Experiment 2 Please, remove the figure containing the young lions and horses.

Experiment 2 (cont.) Please, remove the figure containing the barking dogs and cats. Results: Both hypotheses are confirmed

Word Sketches can make reasonable predictions about how an NP would be understood. But we need more to know from generation point of view: which of the following two NPs is best? The black sheep and the black goats The black sheep and goats (Black Sheep) (Black Goat) (Black Sheep) Goat We seek the answer in next experiment

Clarity-brevity trade-off Recall the pattern: the Adj Noun 1 and Noun 2 Brief descriptions (+b) take the form –the Adj Noun 1 and Noun 2 Non-brief descriptions (-b) take the form –the Adj Noun 1 and the Adj Noun 2 (IR = WS) –the Adj Noun 1 and the Noun 2 (IR = NS) Clear descriptions (+c) –Which have no distractor interpretations Non-clear descriptions (-c) –Which have some distractor interpretations

Hypothesis 1 –(+c, +b) descriptions are preferred over (+c, -b) Hypothesis 2 –(+c, -b) descriptions are preferred over (-c, +b) Each hypothesis is tested under two conditions –C 1 : intended reading is WS –C 2 : intended reading is NS The Hypotheses (Readers Preferences)

Experiment 3: NS Case Which phrase works best to identify the filled area? 1.The barking dogs and cats 2.The barking dogs and the cats

Experiment 3: WS Case Which phrase works best to identify the filled area? 1.The young lions and the young horses 2.The young lions and horses

Experiment 3: Results Both hypotheses are confirmed: –(+c, +b) descriptions are preferred over (+c, -b) –(+c, -b) descriptions are preferred over (-c, +b) Role of length : –In WS cases preferences are very strong –In NS cases preference is not as strong as in WS cases

Summary of Empirical Evidence For the pattern the Adj Noun 1 and Noun 2 –Word Sketches can make reliable predictions –Keeping clarity the same, a brief NP is better than a longer one

Algorithm Development Main knowledge sources –WordNet (for lexicalisation) –SketchEngine (for predicting the most likely reading) Main steps 1.Choose words 2.Use these to construct description in DNF 3.Use transformations to generate alternative structures from DNF 4.Select optimal phrase

Transformation Rules Input –Logical formula in DNF Rule Base 1.(A B 1 ) (A B 2 ) A (B 1 B 2 ) 2.(X Y) (Y X) [A = Adj, B 1 =B 2 =Noun, X=Y=(Adj and/or Noun)] Output –Set of logical formulae

Select optimal phrase 1.(black sheep) (black goats) DNF 2.(black goats) (black sheep) 3.black (goats sheep) 4.black (sheep goats) Optimal (4): Adj has high collocational frequency with N 1 and N 2, so the intended (wide- scope) reading is more likely. Therefore, (4) is selected.

Conclusions GRE should deal with surface ambiguities Word sketches can make distractor interpretation precise Keeping clarity the same, brief descriptions are preferred over longer ones A GRE algorithm is sketched that balances clarity and brevity

THANK YOU