Three Stories on Automated Reasoning for Natural Language Understanding Johan Bos University of Rome "La Sapienza Dipartimento di Informatica.

Slides:



Advertisements
Similar presentations
1 Knowledge Representation Introduction KR and Logic.
Advertisements

© Johan Bos Logic Programming About the course –Taught in English –Tuesday: mostly theory –Thursday: practical work [lab] Teaching material –Learn Prolog.
Charting the Potential of Description Logic for the Generation of Referring Expression SELLC, Guangzhou, Dec Yuan Ren, Kees van Deemter and Jeff.
Logic Programming Automated Reasoning in practice.
First-Order Logic (and beyond)
Introducing Formal Methods, Module 1, Version 1.1, Oct., Formal Specification and Analytical Verification L 5.
Working with Discourse Representation Theory Patrick Blackburn & Johan Bos Lecture 3 DRT and Inference.
Inference and Reasoning. Basic Idea Given a set of statements, does a new statement logically follow from this. For example If an animal has wings and.
Working with Discourse Representation Theory Patrick Blackburn & Johan Bos Lecture 4 Pronouns and Presupposition.
Propositional Logic CMSC 471 Chapter , 7.7 and Chuck Dyer
Copyright © Cengage Learning. All rights reserved.
Everything You Need to Know (since the midterm). Diagnosis Abductive diagnosis: a minimal set of (positive and negative) assumptions that entails the.
Inference Tasks and Computational Semantics. Key Concepts Inference tasks Syntactic versus semantic approach to logic Soundness & completeness Decidability.
INTERNATIONAL CONFERENCE ON NATURAL LANGUAGE PROCESSING NLP-AI IIIT-Hyderabad CIIL, Mysore ICON DECEMBER, 2003.
ISBN Chapter 3 Describing Syntax and Semantics.
Philosophy 120 Symbolic Logic I H. Hamner Hill CSTL-CLA.SEMO.EDU/HHILL/PL120.
Natural Language and Speech Processing Creation of computational models of the understanding and the generation of natural language. Different fields coming.
CPSC 322, Lecture 19Slide 1 Propositional Logic Intro, Syntax Computer Science cpsc322, Lecture 19 (Textbook Chpt ) February, 23, 2009.
Building Spoken Dialogue Systems for Embodied Agents Lecture 2 Johan Bos School of Informatics The University of Edinburgh.
LEARNING FROM OBSERVATIONS Yılmaz KILIÇASLAN. Definition Learning takes place as the agent observes its interactions with the world and its own decision-making.
Natural Language Processing AI - Weeks 19 & 20 Natural Language Processing Lee McCluskey, room 2/07
LEARNING FROM OBSERVATIONS Yılmaz KILIÇASLAN. Definition Learning takes place as the agent observes its interactions with the world and its own decision-making.
CS 330 Programming Languages 09 / 16 / 2008 Instructor: Michael Eckmann.
© Johan Bos April When logical inference helps in determining textual entailment ( and when it doesn’t) Johan Bos & Katja Markert Linguistic Computing.
Describing Syntax and Semantics
Logical and Rule-Based Reasoning Part I. Logical Models and Reasoning Big Question: Do people think logically?
Meaning and Language Part 1.
Statistical Natural Language Processing. What is NLP?  Natural Language Processing (NLP), or Computational Linguistics, is concerned with theoretical.
Lecture 1, 7/21/2005Natural Language Processing1 CS60057 Speech &Natural Language Processing Autumn 2005 Lecture 1 21 July 2005.
February 2009Introduction to Semantics1 Logic, Representation and Inference Introduction to Semantics What is semantics for? Role of FOL Montague Approach.
9/8/20151 Natural Language Processing Lecture Notes 1.
Lecture 12: 22/6/1435 Natural language processing Lecturer/ Kawther Abas 363CS – Artificial Intelligence.
Three Short Stories on Computerised Presupposition Projection Johan Bos University of Rome "La Sapienza“ Dipartimento di Informatica.
For Friday Finish chapter 23 Homework: –Chapter 22, exercise 9.
Assessing the Impact of Frame Semantics on Textual Entailment Authors: Aljoscha Burchardt, Marco Pennacchiotti, Stefan Thater, Manfred Pinkal Saarland.
Working with Discourse Representation Theory Patrick Blackburn & Johan Bos Lecture 5 Applying DRT.
Natural Language Processing Introduction. 2 Natural Language Processing We’re going to study what goes into getting computers to perform useful and interesting.
November 2003CSA4050: Semantics I1 CSA4050: Advanced Topics in NLP Semantics I What is semantics for? Role of FOL Montague Approach.
An Intelligent Analyzer and Understander of English Yorick Wilks 1975, ACM.
CMPF144 FUNDAMENTALS OF COMPUTING THEORY Module 5: Classical Logic.
5.3 Geometric Introduction to the Simplex Method The geometric method of the previous section is limited in that it is only useful for problems involving.
1 Knowledge Representation. 2 Definitions Knowledge Base Knowledge Base A set of representations of facts about the world. A set of representations of.
Formal Semantics Chapter Twenty-ThreeModern Programming Languages, 2nd ed.1.
LOGIC AND ONTOLOGY Both logic and ontology are important areas of philosophy covering large, diverse, and active research projects. These two areas overlap.
Computational Semantics Day 5: Inference Aljoscha.
Indirect Supervision Protocols for Learning in Natural Language Processing II. Learning by Inventing Binary Labels This work is supported by DARPA funding.
October 2005CSA3180 NLP1 CSA3180 Natural Language Processing Introduction and Course Overview.
CSA2050 Introduction to Computational Linguistics Lecture 1 Overview.
Machine Learning Chapter 5. Artificial IntelligenceChapter 52 Learning 1. Rote learning rote( โรท ) n. วิถีทาง, ทางเดิน, วิธีการตามปกติ, (by rote จากความทรงจำ.
Semantic Construction lecture 2. Semantic Construction Is there a systematic way of constructing semantic representation from a sentence of English? This.
1 Introduction to Computational Linguistics Eleni Miltsakaki AUTH Spring 2006-Lecture 8.
Key Concepts Representation Inference Semantics Discourse Pragmatics Computation.
3.2 Semantics. 2 Semantics Attribute Grammars The Meanings of Programs: Semantics Sebesta Chapter 3.
Programming Languages and Design Lecture 3 Semantic Specifications of Programming Languages Instructor: Li Ma Department of Computer Science Texas Southern.
Rules, Movement, Ambiguity
1/21 Automatic Discovery of Intentions in Text and its Application to Question Answering (ACL 2005 Student Research Workshop )
Sight Words.
Chapter 11 Introduction to Computational Complexity Copyright © 2011 The McGraw-Hill Companies, Inc. Permission required for reproduction or display. 1.
A Simple English-to-Punjabi Translation System By : Shailendra Singh.
Recognising Textual Entailment Johan Bos School of Informatics University of Edinburgh Scotland,UK.
Chapter 7. Propositional and Predicate Logic
What is a CAT? What is a CAT?.
Knowledge Representation and Reasoning
Natural Language Processing
CS246: Information Retrieval
Chapter 7. Propositional and Predicate Logic
Propositional Logic CMSC 471 Chapter , 7.7 and Chuck Dyer
Logical and Rule-Based Reasoning Part I
Presentation transcript:

Three Stories on Automated Reasoning for Natural Language Understanding Johan Bos University of Rome "La Sapienza Dipartimento di Informatica

Background My work is in between –natural language processing –computational linguistics –formal and computational semantics Aim of my work –implement linguistic theories –use automated reasoning in modeling natural language understanding

Applications What kind of applications? –Human-machine dialogue systems –Question answering systems –Textual entailment systems Use of logical inference –Off the shelf systems, FOL theorem provers and finite model builders –Empirically successful?

Surprise Perhaps surprisingly, automated reasoning tools rarely make it into NLP applications Why? –Requires interdisciplinary background –Gap between formal semantic theory and practical implementation –It is just not trendy --- statistical approaches dominate the field

Three Stories World Wide Computational Semantics –The worlds first serious implementation of Discourse Representation Theory, with the help of the web and theorem proving Godot, the talking robot –The first robot that computes semantic representations and performs inferences using theorem proving and model building Recognising Textual Entailment –Automated deduction applied in wide-coverage natural language processing

The First Story World Wide Computational Semantics The first serious implementation of Discourse Representation Theory, with the help of the internet…

How it started Implementing tools for the semantic analysis of English Follow linguistic theory as closely as possible –Discourse Representation Theory [DRT] –First-order logic –Presupposition projection Computational Semantics

How can we automate the process of associating semantic representations with expressions of natural language? How can we use logical representations of natural language expressions to automate the process of drawing inferences?

Basic idea Text: Vincent loves Mia.

Basic idea Text: Vincent loves Mia. DRT: x y vincent(x) mia(y) love(x,y)

Basic idea Text: Vincent loves Mia. DRT: FOL: x y(vincent(x) & mia(y) & love(x,y)) x y vincent(x) mia(y) love(x,y)

Basic idea Text: Vincent loves Mia. DRT: FOL: x y(vincent(x) & mia(y) & love(x,y)) BK: x (vincent(x) man(x)) x (mia(x) woman(x)) x (man(x) woman(x)) x y vincent(x) mia(y) love(x,y)

Basic idea Text: Vincent loves Mia. DRT: FOL: x y(vincent(x) & mia(y) & love(x,y)) BK: x (vincent(x) man(x)) x (mia(x) woman(x)) x (man(x) woman(x)) Model: D = {d1,d2} F(vincent)={d1} F(mia)={d2} F(love)={(d1,d2)} x y vincent(x) mia(y) love(x,y)

? Text: Vincent loves Mia. DRT: x y vincent(x) mia(y) love(x,y)

Compositional Semantics The Problem Given a natural language expression, how do we convert it into a logical formula? Freges principle The meaning of a compound expression is a function of the meaning of its parts.

Lexical semantics CategorySemanticsExample N x. spokesman NP/N p. q.(( ;p(x));q(x)) a NP/N p. q.(( ;p(x)) q(x)) the S\NP y. lied spokesman(x) X e lie(e) agent(e,y) X

A derivation NP/N:a N:spokesman S\NP:lied p. q. ;p(x);q(x) z. y. spokesman(z) x e lie(e) agent(e,y)

A derivation NP/N:a N:spokesman S\NP:lied p. q. ;p(x);q(x) z. y (FA) NP: a spokesman p. q. ;p(x);q(x)( z. ) spokesman(z) x e lie(e) agent(e,y) x

A derivation NP/N:a N:spokesman S\NP:lied p. q. ;p(x);q(x) z. y (FA) NP: a spokesman q. ; ;q(x)) spokesman(z) x spokesman(x) e lie(e) agent(e,y) x

A derivation NP/N:a N:spokesman S\NP:lied p. q. ;p(x);q(x) z. y (FA) NP: a spokesman q. ;q(x) spokesman(z) x x spokesman(x) e lie(e) agent(e,y)

A derivation NP/N:a N:spokesman S\NP:lied p. q. ;p(x);q(x) x. y (FA) NP: a spokesman q. ;q(x) (BA) S: a spokesman lied q. ;q(x)( y. ) spokesman(z) x x spokesman(x) e lie(e) agent(e,y) e lie(e) agent(e,y) x spokesman(x)

A derivation NP/N:a N:spokesman S\NP:lied p. q. ;p(x);q(x) x. y (FA) NP: a spokesman q. ;q(x) (BA) S: a spokesman lied ; spokesman(z) x x spokesman(x) e lie(e) agent(e,y) e lie(e) agent(e,x) x spokesman(x)

A derivation NP/N:a N:spokesman S\NP:lied p. q. ;p(x);q(x) x. y (FA) NP: a spokesman q. ;q(x) (BA) S: a spokesman lied spokesman(x) x x e lie(e) agent(e,y) x e spokesman(x) lie(e) agent(e,x)

The DORIS System Reasonable grammar coverage Parsed English sentences, followed by resolving ambiguities –Pronouns –Presupposition Generated many different semantic representation for a text

Texts and Ambiguity Usually, ambiguities cause many possible interpretations Example: Butch walks into his modest kitchen. He opens the refrigerator. He takes out a milk and drinks it.

Texts and Ambiguity Usually, ambiguities cause many possible interpretations Example: Butch walks into his modest kitchen. He opens the refrigerator. He takes out a milk and drinks it.

Texts and Ambiguity Usually, ambiguities cause many possible interpretations Example: Butch walks into his modest kitchen. He opens the refrigerator. He takes out a milk and drinks it.

Texts and Ambiguity Usually, ambiguities cause many possible interpretations Example: Butch walks into his modest kitchen. He opens the refrigerator. He takes out a milk and drinks it.

Basic idea of DORIS Given a text, produce as many different DRSs [semantic interpretations] as possible Filter out `strange` interpretations –Inconsistent interpretations –Uninformative interpretations Applying theorem proving –Use general purpose FOL theorem prover –Bliksem [Hans de Nivelle]

Screenshot

Consistency checking Inconsistent text: –Mia likes Vincent. –She does not like him. Two interpretations, only one consistent: –Mia likes Jody. –She does not like her.

Informativity checking Uninformative text: –Mia likes Vincent. –She likes him. Two interpretations, only one informative: –Mia likes Jody. –She likes her.

Local informativity Example: –Mia is the wife of Marsellus. –If Mia is the wife of Marsellus, Vincent will be disappointed. The second sentence is informative with respect to the first. But…

x y mia(x) marsellus(y) wife-of(x,y) Local informativity

x y z mia(x) marsellus(y) wife-of(x,y) vincent(z) wife-of(x,y) disappointed(z) Local informativity

Local consistency Example: –Jules likes big kahuna burgers. –If Jules does not like big kahuna burgers, Vincent will order a whopper. The second sentence is consistent with respect to the first. But…

x y jules(x) big-kahuna-burgers(y) like(x,y) Local consistency

x y z jules(x) big-kahuna-burgers(y) like(x,y) vincent(z) u order(z,u) whopper(u) Local consistency like(x,y)

Studying Presupposition The DORIS system allowed one to study the behaviour of presupposition Examples such as: –If Mia has a husband, then her husband is out of town. –If Mia is married, then her husband is out of town. –If Mia is dating Vincent, then her husband is out of town.

Applying Theorem Proving The first version of DORIS sort of worked, but… –Many readings to start with, explosion… –The local constraints added a large number of inference tasks –It could take about 10 minutes for a complex sentence

MathWeb MathWeb [by Kohlhase & Franke] came to the rescue Theorem proving services via the internet Interface Doris with MathWeb At the time this was a sensation! What happened exactly?

World Wide Computational Semantics Just in order to find out whether Mia was married or not, thousands of computers world wide were used…

What can we say about DORIS? Demonstrated that FOL could play an interesting role in natural language understanding Helped to develop presupposition theory Generated a new application area for automated deduction [TPTP] Limitations –Scalability –Coverage

Six feet under The DORIS system Why? –Limited grammatical coverage –Unfocussed application domain –It would take at least 20 years to develop a parser that was both robust and accurate [at least that was my belief]

The Second Story Godot the talking robot Or how a mobile, talking robot engaged in conversation using theorem provers and model builders for first-order logic…

Human-Computer Dialogue Focus on small domains –Grammatical coverage ensured –Background knowledge encoding Spoken Dialogue system = killer app –Godot the robot –Speech recognition and synthesis –People could give Godot directions, ask it questions, tell it new information –Godot was a REAL robot

Godot the Robot Godot with Tetsushi Oka

Typical conversation with Godot Simple dialogues in beginning: –Human: Robot? –Robot: Yes? –Human: Where are you? –Robot: I am in the hallway. –Human: OK. Go to the rest room!

Adding inference to Godot Using the components of DORIS, we added semantics and inference –Using DRT –Inference tasks: Consistency checking Informativity checking –Using FOL theorem provers and finite model builders

Advanced conversation with Godot Dealing with inconsistency and informativeness –Human: Robot, where are you? –Robot: I am in the hallway. –Human: You are in my office. –Robot: No, that is not true. –Human: You are in the hallway. –Robot: Yes I know!

Parallel inference Given a text T and a translation into FOL –If theorem prover finds a proof for, then T is inconsistent –If model builder finds a model for, then T is consistent In other words, this is the Yin and Yen of inference

Why is this relevant to natural language? Testing a discourse for consistency DiscourseTheorem proverModel builder

Why is this relevant to natural language? Testing a discourse for consistency DiscourseTheorem proverModel builder Vincent is a man. ??Model

Why is this relevant to natural language? Testing a discourse for consistency DiscourseTheorem proverModel builder Vincent is a man. ??Model Mia loves every man. ??Model

Why is this relevant to natural language? Testing a discourse for consistency DiscourseTheorem proverModel builder Vincent is a man. ??Model Mia loves every man. ??Model Mia does not love Vincent. Proof??

Why is this relevant to natural language? Testing a discourse for informativity DiscourseTheorem proverModel builder

Why is this relevant to natural language? Testing a discourse for informativity DiscourseTheorem proverModel builder Vincent is a man. ??Model

Why is this relevant to natural language? Testing a discourse for informativity DiscourseTheorem proverModel builder Vincent is a man. ??Model Mia loves every man. ??Model

Why is this relevant to natural language? Testing a discourse for informativity DiscourseTheorem proverModel builder Vincent is a man. ??Model Mia loves every man. ??Model Mia loves Vincent. Proof??

Minimal Models Model builders normally generate models by iteration over the domain size As a side-effect, the output is a model with a minimal domain size From a linguistic point of view, this is interesting, as there is no redundant information Minimal in extensions

Using models Examples: Turn on a light. Turn on every light. Turn on everything except the radio. Turn off the red light or the blue light. Turn on another light.

Videos of Godot Video 1: Godot in the basement of Bucceuch Place Video 2: Screenshot of dialogue manager with DRSs and camera view of Godot

What can we say about Godot? Demonstrated that FOL could play an interesting role in human machine dialogue systems Also showed a new application of finite model building Domain known means all background knowledge known Limitations –Scalability, only `small` dialogues –Lack of incremental inference –Minimal models required

Godot the Robot [later] Godot at the Scottish museum

The Third Story 2005-present Recognising Textual Entailment Or how first-order automated deduction is applied to wide- coverage semantic processing of texts

Recognising Textual Entailment What is it? –A task for NLP systems to recognise entailment between two (short) texts –Proved to be a difficult, but popular task. Organisation –Introduced in 2004/2005 as part of the PASCAL Network of Excellence, RTE-1 –A second challenge (RTE-2) was held in 2005/2006 –PASCAL provided a development and test set of several hundred examples

RTE Example (entailment) RTE 1977 (TRUE) His family has steadfastly denied the charges The charges were denied by his family.

RTE Example (no entailment) RTE 2030 (FALSE) Lyon is actually the gastronomical capital of France Lyon is the capital of France.

Aristotles Syllogisms All men are mortal. Socrates is a man Socrates is mortal. ARISTOTLE 1 (TRUE)

Five methods Five different methods to RTE Ranging in sophistication from very basic to advanced

Recognising Textual Entailment Method 1: Flipping a coin

Advantages –Easy to implement –Cheap Disadvantages –Just 50% accuracy

Recognising Textual Entailment Method 2: Calling a friend

Advantages –High accuracy (95%) Disadvantages –Lose friends –High phone bill

Recognising Textual Entailment Method 3: Ask the audience

RTE 893 (????) The first settlements on the site of Jakarta were established at the mouth of the Ciliwung, perhaps as early as the 5 th century AD The first settlements on the site of Jakarta were established as early as the 5 th century AD.

Human Upper Bound RTE 893 (TRUE) The first settlements on the site of Jakarta were established at the mouth of the Ciliwung, perhaps as early as the 5 th century AD The first settlements on the site of Jakarta were established as early as the 5 th century AD.

Recognising Textual Entailment Method 4: Word Overlap

Word Overlap Approaches Popular approach Ranging in sophistication from simple bag of word to use of WordNet Accuracy rates ca. 55%

Word Overlap Advantages –Relatively straightforward algorithm Disadvantages –Hardly better than flipping a coin

RTE State-of-the-Art Pascal RTE challenge Hard problem Requires semantics

Recognising Textual Entailment Method 5: Semantic Interpretation

Basic idea Given a textual entailment pair T/H with text T and hypothesis H: –Produce DRSs for T and H –Translate these DRSs into FOL –Generate Background Knowledge in FOL Use ATPs to determine the likelyhood of entailment

Wait a minute This requires that we have the means to produce semantic representations [DRSs] for any kind of English input Recall DORIS experience Do we have English parsers at our disposal that do this?

Robust Parsing Rapid developments in statistical parsing the last decades These parsers are trained on large annotated corpora [tree banks] Yet most of these parsers produced syntactic analyses not suitable for systematic semantic work This changed with the development of CCG bank and a fast CCG parser

Implementation CCG/DRT Use standard statistical techniques –Robust wide-coverage parser –Clark & Curran (ACL 2004) Grammar derived from CCGbank –409 different categories –Hockenmaier & Steedman (ACL 2002) Compositional Semantics, DRT –Wide-coverage semantics –Bos (IWCS 2005)

Example Output Example: Pierre Vinken, 61 years old, will join the board as a nonexecutive director Nov. 29. Mr. Vinken is chairman of Elsevier N.V., the Dutch publishing group. Semantic representation, DRT Complete Wall Street Journal

Using Theorem Proving Given a textual entailment pair T/H with text T and hypothesis H: –Produce DRSs for T and H –Translate these DRSs into FOL –Give this to the theorem prover: T H If the theorem prover finds a proof, then we predict that T entails H

Vampire (Riazanov & Voronkov 2002) Lets try this. We will use the theorem prover Vampire This gives us good results for: –apposition –relative clauses –coodination –intersective adjectives/complements –passive/active alternations

Example (Vampire: proof) On Friday evening, a car bomb exploded outside a Shiite mosque in Iskandariyah, 30 miles south of the capital A bomb exploded outside a mosque. RTE (TRUE)

Example (Vampire: proof) Initially, the Bundesbank opposed the introduction of the euro but was compelled to accept it in light of the political pressure of the capitalist politicians who supported its introduction The introduction of the euro has been opposed. RTE (TRUE)

Background Knowledge However, it doesnt give us good results for cases requiring additional knowledge –Lexical knowledge –World knowledge We will use WordNet as a start to get additional knowledge All of WordNet is too much, so we create MiniWordNets

MiniWordNets –Use hyponym relations from WordNet to build an ontology –Do this only for the relevant symbols –Convert the ontology into first-order axioms

MiniWordNet: an example Example text: There is no asbestos in our products now. Neither Lorillard nor the researchers who studied the workers were aware of any research on smokers of the Kent cigarettes.

MiniWordNet: an example Example text: There is no asbestos in our products now. Neither Lorillard nor the researchers who studied the workers were aware of any research on smokers of the Kent cigarettes.

x(user(x) person(x)) x(worker(x) person(x)) x(researcher(x) person(x))

x(person(x) risk(x)) x(person(x) cigarette(x)) …….

Using Background Knowledge Given a textual entailment pair T/H with text T and hypothesis H: –Produce DRS for T and H –Translate drs(T) and drs(H) into FOL –Create Background Knowledge for T&H –Give this to the theorem prover: (BK & T) H

MiniWordNets at work Background Knowledge: x(soar(x) rise(x)) Crude oil prices soared to record levels Crude oil prices rise. RTE 1952 (TRUE)

Troubles with theorem proving Theorem provers are extremely precise. They wont tell you when there is almost a proof. Even if there is a little background knowledge missing, Vampire will say: don`t know

Vampire: no proof RTE 1049 (TRUE) Four Venezuelan firefighters who were traveling to a training course in Texas were killed when their sport utility vehicle drifted onto the shoulder of a Highway and struck a parked truck Four firefighters were killed in a car accident.

Using Model Building Need a robust way of inference Use model builders Mace, Paradox –McCune –Claessen & Sorensson (2003) Use size of (minimal) model –Compare size of model of T and T&H –If the difference is small, then it is likely that T entails H

Using Model Building Given a textual entailment pair T/H with text T and hypothesis H: –Produce DRSs for T and H –Translate these DRSs into FOL –Generate Background Knowledge –Give this to the Model Builder: i) BK & T ii) BK & T & H If the models for i) and ii) are similar, then we predict that T entails H

Model similarity When are two models similar? –Small difference in domain size –Small difference in predicate extensions

Example 1 T: John met Mary in Rome H: John met Mary Model T: 3 entities Model T+H: 3 entities Modelsize difference: 0 Prediction: entailment

Example 2 T: John met Mary H: John met Mary in Rome Model T: 2 entities Model T+H: 3 entities Modelsize difference: 1 Prediction: no entailment

Model size differences Of course this is a very rough approximation But it turns out to be a useful one Gives us a notion of robustness Negation –Give not T and not [T & H] to model builder Disjunction –Not necessarily one unique minimal model

How well does this work? We tried this at the RTE-1 and RTE-2 Using standard machine learning methods to build a decision tree using features –Proof (yes/no) –Domain size difference –Model size difference Better than baseline, still room for improvement

RTE Results 2004/5 Accuracy Shallow0.569 Deep0.562 Hybrid (S+D)0.577 Hybrid+Task0.612 Bos & Markert 2005

RTE State-of-the-Art Pascal RTE challenge Hard problem Requires semantics

What can we say about RTE? We can use FOL inference techniques successfully There might be an interesting role for model building The bottleneck is getting the right background knowledge

Lack of Background Knowledge RTE (TRUE) Indonesia says the oil blocks are within its borders, as does Malaysia, which has also sent warships to the area, claiming that its waters and airspace have been violated There is a territorial waters dispute.

Winding up Summary Conclusion Shameless Plug Future

Summary Use of first order inference tools has a major influence on how computational semantics is perceived today –Implementations used in pioneering work of using first-order inference in NLP –Implementations used in spoken dialogue systems –Now also used in wide-coverage NLP systems

Conclusions We have got the tools for doing computational semantics in a principled way using DRT For many applications, success depends on the ability to systematically generate background knowledge –Small restricted domains [dialogue] –Open domain Finite model building has potential Incremental inference

Shameless Plug For more on the basic architecture underlying this work on computational semantics, and particular on implementations on the lambda calculus, and parallel use of theorem provers and model builders, see: