Embodied Construction Grammar in language (acquisition and) use Jerome Feldman Computer Science Division, University of California,

Slides:



Advertisements
Similar presentations
Ontology-Based Computing Kenneth Baclawski Northeastern University and Jarg.
Advertisements

Statistical NLP: Lecture 3
QuASI: Question Answering using Statistics, Semantics, and Inference Marti Hearst, Jerry Feldman, Chris Manning, Srini Narayanan Univ. of California-Berkeley.
For Monday Read Chapter 23, sections 3-4 Homework –Chapter 23, exercises 1, 6, 14, 19 –Do them in order. Do NOT read ahead.
NLP and Speech Course Review. Morphological Analyzer Lexicon Part-of-Speech (POS) Tagging Grammar Rules Parser thethe – determiner Det NP → Det.
NaLIX: A Generic Natural Language Search Environment for XML Data Presented by: Erik Mathisen 02/12/2008.
Amirkabir University of Technology Computer Engineering Faculty AILAB Efficient Parsing Ahmad Abdollahzadeh Barfouroush Aban 1381 Natural Language Processing.
The Neural Basis of Thought and Language Final Review Session.
Knowledge Acquisitioning. Definition The transfer and transformation of potential problem solving expertise from some knowledge source to a program.
NTL – Converging Constraints Basic concepts and words derive their meaning from embodied experience. Abstract and theoretical concepts derive their meaning.
CS 182 Sections slides created by: Eva Mok modified by jgm April 26, 2006.
Semi-Automatic Learning of Transfer Rules for Machine Translation of Low-Density Languages Katharina Probst April 5, 2002.
Artificial Intelligence 2004 Natural Language Processing - Syntax and Parsing - Language Syntax Parsing.
Lecture 1 Introduction: Linguistic Theory and Theories
1 Basic Parsing with Context Free Grammars Chapter 13 September/October 2012 Lecture 6.
Phonetics, Phonology, Morphology and Syntax
March 1, 2009 Dr. Muhammed Al-Mulhem 1 ICS 482 Natural Language Processing INTRODUCTION Muhammed Al-Mulhem March 1, 2009.
Statistical Natural Language Processing. What is NLP?  Natural Language Processing (NLP), or Computational Linguistics, is concerned with theoretical.
Artificial Intelligence Research Centre Program Systems Institute Russian Academy of Science Pereslavl-Zalessky Russia.
Basic Concepts The Unified Modeling Language (UML) SYSC System Analysis and Design.
Lecture 1, 7/21/2005Natural Language Processing1 CS60057 Speech &Natural Language Processing Autumn 2005 Lecture 1 21 July 2005.
Amarnath Gupta Univ. of California San Diego. An Abstract Question There is no concrete answer …but …
35 years of Cognitive Linguistics Session 8: Cognitive Grammar
NTL – Converging Constraints Basic concepts and words derive their meaning from embodied experience. Abstract and theoretical concepts derive their meaning.
9/8/20151 Natural Language Processing Lecture Notes 1.
Some Thoughts to Consider 13 What do we really mean by ‘learning’ in a software system? Can humans or systems learn anything that they don’t already know?
Lecture 12: 22/6/1435 Natural language processing Lecturer/ Kawther Abas 363CS – Artificial Intelligence.
For Friday Finish chapter 23 Homework: –Chapter 22, exercise 9.
A Cognitive Substrate for Natural Language Understanding Nick Cassimatis Arthi Murugesan Magdalena Bugajska.
Interpreting Dictionary Definitions Dan Tecuci May 2002.
Jennie Ning Zheng Linda Melchor Ferhat Omur. Contents Introduction WordNet Application – WordNet Data Structure - WordNet FrameNet Application – FrameNet.
THE BIG PICTURE Basic Assumptions Linguistics is the empirical science that studies language (or linguistic behavior) Linguistics proposes theories (models)
QuASI: Question Answering using Statistics, Semantics, and Inference Marti Hearst, Jerry Feldman, Chris Manning, Srini Narayanan Univ. of California-Berkeley.
Introduction to CL & NLP CMSC April 1, 2003.
A Cascaded Finite-State Parser for German Michael Schiehlen Institut für Maschinelle Sprachverarbeitung Universität Stuttgart
1 CSI 5180: Topics in AI: Natural Language Processing, A Statistical Approach Instructor: Nathalie Japkowicz Objectives of.
Introduction to Embodied Construction Grammar March 4, 2003 Ben Bergen
Linguistic Essentials
Artificial Intelligence Research Center Pereslavl-Zalessky, Russia Program Systems Institute, RAS.
What you have learned and how you can use it : Grammars and Lexicons Parts I-III.
Rules, Movement, Ambiguity
Artificial Intelligence 2004 Ontology
The Minimalist Program
Chapter 10. The Explorer System in Cognitive Systems, Christensen et al. Course: Robots Learning from Humans On, Kyoung-Woon Biointelligence Laboratory.
For Friday Finish chapter 23 Homework –Chapter 23, exercise 15.
CPSC 422, Lecture 27Slide 1 Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 27 Nov, 16, 2015.
Commonsense Reasoning in and over Natural Language Hugo Liu, Push Singh Media Laboratory of MIT The 8 th International Conference on Knowledge- Based Intelligent.
QuASI: Question Answering using Statistics, Semantics, and Inference Marti Hearst, Jerry Feldman, Chris Manning, Srini Narayanan Univ. of California-Berkeley.
RULES Patty Nordstrom Hien Nguyen. "Cognitive Skills are Realized by Production Rules"
A Portrait of the Semantic Web in Action Jeff Heflin and James Hendler IEEE Intelligent Systems December 6, 2010 Hyewon Lim.
For Monday Read chapter 26 Homework: –Chapter 23, exercises 8 and 9.
Understanding Naturally Conveyed Explanations of Device Behavior Michael Oltmans and Randall Davis MIT Artificial Intelligence Lab.
NATURAL LANGUAGE PROCESSING
MENTAL GRAMMAR Language and mind. First half of 20 th cent. – What the main goal of linguistics should be? Behaviorism – Bloomfield: goal of linguistics.
Figure and Ground Part 2 APLNG 597C LEJIAO WANG 03/16/2015.
The Neural Basis of Thought and Language Final Review Session.
The Neural Basis of Thought and Language Week 14.
The Neural Basis of Thought and Language Week 14.
Cognitive Language Processing for Rosie
CSC 594 Topics in AI – Natural Language Processing
QuASI: Question Answering using Statistics, Semantics, and Inference
Statistical NLP: Lecture 3
Basic Parsing with Context Free Grammars Chapter 13
Kenneth Baclawski et. al. PSB /11/7 Sa-Im Shin
Natural Language - General
Semantic Markup for Semantic Web Tools:
CS246: Information Retrieval
Artificial Intelligence 2004 Speech & Natural Language Processing
Information Retrieval
Presentation transcript:

Embodied Construction Grammar in language (acquisition and) use Jerome Feldman Computer Science Division, University of California, Berkeley, and International Computer Science Institute

State of the Art Limited Commercial Speech Applications transcription, simple response systems Statistical NLP for Restricted Tasks tagging, parsing, information retrieval Template-based Understanding programs expensive, brittle, inflexible, unnatural Essentially no NLU in HCI, QA Systems

“Of all the above fields the learning of languages would be the most impressive, since it is the most human of these activities. This field seems however to depend rather too much on sense organs and locomotion to be feasible.” Alan M. Turing Intelligent Machinery (1948)

What does language do? “Harry walked to the cafe.” “Harry walked into the cafe.” A sentence can evoke an imagined scene and resulting inferences : CAFE –Goal of action = at cafe –Source = away from cafe –cafe = point-like location –Goal of action = inside cafe –Source = outside cafe –cafe = containing location

Language understanding Interpretation (Utterance, Situation) Linguistic knowledge Conceptual knowledge Analysis

Language understanding: analysis & simulation “Harry walked to the cafe.” SchemaTrajector Goal walkHarrycafe Cafe Lexicon Constructicon General Knowledge Belief State Analysis Process Semantic Specification Utterance Simulation

Interpretation: x-schema simulation Constructions can specify which schemas and entities are involved in an event, and how they are related profile particular stages of an event set parameters of an event energy walker at goal walker =Harry goal =home Harry is walking home.

Phonetics Semantics Pragmatics Morphology Syntax Traditional Levels of Analysis

Phonetics Semantics Pragmatics Morphology Syntax “Harry walked into the cafe.” Utterance

Construction Grammar to block walk FormMeaning A construction is a form-meaning pair whose properties may not be strictly predictable from other constructions. (Construction Grammar, Goldberg 1995) Source Path Goal Trajector

Form-meaning mappings for language Form phonological cues word order intonation inflection Meaning event structure sensorimotor control attention/perspective social goals... Linguistic knowledge consists of form-meaning mappings : Cafe

Constructions as maps between relations Mover + Motion + Direction before(Motion, Direction) before(Mover, Motion) “is” + Action + “ing” before(“is”, Action) suffix(Action, “ing”) Mover + Motion before(Mover, Motion) FormMeaning ProgressiveAction aspect(Action, ongoing) MotionEvent mover(Motion, Mover) DirectedMotionEvent direction(Motion, Direction) mover(Motion, Mover) Complex constructions are mappings between relations in form and relations in meaning.

Embodied Construction Grammar (Bergen and Chang 2002) Embodied representations –active perceptual and motor schemas –situational and discourse context Construction Grammar –Linguistic units relate form and meaning / function. –Both constituency and (lexical) dependencies allowed. Constraint-based (Unification) –based on feature structures (as in HPSG) –Diverse factors can flexibly interact.

schema Container roles interior exterior portal boundary Representing image schemas Interior Exterior Boundary Portal Source Path Goal Trajector These are abstractions over sensorimotor experiences. schema Source-Path-Goal roles source path goal trajector schema name role name

Inference and Conceptual Schemas Hypothesis: –Linguistic input is converted into a mental simulation based on bodily-grounded structures. Components: –Semantic schemas image schemas and executing schemas are abstractions over neurally grounded perceptual and motor representations –Linguistic units lexical and phrasal construction representations invoke schemas, in part through metaphor Inference links these structures and provides parameters for a simulation engine

Early Example Understanding News Stories France fell into recession. Pulled out by Germany In1991, India set out on a path of liberalization. The Government started to loosen its stranglehold on business and removed obstacles to international trade. Now the Government is stumbling in implementing the liberalization plan.

Task Interpret simple discourse fragments/blurbs –France fell into recession. Pulled out by Germany –Economy moving at the pace of a Clinton jog. –US Economy on the verge of falling back into recession after moving forward on an anemic recovery. –Indian Government stumbling in implementing Liberalization plan. –Moving forward on all fronts, we are going to be ongoing and relentless as we tighten the net of justice. –The Government is taking bold new steps. We are loosening the stranglehold on business, slashing tariffs and removing obstacles to international trade.

I/O as Feature Structures Indian Government stumbling in implementing liberalization plan

Language understanding: analysis & simulation “Harry walked into the cafe.” Analysis Process Semantic Specification Utterance Constructions General Knowledge Belief State CAFE Simulation construction W ALKED form self f.phon  [wakt] meaning : Walk-Action constraints self m.time before Context.speech-time self m..aspect  encapsulated

Embodied Construction Grammar provides formal tools for linguistic description and analysis motivated largely by cognitive/functional concerns. Allows precise specifications of structures/processes involved in acquisition of early constructions –Embodied constructions (structured maps between form and meaning); lexically specific and more general –Usage-based processes of learning new constructions to account for co-occurring utterance-situation pairs Bridge to detailed psycholinguistic and neural imaging experiments

Formal Cognitive Linguistics Schemas and frames –Image schemas, force dynamics, executing schemas… Constructions –Lexical, grammatical, morphological, gestural… Maps –Metaphor, metonymy, mental space maps… Mental spaces –Discourse, hypothetical, counterfactual…

Embodied constructions construction H ARRY form : [hEriy] meaning : Harry construction C AFE form : [k h aefej] meaning : Cafe Harry CAFE cafe Notation FormMeaning Constructions have form and meaning poles that are subject to type constraints.

Schema Formalism SCHEMA SUBCASE OF EVOKES AS ROLES : CONSTRAINTS :: :: |

A Simple Example SCHEMA hypotenuse SUBCASE OF line-segment EVOKES right-triangle AS rt ROLES Comment inherited from line-segment CONSTRAINTS SELF rt.long-side

Source-Path-Goal SCHEMA: spg ROLES: source: Place path: Directed Curve goal: Place trajector: Entity

Translational Motion SCHEMA translational motion SUBCASE OF motion EVOKES spg AS s ROLES mover s.trajector source s.source goal s.goal CONSTRAINTS before:: mover.location source after:: mover.location goal

Construction Formalism CONSTRUCTION SUBCASE OF CONSTRUCTIONAL EVOKES AS CONSTITUENTS : CONSTRAINTS // as in SCHEMAs FORM ELEMENTS CONSTRAINTS // as in SCHEMAs MEANING // as in SCHEMAs

The meaning pole may evoke schemas (e.g., image schemas) with a local alias. The meaning pole may include constraints on the schemas (e.g., identification constraints  ). construction T O form self f.phon  [t h u w ] meaning evokes Trajector-Landmark as tl Source-Path-Goal as spg constraints: tl.trajector  spg.trajector tl.landmark  spg.goal construction T O form self f.phon  [t h u w ] meaning evokes Trajector-Landmark as tl Source-Path-Goal as spg constraints: tl.trajector  spg.trajector tl.landmark  spg.goal Representing constructions: T O local alias identification constraint

T O vs. I NTO : I NTO adds a Container schema and appropriate bindings. The I NTO construction construction I NTO form self f.phon  [ I nt h u w ] meaning evokes Trajector-Landmark as tl Source-Path-Goal as spg Container as cont constraints: tl.trajector  spg.trajector tl.landmark  cont cont.interior  spg.goal cont.exterior  spg.source

construction S PATIAL- P REDICATION constructional constituents sp : Trajector-Landmark lm : Thing form sp f before lm f meaning sp m.landmark  lm m Constructions with constituents: The S PATIAL- P REDICATION construction Constructions may also specify constructional constituents and impose form and meaning constraints on them: –order constraints –identification constraints order constraint local alias identification constraint

Grammatical Construction Example CONSTRUCTION Spatial-PP SUBCASE OF Phrase CONSTRUCTIONAL CONSTITUENTS rel: Spatial-Preposition lm: Referring-Exp CONSTRAINTS rel.case lm.case FORM rel < lm MEANING CONSTRAINTS rel.landmark lm

The D IRECTED- M OTION construction construction D IRECTED -M OTION constructional constituents mover : Thing motion : Motion-Process direction : Source-Path-Goal form mover f before motion f motion f before direction f meaning evokes Motion-Event as m m.mover  mover m m.motion  motion m m.path  direction m direction m.trajector  mover m motion m.mover  mover m

Semantic specification The analysis process produces a semantic specification that includes image-schematic, motor control and conceptual structures provides parameters for a mental simulation

Language Understanding Process

Constructional analysis

Semantic Specification

Language understanding: analysis & simulation “Harry walked into the cafe.” Analysis Process Semantic Specification Utterance Constructions General Knowledge Belief State CAFE Simulation construction W ALKED form self f.phon  [wakt] meaning : Walk-Action constraints self m.time before Context.speech-time self m..aspect  encapsulated

Simulation-based sense disambiguation The scientist walked into the laboratory. The scientist walked into the wall. Ease of construing nominal as a CONTAINER determines what sense of into is appropriate: CONTAINER senseCONTACT sense LABWALL Bonk!!

Simulation-based inference The teacher drifted into the house. The smoke drifted into the house. Detailed inferences can result from simulation. Image-schematic content of prepositions must fit with properties of other elements of sentence. –Final location of Trajector = inside cafe –Portal = door –Final location of Trajector = inside (possibly throughout) cafe –Portal = door/window

World knowledge informs simulation Physical knowledge of how people and gases interact with houses determines: –Relation between Trajector and Interior The smoke drifted into the house and filled it. ?The teacher drifted into the house and filled it. –Portal for motion across Boundary The smoke drifted into the house because the window had been left open. ?The teacher drifted into the house because the window had been left open.

ECG applications Grammar (Note: Theme Session on ECG at ICLC 2003, La Rioja) –Spatial relations/events (Bergen & Chang 1999; Bretones et al. In press) –Verbal morphology (Gurevich 2003, Bergen ms.) –Reference: measure phrases (Dodge and Wright 2002), construal resolution (Porzel & Bryant 2003), reflexive pronouns (Sanders 2003) Semantic representations / inference –Aspectual inference (Narayanan 1997; Chang, Gildea & Narayanan 1998) –Perspective / frames (Chang, Narayanan & Petruck 2002) –Metaphorical inference (Narayanan 1997, 1999) –Simulation semantics (Narayanan 1997, 1999) Language acquisition –Lexical acquisition (Regier 1996, Bailey 1997) – Multi-word constructions (Chang & Maia 2001)

Getting From the Utterance to the SemSpec Johno Bryant Need a grammar formalism –Embodied Construction Grammar (Bergen & Chang 2002) Need new models for language analysis –Traditional methods too limited –Traditional methods also don’t get enough leverage out of the semantics.

Embodied Construction Grammar Semantic Freedom –Designed to be symbiotic with cognitive approaches to meaning –More expressive semantic operators than traditional grammar formalisms Form Freedom –Free word order, over-lapping constituency Precise enough to be implemented

Traditional Parsing Methods Fall Short PSG parsers too strict –Constructions not allowed to leave constituent order unspecified Traditional way of dealing with incomplete analyses is ad-hoc –Making sense of incomplete analyses is important when an application must deal with “ill-formed” input (For example, modeling language learning) Traditional unification grammar can’t handle ECG’s deep semantic operators.

Our Analyzer Replaces the FSMs used in traditional chunking (Abney 96) with much more powerful machines capable of backtracking called construction recognizers Arranges these recognizers into levels just like in Abney’s work But uses a chart to deal with ambiguity

Our Analyzer (cont’d) Uses specialized feature structures to deal with ECG’s novel semantic operators Supports a heuristic evaluation metric for finding the “right” analysis Puts partial analyses together when no complete analyses are available –The analyzer was designed under the assumption that the grammar won’t cover every meaningful utterance encountered by the system.

System Architecture Learner Semantic Chunker Semantic Integration Grammar/Utterance Chunk Chart Ranked Analyses

The Levels The analyzer puts the recognizer on the level assigned by the grammar writer. –Assigned level should be greater than or equal to the levels of the construction’s constituents. The analyzer runs all the recognizers on level 1, then level 2, etc. until no more levels. Recognizers on the same level can be mutually recursive.

Recognizers Each Construction is turned into a recognizer Recognizer = active representation –seeks form elements/constituents when initiated –Unites grammar and process - grammar isn’t just a static piece of knowledge in this model. Checks both form and semantic constraints –Contains an internal representation of both the semantics and the form –A graph data structure used to represent the form and a feature structure representation for the meaning.

Recognizer Example Path Patient ActionAgent Mary kicked the ball into the net. This is the initial Constituent Graph for caused-motion.

Recognizer Example Construct: Caused-Motion Constituent: Agent Constituent: Action Constituent: Patient Constituent: Path The initial constructional tree for the instance of Caused-Motion that we are trying to create.

Recognizer Example

processed Mary kicked the ball into the net. Path Patient ActionAgent A node filled with gray is removed.

Recognizer Example Construct: Caused-Motion Constituent: Action Constituent: Patient Constituent: Path RefExp: Mary Mary kicked the ball into the net.

Recognizer Example

processed Mary kicked the ball into the net. Path Patient ActionAgent

Recognizer Example Construct: Caused-Motion Verb: kicked Constituent: Patient Constituent: Path RefExp: Mary Mary kicked the ball into the net.

Recognizer Example

processed Mary kicked the ball into the net. Path Patient ActionAgent According to the Constituent Graph, The next constituent can either be the Patient or the Path.

Recognizer Example processed Mary kicked the ball into the net. Path Patient ActionAgent

Recognizer Example Construct: Caused-Motion Verb: kicked RefExp: Det Noun Constituent: Path RefExp: Mary Mary kicked the ball into the net. Noun Det

Recognizer Example

processed Mary kicked the ball into the net. Path Patient ActionAgent

Recognizer Example Construct: Caused-Motion Verb: kicked RefExp: Det Noun Spatial-Pred: Prep RefExp RefExp: Mary Mary kicked the ball into the net. Noun Det Noun DetPrep RefExp

Recognizer Example

Scene = Caused-Motion Agent = Mary Action = Kick Patient = Path.Trajector = The Ball Path = Into the net Path.Goal = The net After analyzing the sentence, the following identities are asserted in the resulting SemSpec: Resulting SemSpec

Progress The analyzer (as described so far) is already being put to use in Chang’s thesis work. –The levels are well-suited to incremental learning. –Syntactic robustness important for generating partial analyses with poor coverage It will also be used this semester for producing SemSpecs for Narayanan’s enactment engine. –Put the deep semantics towards parameterizing x-schemas

Chunking the woman in the lab coat thought you were sleeping L0 D N P D N N V-tns Pron Aux V-ing L1 ____NP P_______NP VP NP ______VP L2 ____NP _________PP VP NP ______VP L3 ________________________S_____________S Cite/description

Construction Recognizers You want to put a cloth on your hand ? NP Form Meaning “you” [Addressee] Form Meaning D,N [Cloth num:sg] Form Meaning PP$,N [Hand num:sg poss:addr] Like Abney:Unlike Abney: One recognizer per rule Bottom up and level-based Check form and semantics More powerful/slower than FSMs

Chunk Chart Interface between chunking and structure merging Each edge is linked to its corresponding semantics. You want to put a cloth on your hand ?

Combining Partial Parses Prefer an analysis that spans the input utterance with the minimum number of chunks. When no spanning analysis exists, however, we still have a chart full of semantic chunks. The system tries to build a coherent analysis out of these semantics chunks. This is where structure merging comes in.

Structure Merging Closely related to abductive inferential mechanisms like abduction (Hobbs) Unify compatible structures (find fillers for frame roles) Intuition: Unify structures that would have been co- indexed had the missing construction been defined. There are many possible ways to merge structures. In fact, there are an exponential number of ways to merge structures (NP Hard). But using heuristics cuts down the search space.

Structure Merging Example Utterance:You used to hate to have the bib put on. [Addressee < Animate] Bib < Clothing num:sg givenness:def Caused-Motion-Action Agent: [Animate] Patient: [Entity] Path:On Before Merging: After Merging: Caused-Motion-Action Agent: [Addressee] Patient: Path:On Bib < Clothing num:sg givenness:def

Semantic Density Semantic density is a simple heuristic to choose between competing analyses. Density of an analysis = (filled roles) / (total roles) The system prefers higher density analyses because a higher density suggests that more frame roles are filled in than in competing analyses. Extremely simple / useful? but it certainly can be improved upon.

Summary: ECG Linguistic constructions are tied to a model of simulated action and perception Embedded in a theory of language processing –Constrains theory to be usable –Frees structures to be just structures, used in processing Precise, computationally usable formalism –Practical computational applications, like MT and NLU –Testing of functionality, e.g. language learning A shared theory and formalism for different cognitive mechanisms –Constructions, metaphor, mental spaces, etc.

Issues in Scaling up to Language Knowledge –Lexicon (FrameNet ) –Constructicon (ECG) –Maps (Metaphors, Metonymies) (MetaNet) –Conceptual Relations (Image Schemas, X-schemas) Computation –Representation (ECG) expressiveness, modularity, compositionality –Inference (Simulation Semantics) tractable, distributed, probabilistic concurrent, context-sensitive

The Buy schema schema Buy subcase of Action evokes Commercial-Transaction as ct roles self  ct.nucleus buyer  actor  ct.customer  ct.agent goods  undergoer  ct.goods

The Sell schema schema Sell subcase of Action evokes Commercial-Transaction as ct roles self  ct.nucleus seller  actor  ct.vendor  ct.agent goods  undergoer  ct.goods

Extending Inferential Capabilities Given the formalization of the conceptual schemas –How to use them for inferencing? Earlier pilot systems –Used metaphor and Bayesian belief networks –Successfully construed certain inferences –But don’t scale New approach –Probabilistic relational models –Support an open ontology

Semantic Web The World Wide Web (WWW) contains a large and expanding information base. HTML is accessible to humans but does not formally describe data in a machine interpretable form. XML remedies this by allowing for the use of tags to describe data (ex. disambiguating crawl) Ontologies are useful to describe objects and their inter- relationships. DAML+OIL ( is an markup language based on XML and RDF that is grounded in description logic and is designed to allow for ontology development, transfer, and use on the web.

The ICSI/Berkeley Neural Theory of Language Project Acquisition of early constructions ECG

Probabilistic Relation Inference Scalable Representation of –States, domain knowledge, ontologies (Avi Pfeffer 2000, Koller et al. 2001) Merges relational database technolgy with Probabilistic reasoning based on Graphical Models. –Domain entities and relational entities –Inter-entity relations are probabilistic functions –Can capture complex dependencies with both simple and composite slot (chains). Inference exploits structure of the domain

Status of PRMs Summer Project –Build the basic PRM codebase/infrastructure Fall Project –Design Coordinated PRM (CPRM) –Build Interface for testing Spring/Summer 03 –Implement CPRM to replace Pilot System DBN –Test CPRM for QA Related Work –Probabilistic OWL (PrOWL) –Probabilistic FrameNet

Articulating Projects FrameNet – NSF (with Colorado, USD) SmartKom – International Consortium EDU – European Media Lab Acquaint – ARDA (with SIMS, Stanford)

Conclusion NLU is essential to large, open domain QA. –Much of the web in unstructured data Substantial Progress in Enabling Technologies –Knowledge Representation/Inference Techniques Active Knowledge – X-schemas, Simulation Semantics Dealing With Uncertainty – PRM’s Combining Statistics and Structure. Conceptual Relations – Schemas, Metaphor, ECG –Scaling Up CYC, Wordnet, Term-bases FrameNet, Semantic Web, MetaNet Open Source The goal of NLU can be realized, perhaps! –Anyway, it’s time to try again.