Presentation is loading. Please wait.

Presentation is loading. Please wait.

A Best-Fit Approach for Productive Analysis of Omitted Arguments Eva Mok & John Bryant University of California, Berkeley International Computer Science.

Similar presentations


Presentation on theme: "A Best-Fit Approach for Productive Analysis of Omitted Arguments Eva Mok & John Bryant University of California, Berkeley International Computer Science."— Presentation transcript:

1 A Best-Fit Approach for Productive Analysis of Omitted Arguments Eva Mok & John Bryant University of California, Berkeley International Computer Science Institute

2 Simplify grammar by exploiting the language understanding process  Omission of arguments in Mandarin Chinese  Construction grammar framework  Model of language understanding  Our best-fit approach

3  Mother (I) give you this (a toy). CHILDES Beijing Corpus (Tardiff, 1993; Tardiff, 1996) ma1+magei3ni3zhei4+ge mothergive2PSthis+CLS  You give auntie [the peach].  Oh (go on)! You give [auntie] [that]. Productive Argument Omission (in Mandarin ) 1 2 3 ni3gei3yi2 2PSgiveauntie aoni3gei3ya EMP2PSgiveEMP 4 gei3 give  [I] give [you] [some peach].

4 Arguments are omitted with different probabilities All arguments omitted: 30.6%No arguments omitted: 6.1%

5 Construction grammar approach  Kay & Fillmore 1999; Goldberg 1995  Grammaticality: form and function  Basic unit of analysis: construction, i.e. a pairing of form and meaning constraints  Not purely lexically compositional  Implies early use of semantics in processing  Embodied Construction Grammar (ECG) (Bergen & Chang, 2005)

6 Problem: Proliferation of constructions SubjVerbObj1Obj2 ↓↓↓↓ GiverTransferRecipientTheme VerbObj1Obj2 ↓↓↓ TransferRecipientTheme … SubjVerbObj2 ↓↓↓ GiverTransferTheme SubjVerbObj1 ↓↓↓ GiverTransferRecipient

7 If the analysis process is smart, then...  The grammar needs only state one construction  Omission of constituents is flexibly allowed  The analysis process figures out what was omitted SubjVerbObj1Obj2 ↓↓↓↓ GiverTransferRecipientTheme

8 Best-fit analysis process takes burden off the grammar representation Constructions Simulation Utterance Discourse & Situational Context Semantic Specification: image schemas, frames, action schemas Analyzer: incremental, competition-based, psycholinguistically plausible

9 Competition-based analyzer finds the best analysis  An analysis is made up of:  A constructional tree  A set of resolutions  A semantic specification The best fit has the highest combined score

10 Combined score that determines best-fit  Syntactic Fit:  Constituency relations  Combine with preferences on non-local elements  Conditioned on syntactic context  Antecedent Fit:  Ability to find referents in the context  Conditioned on syntactic information, feature agreement  Semantic Fit:  Semantic bindings for frame roles  Frame roles’ fillers are scored

11 Analyzing ni3 gei3 yi2 (You give auntie)  Syntactic Fit:  P(Theme omitted | ditransitive cxn) = 0.65  P(Recipient omitted | ditransitive cxn) = 0.42 Two of the competing analyses: ni3gei3yi2omitted ↓↓↓↓ GiverTransferRecipientTheme ni3gei3omittedyi2 ↓↓↓↓ GiverTransferRecipientTheme (1-0.78)*(1-0.42)*0.65 = 0.08(1-0.78)*(1-0.65)*0.42 = 0.03

12 Using frame and lexical information to restrict type of reference Lexical Unit gei3 Giver (DNI) Recipient (DNI) Theme (DNI) The Transfer Frame Giver Recipient Theme Manner Means Place Purpose Reason Time

13 Can the omitted argument be recovered from context?  Antecedent Fit: ni3gei3yi2omitted ↓↓↓↓ GiverTransferRecipientTheme ni3gei3omittedyi2 ↓↓↓↓ GiverTransferRecipientTheme Discourse & Situational Context childmother peachauntie table ?

14 How good of a theme is a peach? How about an aunt? The Transfer Frame Giver (usually animate) Recipient (usually animate) Theme (usually inanimate) ni3gei3yi2omitted ↓↓↓↓ GiverTransferRecipientTheme ni3gei3omittedyi2 ↓↓↓↓ GiverTransferRecipientTheme  Semantic Fit:

15 The argument omission patterns shown earlier can be covered with just ONE construction  Each cxn is annotated with probabilities of omission  Language-specific default probability can be set SubjVerbObj1Obj2 ↓↓↓↓ GiverTransferRecipientTheme 0.780.420.65 P(omitted|cxn):

16 16 Research goal  A computationally-precise modeling framework for learning early constructions Language Data Linguistic Knowledge Learning

17 17 Frequent argument omission in pro-drop languages  Mandarin example: ni3 gei3 yi2 (“you give auntie”)  Even in English, there are often no spoken antecedents to pronouns in conversations Learner must integrate cues from intentions, gestures, prior discourse, etc

18 18 A short dialogue  bie2 mo3 wai4+tou2 a: #1_3 ! ( 別抹外頭啊 )  NEG-IMP apply forehead  Don ’ t apply [lotion to your] forehead  mo3 wai4+tou2 ke3 jiu4 bu4 hao3+kan4 le a:. ( 抹外頭可就不好看了啊 )  apply forehead LINKER LINKER NEG good looking CRS SFP  [If you] apply [lotion to your] forehead then [you will] not be pretty …  ze ya a: # bie2 gei3 ma1+ma wang3 lian3 shang4 mo:3 e: ! ( 嘖呀啊 # 別給媽媽往臉上抹呃 )  INTERJ # NEG-IMP BEN mother CV-DIR face on apply  INTERJ # Don ’ t apply [the lotion] on [your mom ’ s] face (for mom)  [- low pitch motherese] ma1+ma bu4 mo:3 you:2. ( 媽媽不抹油 )  mother NEG apply lotion  Mom doesn ’ t apply (use) lotion

19 Goals, refined  Demonstrate learning given  embodied meaning representation  structured representation of context  Based on  Usage-based learning  Domain-general statistical learning mechanism  Generalization / linguistic category formation 19

20 Towards a precise computational model  Modeling early grammar learning  Context model & Simulation  Data annotation  Finding the best analysis for learning  Hypothesizing and reorganizing constructions  Pilot results 20

21 Embodied Construction Grammar 21 construction yi2-N subcase of Morpheme form constraints self.f.orth <-- "yi2" meaning : @Aunt evokes RD as rd constraints self.m rd.referent self.m rd.ontological_category

22 “you” specifies discourse role 22 construction ni3-N subcase of Morpheme form constraints self.f.orth <-- "ni3" meaning : @Human evokes RD as rd constraints self.m rd.referent self.m rd.ontological_category rd.discourse_participant_role <-- @Addressee rd.set_size <-- @Singleton

23 The meaning of “give” is a schema with roles 23 construction gei3-V2 subcase of Morpheme form constraints self.f.orth <-- "gei3" meaning : Give schema Give subcase of Transfer constraints inherent_aspect <-- @Inherent_Achievement giver <-- @Animate recipient <-- @Animate theme <-- @Manipulable_Inanimate_Object schema Transfer subcase of Action roles giver : @Entity recipient : @Entity theme : @Entity constraints giver protagonist

24 Finally, you-give-aunt links up the roles 24 construction ni3-gei3-yi2 subcase of Finite_Clause constructional constituents n : ni3-N g : gei3-V2 y : yi2-N form constraints n.f meets g.f g.f meets y.f meaning : Give constraints self.m g.m self.m.giver n.m self.m.recipient y.m

25 The learning loop: Hypothesize & Reorganize 25 Linguistic Knowledge Discourse & Situational Context Analysis World Knowledge Context Fitting

26 26 XIXI INV Discourse Segment addresseespeaker If the learner has a ditransitive cxn meets ni3 Addressee giver Give gei3 recipient Aunt yi2 omitted theme Peach MOT

27 27 XIXI giver Give INV recipient Discourse Segment addresseespeaker Context fitting recovers more relations meets ni3 Addressee giver Give gei3 recipient Aunt yi2 omitted theme Peach MOT theme attentional- focus

28 28 giver Discourse Segment addresseespeaker Peach MOT theme attentional- focus recipient XIXI Give INV But the learner does not yet have phrasal cxns ni3 Addressee Give gei3 Aunt yi2 giver recipient meets

29 29 Context bootstraps learning ni3 Addressee Give gei3 Aunt yi2 meets construction ni3-gei3-yi2 subcase of Finite_Clause constructional constituents n : ni3 g : gei3 y : yi2 form constraints n.f meets g.f g.f meets y.f meaning : Give constraints self.m g.m self.m.giver n.m self.m.recipient y.m giver recipient

30 30 A model of context is key to learning  The context model makes it possible for the learning model to:  learn new constructions using contextually available information  learn argument-structure constructions in pro-drop languages

31 31 Understanding an utterance in context Schemas + Constructions Simulation Transcripts Events + Utterances Semantic Specification Analysis + Resolution Context Fitting Context Model Recency Model

32 32 Context model: Events + Utterances Setting participants, entities, & relations Setting participants, entities, & relations Start Event DS Sub-Event

33 33 Entities and Relations are instantiated Setting CHI, MOT (incl. body parts) livingroom(incl. ground, ceiling, chair, etc), lotion Setting CHI, MOT (incl. body parts) livingroom(incl. ground, ceiling, chair, etc), lotion Start apply02 applier = CHI substance = lotion surface = face(CHI) apply02 applier = CHI substance = lotion surface = face(CHI) ds04 admonishing05 speaker = MOT addressee = CHI forcefulness = normal ds04 admonishing05 speaker = MOT addressee = CHI forcefulness = normal caused_motion01 forceful_motionmotion caused_motion01 forceful_motionmotion translational_motion03 mover = lotion spg = SPG translational_motion03 mover = lotion spg = SPG

34 The context model is updated dynamically  Extended transcript annotation: speech acts & events  Simulator inserts events into context model & updates it with the effects  Some relations persists over time; some don’t. 34 Simulation Events Context Model Recency Model

35 35 Competition-based analyzer finds the best analysis  An analysis is made up of:  A constructional tree  A semantic specification  A set of resolutions Bill gaveMarythe book MaryBill Ref-Exp Give A-GIVE-B-X subj vobj1 obj2 book01@Man@WomanGive-Action @Book giver recipient theme

36 36 Combined score that determines best-fit  Syntactic Fit:  Constituency relations  Combine with preferences on non-local elements  Conditioned on syntactic context  Antecedent Fit:  Ability to find referents in the context  Conditioned on syntactic information, feature agreement  Semantic Fit:  Semantic bindings for frame roles  Frame roles’ fillers are scored

37 37 XIXI giver Give INV recipient Discourse Segment addresseespeaker Context Fitting goes beyond resolution meets ni3 Addressee giver Give gei3 recipient Aunt yi2 omitted theme Peach MOT theme attentional- focus

38 38 Context Fitting, a.k.a. intention reading  Context Fitting takes resolution a step further  considers entire context model, ranked by recency  considers relations amongst entities  heuristically fits from top down, e.g. discourse-related entities complex processes simple processes other structured and unstructured entities  more heuristics for future events (e.g. in cases of commands or suggestions)

39 39 Adult grammar size  ~615 constructions total  ~100 abstract cxns (26 to capture lexical variants)  ~70 phrasal/clausal cxns  ~440 lexical cxns (~260 open class)  ~195 schemas (~120 open class, ~75 closed class)

40 40 Starter learner grammar size  No grammatical categories (except interjections)  Lexical items only  ~440 lexical constructions  ~260 open class: schema / ontology meanings  ~40 closed class: pronouns, negation markers, etc  ~60 function words: no meanings  ~195 schemas (~120 open class, ~75 closed class)

41 41 The process hierarchy defined in schemas Process State Action State_ Change Complex_Process Proto_Transitive Intransitive_ State Two_Participant_ State Mental_State Joint_Motion Caused_Motion Concurrent_ Processes Cause_Effect Serial_Processes

42 42 The process hierarchy defined in schemas Action Translational_ Motion Translational_ Self_Motion Motion Intransitive_Action Expression Self_Motion Force_Application Continuous_ Force_Application Agentive_Impact Forceful_Motion

43 43 The process hierarchy defined in schemas Action Perception Ingestion Communication Transfer Cause_Change Other_ Transitive_Action Obtainment

44 44 Understanding an utterance in context Schemas + Constructions Simulation Transcripts Events + Utterances Semantic Specification Analysis + Resolution Context Fitting Context Model Recency Model

45 Hypothesize & Reorganize  Hypothesize:  utterance-driven;  relies on the analysis (SemSpec & context)  operations: compose  Reorganize:  grammar-driven;  can be triggered by usage (to be determined)  operations: generalize 45

46 Composing new constructions  Compose operation: If roles from different constructions point to the same context element, propose a new construction and set up a meaning binding. 46 ni3 Addressee Give gei3 giver recipient theme XIXI INV Peach MOT

47 Creating pivot constructions  Pivot generalization: Given a phrasal cxn, look for another cxn that shares 1+ constituents. Line up roles and bindings. Create new cxn category for the slot. 47 ni3 Addressee Give gei3 @Aunt yi2 giver recipient meets ni3 Addressee Give gei3 @Human wo3 giver recipient meets

48 48 Resulting constructions construction ni3-gei3-cat01 constituents ni3, gei3, cat01 meaning : Give constraints self.m.recipient g.m construction wo3 subcase of cat01 meaning: @Human construction yi2 subcase of cat01 meaning: @Aunt general construction cat01 subcase of Morpheme meaning: @Human

49 Pilot Results: Sample constructions learned  Composed:  Pivot Cxns: 49 chi1_fan4 ni3_chuan1_xie2 ni3_shuo1 bu4_na2 wo3_qu4 ni3_ping2zi_gei3_wo3 ni3_gei3_yi2 wo3_bu4_chi1 eat rice you wear shoe you say NEG take I go you bottle give me you give aunt I NEG eat ni3 {shuo1, chuan1} ni3 {shuo1, hua4} wo3 {zhao3, qu4} bu4 {na2, he1} {wo3, ma1} cheng2 you {say, wear} you {say, draw} I {find, go} NEG {take, drink} {I, mom} scoop

50 Challenge #1: Non-compositional meaning  Non-compositional meaning: Search for additional meaning schemas (in context or in general) that relate the meanings of the individual constructions 50 you Addressee Bake bake baker baked Bake-Event CHI Cake MOT a cake @Cake Give-Event

51 Challenge #2: Function words  Function words tend to indicate relations rather than events or entities 51 you Addressee Bake bake baker baked Bake-Event CHI Cake MOT a cake @Cake for Benefaction

52 Challenge #3: How far up to generalize  Eat rice  Eat apple  Eat watermelon  Want rice  Want apple  Want chair 52 Inanimate Object Manipulable Objects Manipulable Objects Unmovable Objects Food Furniture Fruit Savory Chair Sofa apple watermelon rice

53 Challenge #4: Beyond pivot constructions  Pivot constructions: indexing on particular constituent type Eat rice; Eat apple; Eat watermelon  Abstract constructions: indexing on role-filler relations between constituents 53 Schema Eat roles eater agent food patient food Eat catX Schema Want roles wanter agent wanted patient wanted Want catY

54 Challenge #5: Omissible constituents  Intuition:  Same context, two expressions that differ by one constituent  a general construction with the constituent being omissible  May require verbatim memory traces of utterances + “relevant” context 54

55 When does the learning stop?  Most likely grammar given utterances and context  The grammar prior is a preference for the “kind” of grammar  In practice, take the log and minimize cost  Minimum Description Length (MDL) 55 Bayesian Learning Framework Schemas + Constructions SemSpec Analysis +Resolution Context Fitting

56 Intuition for MDL  S -> Give me NP  NP -> the book  NP -> a book  S -> Give me NP  NP -> DET book  DET -> the  DET -> a 56 Suppose that the prior is inversely proportional to the size of the grammar (e.g. number of rules) It’s not worthwhile to make this generalization

57 Intuition for MDL  S -> Give me NP  NP -> the book  NP -> a book  NP -> the pen  NP -> a pen  NP -> the pencil  NP -> a pencil  NP -> the marker  NP -> a marker  S -> Give me NP  NP -> DET N  DET -> the  DET -> a  N -> book  N -> pen  N -> pencil  N -> marker 57

58 How to calculate the prior of this grammar  (Yet to be determined)  There is evidence that the lexicalized constructions do not completely go away  If the more lexicalized constructions are retained, the size of grammar is a bad indication of degree of generality 58


Download ppt "A Best-Fit Approach for Productive Analysis of Omitted Arguments Eva Mok & John Bryant University of California, Berkeley International Computer Science."

Similar presentations


Ads by Google