The Neural Basis of Thought and Language Final Review Session.

Slides:



Advertisements
Similar presentations
CSCTR Session 8 Dana Retová. group at UC Berkeley & Uni of Hawaii Nancy Chang Benjamin Bergen Jerome Feldman, … General assumption Semantic relations.
Advertisements

Big Ideas in Cmput366. Search Blind Search State space representation Iterative deepening Heuristic Search A*, f(n)=g(n)+h(n), admissible heuristics Local.
BAYESIAN NETWORKS. Bayesian Network Motivation  We want a representation and reasoning system that is based on conditional independence  Compact yet.
Embodied Construction Grammar in language (acquisition and) use Jerome Feldman Computer Science Division, University of California,
Embodied Compositional Semantics Ellen Dodge
ISBN Chapter 3 Describing Syntax and Semantics.
NLP and Speech Course Review. Morphological Analyzer Lexicon Part-of-Speech (POS) Tagging Grammar Rules Parser thethe – determiner Det NP → Det.
Multiobjective VLSI Cell Placement Using Distributed Simulated Evolution Algorithm Sadiq M. Sait, Mustafa I. Ali, Ali Zaidi.
CS 182 Sections slides created by Eva Mok modified by JGM March 22, 2006.
The Neural Basis of Thought and Language Final Review Session.
Unified Cognitive Science Neurobiology Psychology Computer Science Linguistics Philosophy Social Sciences Experience Take all the Findings and Constraints.
CS 182 Sections slides created by Eva Mok modified by jgm April 13, 2005.
NTL – Converging Constraints Basic concepts and words derive their meaning from embodied experience. Abstract and theoretical concepts derive their meaning.
CS 182 Sections slides created by: Eva Mok modified by jgm April 26, 2006.
CS 182 Sections Eva Mok April 21, 2004.
1/31 CS 426 Senior Projects Chapter 1: What is UML? Chapter 2: What is UP? [Arlow and Neustadt, 2005] January 22, 2009.
Describing Syntax and Semantics
PCFG Based Synthetic Mobility Trace Generation S. C. Geyik, E. Bulut, and B. K. Szymanski Department of Computer Science, Rensselaer Polytechnic Institute.
CS Bayesian Learning1 Bayesian Learning. CS Bayesian Learning2 States, causes, hypotheses. Observations, effect, data. We need to reconcile.
EE513 Audio Signals and Systems Statistical Pattern Classification Kevin D. Donohue Electrical and Computer Engineering University of Kentucky.
NTL – Converging Constraints Basic concepts and words derive their meaning from embodied experience. Abstract and theoretical concepts derive their meaning.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 26 of 41 Friday, 22 October.
Bayesian Learning By Porchelvi Vijayakumar. Cognitive Science Current Problem: How do children learn and how do they get it right?
A Cognitive Substrate for Natural Language Understanding Nick Cassimatis Arthi Murugesan Magdalena Bugajska.
Introduction to Embodied Construction Grammar March 4, 2003 Ben Bergen
CS 182 Sections slides created by Eva Mok modified by JGM March
Overview of the final test for CSC Overview PART A: 7 easy questions –You should answer 5 of them. If you answer more we will select 5 at random.
Rules, Movement, Ambiguity
Bayesian Optimization Algorithm, Decision Graphs, and Occam’s Razor Martin Pelikan, David E. Goldberg, and Kumara Sastry IlliGAL Report No May.
EEL 5937 Content languages EEL 5937 Multi Agent Systems Lecture 10, Feb. 6, 2003 Lotzi Bölöni.
Artificial Intelligence Knowledge Representation.
10/31/00 1 Introduction to Cognitive Science Linguistics Component Topic: Formal Grammars: Generating and Parsing Lecturer: Dr Bodomo.
The Neural Basis of Thought and Language Final Review Session.
The Neural Basis of Thought and Language Week 11 Metaphor and Automatic Reasoning.
CS 182 Sections Leon Barrett with acknowledgements to Eva Mok and Joe Makin April 4, 2007.
slides derived from those of Eva Mok and Joe Makin March 21, 2007
The Neural Basis of Thought and Language Week 14.
CS 182 Sections Leon Barrett with slides inspired by Eva Mok and Joe Makin April 18, 2007.
The Neural Basis of Thought and Language Week 14.
CS182 Leon Barrett with slides inspired by Eva Mok and Joe Makin April 25, 2008.
Announcements a3 is out, due 2/13 and 2/22 11:59pm Computational: part 1 isn't too hard; part 2 is harder quiz will be graded in about a week.
The Neural Basis of Thought and Language Week 12 Metaphor and Automatic Reasoning, plus Grammars.
CS 182 Sections slide credit to Eva Mok and Joe Makin Updated by Leon Barrett April 25, 2007.
What is cognitive psychology?
Learning linguistic structure with simple and more complex recurrent neural networks Psychology February 2, 2017.
Chapter 4: Business Process and Functional Modeling, continued
“Perceptual Symbol Systems”
CS Fall 2015 (Shavlik©), Midterm Topics
What is UML? What is UP? [Arlow and Neustadt, 2005] October 5, 2017
ICS 280 Learning in Graphical Models
slides derived from those of Eva Mok and Joe Makin March 21, 2007
with acknowledgments to Eva Mok and Joe Makin
Intelligent Information System Lab
Writing a Scientific Research Article
Data Mining Lecture 11.
Service-Oriented Computing: Semantics, Processes, Agents
Logical architecture refinement
Artificial Intelligence: Logic agents
CS4705 Natural Language Processing
EE513 Audio Signals and Systems
Class #19 – Tuesday, November 3
Class #16 – Tuesday, October 26
Learning linguistic structure with simple recurrent neural networks
6.001 SICP Interpretation Parts of an interpreter
CS 188: Artificial Intelligence Spring 2006
Neural Network Models in Vision
Toward a Great Class Project: Discussion of Stoianov & Zorzi’s Numerosity Model Psych 209 – 2019 Feb 14, 2019.
Programming Languages and Compilers (CS 421)
Chapter 14 February 26, 2004.
CS 188: Artificial Intelligence Fall 2008
Presentation transcript:

The Neural Basis of Thought and Language Final Review Session

Administrivia Final in class next Tuesday, May 8 th Be there on time! Format: –closed books, closed notes –short answers, no blue books Final paper due on bSpace on Friday, May 11

Resources Textbook! Class slides Section slides Joe Makin's class notes from last year –on notes page

The Second Half Cognition and Language Computation Structured Connectionism Computational Neurobiology Biology Midterm Final abstraction Motor Control Metaphor Grammar Bailey Model KARMA ECG SHRUTI ECG Learning Bayes Nets

Overview Bailey Model –feature structures –Bayesian model merging –recruitment learning KARMA –X-schema, frames –aspect –event-structure metaphor –inference Grammar Learning –parsing –construction grammar –learning algorithm SHRUTI FrameNet Bayesian Model of Human Sentence Processing

Important topics Regier's model of spatial relation learning Bailey's model of verb learning KARMA model of metaphor Binding and inference –SHRUTI, short signatures Grammars and learning ECG –Learning ECG Bayes nets Model merging, MDL Petri nets Language –Metaphor –Aspect –Grammars –Schemas –Frames –SemSpec

Q & A

Bayes Nets Bayes' Rule / Product Rule –P(A|B) = P(A,B) / P(B) –P(A,B) = P(A|B) P(B) –P(B|A) = P(A|B) P(B) / P(A) –All the same! Variables have distributions Variables depend on other variables

Regier's model Learn spatial relation terms –e.g. in, on, above Neural network + hand-designed “vision” parts

Bailey's model Verb learning Learn parameters matched to words –word senses –can connect to simulator Model merging!

6palmextendslide accelpostureelbow jntschema 8palmextendslide accelpostureelbow jntschema [6]palm 0.9extend 0.9slide 0.9 accelpostureelbow jntschema data #1 data #2 data #3 data #4 [2]index 0.9fixed 0.9depress 0.9 accelpostureelbow jntschema [6 - 8]palm 0.9extend 0.9slide 0.9 accelpostureelbow jntschema palm 0.7 grasp 0.3 extend 0.9slide 0.9 postureelbow jntschema 2indexfixeddepress accelpostureelbow jntschema 2graspextendslide accelpostureelbow jntschema

Computational Details complexity of model + ability to explain data maximum a posteriori (MAP) hypothesis how likely is the data given this model?penalize complex models – those with too many word senses wants the best model given data

Model merging Start with a simple model Merge to refine it –“Greedy” merges: reduce cost without thought for future Cost metric –prefer simple representation –prefer to explain data well

Metaphor There are LOTS of metaphors we use –Power is size –Knowing is seeing –Event structure is motion

Event Structure Metaphor States are Locations Changes are Movements Causes are Forces Causation is Forced Movement Actions are Self-propelled Movements Purposes are Destinations Means are Paths Difficulties are Impediments to Motion External Events are Large, Moving Objects Long-term, Purposeful Activities are Journeys

Ego Moving versus Time Moving

Results 30.8%69.2% Object Moving 73.3%26.7% Ego Moving Meeting is FridayMeeting is MondayPRIME

KARMA simulator Invented by Carson Daly Allows metaphor understanding –Event structure metaphor Source domain is Petri net –Target domain is Bayes net –Metaphor maps connect

KARMA DBN to represent target domain knowledge Metaphor maps link target and source domain X-schema to represent source domain knowledge

Temporal synchrony and SHRUTI Binding problem –bind properties to objects –don't mix them up! Reflexive reasoning –understand implied information –not conscious of this

SHRUTI entity, type, and predicate focal clusters An “entity” is a phase in the rhythmic activity. Bindings are synchronous firings of role and entity cells Rules are interconnection patterns mediated by coincidence detector circuits that allow selective propagation of activity An episode of reflexive processing is a transient propagation of rhythmic activity

asserting that walk(Harry, café) Harry fires in phase with agent role cafe fires in phase with goal role + - ? agtagt goalgoal +e+e +v+v ?e?e ?v?v +? walk cafe Harry type entit y predicate “Harry walked to the café.”

asserting that walk(Harry, café) Harry fires in phase with agent role cafe fires in phase with goal role + - ? agtagt goalgoal +e+e +v+v ?e?e ?v?v +? walk cafe Harry type entit y predicate “Harry walked to the café.”

Activation Trace for walk(Harry, café) +: Harry +: walk +e: cafe walk-agt walk-goal 1234

Alternative: short signatures + - ? agtagt goalgoal +e+e +v+v ?e?e ?v?v +? walk cafe Harry type entit y predicate = 010 = 110

Language Grammar –Syntax Tense Aspect Semantics Metaphor Simulation Unification

Computer-science style grammar Regular grammar –X -> a b c Y Context-free grammar –X -> a Y b Z W

Analysis Process Utterance Simulatio n Belief State General Knowledge Construction s Semantic Specification “Harry walked into the café.”

The INTO construction construction INTO subcase of Spatial-Relation form self f.orth ← “into” meaning: Trajector-Landmark evokes Container as cont evokes Source-Path-Goal as spg trajector ↔ spg.trajector landmark ↔ cont cont.interior ↔ spg.goal cont.exterior ↔ spg.source

Unification Basic idea: capture agreement and semantic features in feature structures Enforce constraints on these features using unification rules Pat number : SG person : 3rd agreement I number : SG person : 1st agreement Went agreement Shop agreement number : person : 1st VP  Verb NP VP.agreement ↔ Verb.agreement S  NP VP NP.agreement ↔ VP.agreement

The Spatial-Phrase construction construction S PATIAL- P HRASE constructional constituents sr : Spatial-Relation lm : Ref-Expr form sr f before lm f meaning sr m.landmark ↔ lm m

The Directed-Motion construction construction DIRECTED-MOTION constructional constituents a : Ref-Exp m: Motion-Verb p : Spatial-Phrase form a f before m f m f before p f meaning evokes Directed-Motion as dm self m.scene ↔ dm dm.agent ↔ a m dm.motion ↔ m m dm.path ↔ p m schema Directed-Motion roles agent : Entity motion : Motion path : SPG

What exactly is simulation? Belief update and/or X-schema execution hung ry meet ing cafe time of day ready start ongoing finish done iterate WAL K at goal

“Harry walked into the café.” read y wal k don e walker=Harr y goal=caf e

Analysis Process Utterance Simulatio n Belief State General Knowledge Construction s Semantic Specification “Harry is walking to the café.”

read y start ongoin g finis h don e iterat e abor t cancelle d interrup t resum e suspende d WALK walker=Harr y goal=caf e

Perhaps a different sense of INTO? construction INTO subcase of spatial-prep form self f.orth ← “into” meaning evokes Trajector-Landmark as tl evokes Container as cont evokes Source-Path-Goal as spg tl.trajector ↔ spg.trajector tl.landmark ↔ cont cont.interior ↔ spg.goal cont.exterior ↔ spg.source construction INTO subcase of spatial-prep form self f.orth ← “into” meaning evokes Trajector-Landmark as tl evokes Impact as im evokes Source-Path-Goal as spg tl.trajector ↔ spg.trajector tl.landmark ↔ spg.goal im.obj1 ↔ tl.trajector im.obj2 ↔ tl.landmark

“Harry has walked into the wall.” read y start ongoin g finis h don e iterat e abor t cancelle d interrup t resum e suspende d WALK walker=Harr y goal=wal l

Map down to timeline S R E read y start ongoin g finish done consequence

further questions?

How do you learn… the meanings of spatial relations, the meanings of verbs, the metaphors, and the constructions?

How do you learn… the meanings of spatial relations, the meanings of verbs, the metaphors, and the constructions? That’s the Regier model.

How do you learn… the meanings of spatial relations, the meanings of verbs, the metaphors, and the constructions? That's Bailey's model

How do you learn… the meanings of spatial relations, the meanings of verbs, the metaphors, and the constructions? conflation hypothesis (primary metaphors)

How do you learn… the meanings of spatial relations, the meanings of verbs, the metaphors, and the constructions? construction learning

Acquisitio n Reorganize Hypothesize Production Utterance (Comm. Intent, Situation) Generate Construction s (Utterance, Situation) Analysis Comprehensio n Analyze Partial Usage-based Language Learning

Main Learning Loop while available and cost > stoppingCriterion analysis = analyzeAndResolve(utterance, situation, currentGrammar); newCxns = hypothesize(analysis); if cost(currentGrammar + newCxns) < cost(currentGrammar) addNewCxns(newCxns); if (re-oganize == true) // frequency depends on learning parameter reorganizeCxns();

Three ways to get new constructions Relational mapping –throw the ball Merging –throw the block –throwing the ball Composing –throw the ball –ball off –you throw the ball off THROW < BALL < OFFTHROW < OBJECT THROW < BALL

Minimum Description Length Choose grammar G to minimize cost(G|D): –cost(G|D) = α size(G) + β complexity(D|G) –Approximates Bayesian learning; cost(G|D) ≈ posterior probability P(G|D) Size of grammar = size(G) ≈ 1/prior P(G) –favor fewer/smaller constructions/roles; isomorphic mappings Complexity of data given grammar ≈ 1/likelihood P(D|G) –favor simpler analyses (fewer, more likely constructions) –based on derivation length + score of derivation

further questions?