The Neural Basis of Thought and Language Week 14.

Slides:



Advertisements
Similar presentations
CILC2011 A framework for structured knowledge extraction and representation from natural language via deep sentence analysis Stefania Costantini Niva Florio.
Advertisements

Statistical Machine Translation Part II: Word Alignments and EM Alexander Fraser ICL, U. Heidelberg CIS, LMU München Statistical Machine Translation.
GRAMMAR & PARSING (Syntactic Analysis) NLP- WEEK 4.
The Neural Basis of Thought and Language Final Review Session.
CS 182 Sections slides created by: Eva Mok modified by jgm April 26, 2006.
Embodied Models of Language Learning and Use Embodied language learning Nancy Chang UC Berkeley / International Computer Science Institute.
Models of Grammar Learning CS 182 Lecture April 26, 2007.
Constructing Grammar: a computational model of the acquisition of early constructions CS 182 Lecture April 25, 2006.
Grammar induction by Bayesian model averaging Guy Lebanon LARG meeting May 2001 Based on Andreas Stolcke’s thesis UC Berkeley 1994.
Part of speech (POS) tagging
Latent Semantic Analysis (LSA). Introduction to LSA Learning Model Uses Singular Value Decomposition (SVD) to simulate human learning of word and passage.
Artificial Intelligence Research Centre Program Systems Institute Russian Academy of Science Pereslavl-Zalessky Russia.
Natural Language Understanding
A Survey for Interspeech Xavier Anguera Information Retrieval-based Dynamic TimeWarping.
Learning Patterns on the World Wide Web Andrew Hogue Advisor: David Karger October 17, 2003.
SEMANTIC ANALYSIS WAES3303
Terminology and documentation*  Object of the study of terminology:  analysis and description of the units representing specialized knowledge in specialized.
Artificial Intelligence Research Center Pereslavl-Zalessky, Russia Program Systems Institute, RAS.
1 Compiler Design (40-414)  Main Text Book: Compilers: Principles, Techniques & Tools, 2 nd ed., Aho, Lam, Sethi, and Ullman, 2007  Evaluation:  Midterm.
Artificial Intelligence: Natural Language
For Monday Read chapter 26 Last Homework –Chapter 23, exercise 7.
Jane Reid, AMSc IRIC, QMUL, 30/10/01 1 Information seeking Information-seeking models Search strategies Search tactics.
For Friday Finish chapter 23 Homework –Chapter 23, exercise 15.
Supertagging CMSC Natural Language Processing January 31, 2006.
1 Minimum Error Rate Training in Statistical Machine Translation Franz Josef Och Information Sciences Institute University of Southern California ACL 2003.
CS 4705 Lecture 17 Semantic Analysis: Robust Semantics.
For Monday Read chapter 26 Homework: –Chapter 23, exercises 8 and 9.
Overview of Statistical NLP IR Group Meeting March 7, 2006.
BRAIN SCAN  Brain scan is an interactive quiz for use as a revision/ learning reinforcement tool that accompanies the theory package.  To answer a question.
11 Thoughts on STS regarding Machine Reading Ralph Weischedel 12 March 2012.
The Neural Basis of Thought and Language Final Review Session.
The Neural Basis of Thought and Language Final Review Session.
slides derived from those of Eva Mok and Joe Makin March 21, 2007
The Neural Basis of Thought and Language Week 14.
CS 182 Sections Leon Barrett with slides inspired by Eva Mok and Joe Makin April 18, 2007.
CS182 Leon Barrett with slides inspired by Eva Mok and Joe Makin April 25, 2008.
CS 182 Sections slide credit to Eva Mok and Joe Makin Updated by Leon Barrett April 25, 2007.
Teaching for Mastery in Maths
Statistical Machine Translation Part II: Word Alignments and EM
Review course concepts
PSYC 206 Lifespan Development Bilge Yagmurlu.
Compiler Design (40-414) Main Text Book:
Assessing Grammar Module 5 Activity 5.
CSC 594 Topics in AI – Natural Language Processing
How to Write a Constructed and Visual Response for the EOC
Course Organization.
Basic Parsing with Context Free Grammars Chapter 13
Kenneth Baclawski et. al. PSB /11/7 Sa-Im Shin
slides derived from those of Eva Mok and Joe Makin March 21, 2007
Natural Language Processing (NLP)
Robust Semantics, Information Extraction, and Information Retrieval
We have defined what the student should be able to do.
Data Mining Lecture 11.
Lecture 12: Data Wrangling
CSc4730/6730 Scientific Visualization
School of Library and Information Science
HCC class lecture 13 comments
CSCI 5832 Natural Language Processing
Authors: Wai Lam and Kon Fan Low Announcer: Kyu-Baek Hwang
Information Networks: State of the Art
AICE General Paper What IS this class?.
Deputy Commissioner Jeff Wulfson Associate Commissioner Michol Stapel
Year 2 – Reading Fiction Session
A User study on Conversational Software
Natural Language Processing (NLP)
Lecture 5 Scanning.
Artificial Intelligence 2004 Speech & Natural Language Processing
Theoretical approaches to helping children to learn to read:
AICE English General Paper
Natural Language Processing (NLP)
Presentation transcript:

The Neural Basis of Thought and Language Week 14

Administrivia Final exam review session Monday –7-9pm in 310 Soda Final in class next Thursday, May 8 th Be there on time! Format: –closed books, closed notes –short answers, no blue books Final paper due on bSpace on Monday, May 11

Questions 1.How does the analyzer use the constructions to parse a sentence? 2.How can we learn new ECG constructions? 3.What are ways to re-organize and consolidate the current grammar? 4.What metric is used to determine when to form a new construction?

“you” you schema Addressee subcase of Human FORM (sound) MEANING (stuff) Analyzing “You Throw The Ball” “throw” throw schema Throw roles: thrower throwee “ball” ball schema Ball subcase of Object “block” block schema Block subcase of Object t1 before t2 t2 before t3 Thrower-Throw-Object t2.thrower ↔ t1 t2.throwee ↔ t3 “the” Addressee Throw thrower throwee Ball

Do not forget the SemSpec!

Analyzing in ECG create a recognizer for each construction in the grammar for each level j (in ascending order) repeat for each recognizer r in j for each position p of utterance initiate r starting at p until we don't find anything new

Recognizer for the Transitive-Cn an example of a level-1 construction is Red-Ball-Cn each recognizer looks for its constituents in order (the ordering constraints on the constituents can be a partial ordering) ag t vobj I ge t cu p level 0 level 2 level 1

Questions 1.How does the analyzer use the constructions to parse a sentence? 2.How can we learn new ECG constructions? 3.What are ways to re-organize and consolidate the current grammar? 4.What metric is used to determine when to form a new construction?

Constructions (Utterance, Situation) 1.Learner passes input (Utterance + Situation) and current grammar to Analyzer. Analyze Semantic Specification, Constructional Analysis 2.Analyzer produces SemSpec and Constructional Analysis. 3. Learner updates grammar: Hypothesize a.Hypothesize new map. Reorganize b.Reorganize grammar (merge or compose). c.Reinforce (based on usage). Learning-Analysis Cycle (Chang, 2004)

Acquisitio n Reorganize Hypothesize Productio n Utteranc e (Comm. Intent, Situation) Generate Constructions (Utterance, Situation) Analysis Comprehension Analyze Partia l Usage-based Language Learning

Basic Learning Idea The learner’s current grammar produces a certain analysis for an input sentence The context contains richer information (e.g. bindings) that are unaccounted for in the analysis Find a way to account for these meaning relations (by looking for corresponding form relations) SemSpecr1r2r1r2r3r1r2Contextr1r2r1r2r3r1r2

“you” “throw” “ball” you throw ball “block” block schema Addressee subcase of Human FORM (sound) MEANING (stuff) lexical constructions Initial Single-Word Stage schema Throw roles: thrower throwee schema Ball subcase of Object schema Block subcase of Object

“you” you schema Addressee subcase of Human FORMMEANING New Data: “You Throw The Ball” “throw” throw schema Throw roles: thrower throwee “ball” ball schema Ball subcase of Object “block” block schema Block subcase of Object “the” Addressee Throw thrower throwee Ball Self SITUATION Addressee Throw thrower throwee Ball before role-filler throw-ball

Relational Mapping Scenarios AfAf BfBf AmAm BmBm A B form- relatio n role - fille r throw ball throw.throwee ↔ ball AfAf BfBf AmAm BmBm A B form- relatio n role - fille r X put ball down put.mover ↔ ball down.tr ↔ ball AfAf BfBf AmAm BmBm A B form- relatio n role - fille r Y Nomi ball possession.possessor ↔ Nomi possession.possessed ↔ ball

Questions 1.How does the analyzer use the constructions to parse a sentence? 2.How can we learn new ECG constructions? 3.What are ways to re-organize and consolidate the current grammar? 4.What metric is used to determine when to form a new construction?

Merging Similar Constructions throw before block Throw.throwee = Block throw before ball Throw.throwee = Ball throw before-s ing Throw.aspect = ongoing throw-ing the ball throw the block throw before Object f THROW.throwee = Object m THROW-OBJECT

Resulting Construction construction THROW-OBJECT constructional constituents t : THROW o : OBJECT form t f before o f meaning t m.throwee ↔ o m

Composing Co-occurring Constructions ball before off Motion m m.mover = Ball m.path = Off ball off throw before ball Throw.throwee = Ball throw the ball throw before ball ball before off THROW.throwee = Ball Motion m m.mover = Ball m.path = Off THROW-BALL-OFF

Resulting Construction construction THROW-BALL-OFF constructional constituents t : THROW b : BALL o : OFF form t f before b f b f before o f meaning evokes MOTION as m t m.throwee ↔ b m m.mover ↔ b m m.path ↔ o m

Questions 1.How does the analyzer use the constructions to parse a sentence? 2.How can we learn new ECG constructions? 3.What are ways to re-organize and consolidate the current grammar? 4.What metric is used to determine when to form a new construction?

Size Of Grammar Size of the grammar G is the sum of the size of each construction: Size of each construction c is: where n c = number of constituents in c, m c = number of constraints in c, length(e) = slot chain length of element reference e

Example: The Throw-Ball Cxn construction THROW-BALL constructional constituents t : THROW b : BALL form t f before b f meaning t m.throwee ↔ b m size ( THROW-BALL ) = (2 + 3) = 9

Complexity of Data Given Grammar Complexity of the data D given grammar G is the sum of the analysis score of each input token d: Analysis score of each input token d is: where c is a construction used in the analysis of d weight c ≈ relative frequency of c, |type r | = number of ontology items of type r used, height d = height of the derivation graph, semfit d = semantic fit provide by the analyzer

Minimum Description Length Choose grammar G to minimize cost(G|D): –cost(G|D) = α size(G) + β complexity(D|G) –Approximates Bayesian learning; cost(G|D) ≈ 1/posterior probability ≈ 1/P(G|D) Size of grammar = size(G) ≈ 1/prior ≈ 1/P(G) –favor fewer/smaller constructions/roles; isomorphic mappings Complexity of data given grammar ≈ 1/likelihood ≈ 1/P(D|G) –favor simpler analyses (fewer, more likely constructions) –based on derivation length + score of derivation

Final Remark The goal here is to build a cognitive plausible model of language learning A very different game that one could play is unsupervised / semi-supervised language learning using shallow or no semantics –statistical NLP –automatic extraction of syntactic structure –automatic labeling of frame elements Fairly reasonable results for use in tasks such as information retrieval, but the semantic representation is very shallow