Jamie Alexandre. ≠ = would you like acookie jason.

Slides:



Advertisements
Similar presentations
Flexible Shaping: How learning in small steps helps Hierarchical Organization of Behavior, NIPS 2007 Kai Krueger and Peter Dayan Gatsby Computational Neuroscience.
Advertisements

Chapter 11 user support. Issues –different types of support at different times –implementation and presentation both important –all need careful design.
Chapter 4 Key Concepts.
What ’ s New? Acquiring New Information as a Process in Comprehension Suan E. Haviland & Herbert H. Clark.
Introduction to Computational Natural Language Learning Linguistics (Under: Topics in Natural Language Processing ) Computer Science (Under:
Artificial Grammar Learning (AGL) First developed by Reber in 1967 Standard procedure: Subjects are shown a series of letter strings that follow particular.
All slides © S. J. Luck, except as indicated in the notes sections of individual slides Slides may be used for nonprofit educational purposes if this copyright.
Lecture # 7 Chapter 4: Syntax Analysis. What is the job of Syntax Analysis? Syntax Analysis is also called Parsing or Hierarchical Analysis. A Parser.
Learning linguistic structure with simple recurrent networks February 20, 2013.
For Monday Read Chapter 23, sections 3-4 Homework –Chapter 23, exercises 1, 6, 14, 19 –Do them in order. Do NOT read ahead.
Computational Analysis of Motor Learning. Three paradigms Force field adaptation Visuomotor transformations Sequence learning Does one term (motor learning)
Connectionist Simulation of the Empirical Acquisition of Grammatical Relations – William C. Morris, Jeffrey Elman Connectionist Simulation of the Empirical.
PS: Introduction to Psycholinguistics Winter Term 2005/06 Instructor: Daniel Wiechmann Office hours: Mon 2-3 pm Phone:
9.012 Brain and Cognitive Sciences II Part VIII: Intro to Language & Psycholinguistics - Dr. Ted Gibson.
Help and Documentation zUser support issues ydifferent types of support at different times yimplementation and presentation both important yall need careful.
Tom Griffiths CogSci C131/Psych C123 Computational Models of Cognition.
Implicit learning Zoltán Dienes Conscious and unconscious mental processes.
Help and Documentation CSCI324, IACT403, IACT 931, MCS9324 Human Computer Interfaces.
1 CONTEXT-FREE GRAMMARS. NLE 2 Syntactic analysis (Parsing) S NPVP ATNNSVBD NP AT NNthechildrenate thecake.
Models of Generative Grammar Smriti Singh. Generative Grammar  A Generative Grammar is a set of formal rules that can generate an infinite set of sentences.
Lecture 1 Introduction: Linguistic Theory and Theories
Statistical Natural Language Processing. What is NLP?  Natural Language Processing (NLP), or Computational Linguistics, is concerned with theoretical.
Conclusions Funding: NIH R01DC  The ERP findings for the language task are not surprising given that the P600 component has often been evoked by.
The mental representation of sentences Tree structures or state vectors? Stefan Frank
1 Lecture 16 - Chapter 11 User support Issues –different types of support at different times –implementation and presentation both important –all need.
Modelling Language Evolution Lecture 2: Learning Syntax Simon Kirby University of Edinburgh Language Evolution & Computation Research Unit.
Tree Kernels for Parsing: (Collins & Duffy, 2001) Advanced Statistical Methods in NLP Ling 572 February 28, 2012.
For Friday Finish chapter 23 Homework: –Chapter 22, exercise 9.
Adaptor Grammars Ehsan Khoddammohammadi Recent Advances in Parsing Technology WS 2012/13 Saarland University 1.
THE BIG PICTURE Basic Assumptions Linguistics is the empirical science that studies language (or linguistic behavior) Linguistics proposes theories (models)
Connectionist Models of Language Development: Grammar and the Lexicon Steve R. Howell McMaster University, 1999.
1 Compiler Construction (CS-636) Muhammad Bilal Bashir UIIT, Rawalpindi.
University of Windsor School of Computer Science Topics in Artificial Intelligence Fall 2008 Sept 11, 2008.
1 CSC 9010 Spring Paula Matuszek CSC 9010 ANN Lab Paula Matuszek Spring, 2011.
Distributed Representative Reading Group. Research Highlights 1Support vector machines can robustly decode semantic information from EEG and MEG 2Multivariate.
Material from Authors of Human Computer Interaction Alan Dix, et al
Chap#11 What is User Support?
Cognitive Processes PSY 334 Chapter 11 – Language Structure June 2, 2003.
Artificial Intelligence: Natural Language
Evaluating Models of Computation and Storage in Human Sentence Processing Thang Luong CogACLL 2015 Tim J. O’Donnell & Noah D. Goodman.
CPSC 422, Lecture 27Slide 1 Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 27 Nov, 16, 2015.
1 Chapter 4 Syntax Part III. 2 The infinity of language pp The number of sentences in a language is infinite. 2. The length of sentences is.
LEARNING Prof.Elham Aljammas May 2015 L3. Relatively permanent change in behavior as a result of prior experience Tasks used to study the phenomenon can.
CS : Speech, NLP and the Web/Topics in AI Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture-15: Probabilistic parsing; PCFG (contd.)
Chapter 2: The Cognitive Science Approach
Dependency Parsing Niranjan Balasubramanian March 24 th 2016 Credits: Many slides from: Michael Collins, Mausam, Chris Manning, COLNG 2014 Dependency Parsing.
Expert System / Knowledge-based System Dr. Ahmed Elfaig 1.ES can be defined as computer application program that makes decision or solves problem in a.
Implicit an Explicit Motor Learning Two equal routes, or is one better than the other?
Chapter 2. Formal Languages Dept. of Computer Engineering, Hansung University, Sung-Dong Kim.
Chapter 11 user support. Overview Users require different types of support at different times. There are four main types of assistance that users require:
Roadmap Probabilistic CFGs –Handling ambiguity – more likely analyses –Adding probabilities Grammar Parsing: probabilistic CYK Learning probabilities:
General Information on Context-free and Probabilistic Context-free Grammars İbrahim Hoça CENG784, Fall 2013.
Natural Language Processing Vasile Rus
Human Computer Interaction Lecture 21 User Support
Neural representation and decoding of the meanings of words
Learning linguistic structure with simple and more complex recurrent neural networks Psychology February 2, 2017.
Statistical NLP Winter 2009
Simple recurrent networks.
SYNTAX.
James L. McClelland SS 100, May 31, 2011
Chapter Eight Syntax.
Formal Language Theory
A First Look at Music Composition using LSTM Recurrent Neural Networks
Learning linguistic structure with simple and more complex recurrent neural networks Psychology February 8, 2018.
Modelling Language Evolution Lecture 3: Evolving Syntax
Chapter 11 user support.
Learning linguistic structure with simple recurrent neural networks
David Kauchak CS159 – Spring 2019
CS249: Neural Language Model
Statistical NLP Winter 2009
Presentation transcript:

Jamie Alexandre

≠ =

would you like acookie jason

Grammatical Complexity The Chomsky Hierarchy

Recursion Something containing an instance of itself.

Recursion in Language The dog walked down the street. The dog the cat rode walked down the street. The dog the cat the rat grabbed rode walked down the street.

Recursion: “Stack” Memory The dog the cat the rat grabbed rode walked down the street. DOGCATRATWALKRIDEGRAB

Recursion: “Stack” Memory The dog the cat the rat grabbed rode walked down the street. DOG CAT RAT WALKRIDEGRAB “Limited performance…” “Infinite competence…”

?

SRN Simple Recurrent Network (Elman, 1990) Some ability to use longer contexts Incremental learning: no looking back No “rules”: distributed representation

PCFG Easily handles recursive structure, long-range context Hierarchical, “rule”-based representation More computationally complex, non-incremental learning Probabilistic Context-Free Grammar S  NP VP N’  AdjP N’ N’  N Adj  green … …

Serial Reaction Time (SRT) Study Buttons flash in short sequences –“press the button as quickly as possible when it lights up” Dependent measure: RT –time from light on  correct button pressed Subjects seem to be making sequential predictions RT ∝ P(button|context) also: RT ∝ -log(P(button|context)) (“surprisal”, e.g. Hale, 2001; Levy, 2008)

Training the Humans Eight subjects per experimental condition Same sequences, different mappings Broken into 16 blocks, with breaks About an hour of button-pressing total Emphasized speed, while minimizing errors

Training the Models Trained on exactly the same sequences as the humans, but not fit to human data Predictions at every point based solely on sequences seen prior to that Results in sequence of probabilities –correlated with sequence of human RTs, through surprisal (negative log probability)

Analysis

A Case Study in Recursion: Palindromes A C L Q L C A (Sequences of length 5 through 15; total of 3728 trials per subject)

PCFG SRN PCFG SRN “Did you notice any patterns?” Subjects with no awareness of pattern: “No”, “None”, “Not really” (n=5) Those with explicit awareness of pattern: “Circular pattern”, “Mirror pattern” (n=3) SRN (implicit task performance) PCFG (explicit task performance) Will this replicate?

Block Correlation (Surprisal vs RT) Implicit, didn't notice (n=8) PCFG SRN

Differences between individuals? –or actually between modes of processing? What if we explicitly train subjects on the pattern? First half implicit, second half explicit

“ This is the middle button in every sequence (and it only occurs in the middle position, halfway through the sequence): This means that as soon as you see this button, you know that the sequence will start to reverse. Here are some example sequences of various lengths: Explicit Training Worksheet

And Quiz Sheet “ Now, complete these sequences using the same pattern (crossing out any unneeded boxes at the end of a sequence):

Block Correlation (Surprisal vs RT) Fully explicit from middle (n=8) PCFG SRN (explicit instruction given here)

Before explicit instruction After

Context-free vs Context-sensitive A  A B  B C  C D  D

CFG: CSG: Explicit Instruction (after block 4)

Methods Four conditions, with 8 subjects in each –Implicit context-free grammar (CFG) –Implicit context-sensitive grammar (CSG) –Explicit context-free grammar (CFG) –Explicit context-sensitive grammar (CSG) Total of 640 sequences (4,120 trials) per subject –Sequences of length 4, 6, 8, and 10 –Around 1.5 hours of button-pressing –In blocks 9-16, 5% of the trials were “errors” A 1 B 1 C 1 C 2 B 2 A 2 D 2

Blocks 1-4 Blocks 5-8 Blocks 9-12 (errors thicker) Blocks (errors thicker)

** (6ms) ** (27ms)(2ms) ** (11ms) RT (ms)

(1ms) ** (6ms) ** (7ms) ** (3ms)

Conclusions Explicit/Implicit processing –Implicit performance correlated with the predictions made by an SRN (a connectionist model) –Explicit performance correlated with the predictions made by a PCFG (a rule-based model) Grammatical complexity –Able to process context-free, recursive structures at a very rapid timescale –More limited ability to process context-sensitive structures

Longer training More complex grammars –Determinism Other response measures –EEG: more sensitive than RTs to initial stages of learning Field studies in Switzerland or Brazil…? Future Directions

Broader Goals L2-learning pedagogy

Thankyous! Mentorship Jeff Elman Roger Levy Marta Kutas Advice Micah Bregman Ben Cipollini Vicente Malave Nathaniel Smith Angela Yu Rachel Mayberry Tom Urbach Andrea, Seana and the 3 rd Year Class! Research Assistants Frances Martin (2010) Ryan Cordova (2009) Wai Ho Chiu (2009)

Palindromes

AGL and Language Areas associated with syntax may be involved –Bahlmann, Schubotz, and Friederici (2008). Hierarchical artificial grammar processing engages Broca's area. NeuroImage, 42(2): P600-like effects can be seen in AGL –Christiansen, Conway, & Onnis (2007). Neural Responses to Structural Incongruencies in Language and Statistical Learning Point to Similar Underlying Mechanisms. –“violations in an artificial grammar can elicit late positivities qualitatively and topographically comparable to the P600 seen with syntactic violations in natural language”

Sanity Check: Effect is Local

Context-free Grammar The dog the cat the rat grabbed rode walked. S  NP VP NP  N NP  N S N  the dog N  the cat N  the rat VP  grabbed VP  rode VP  walked