Introduction to Language Acquisition Theory Janet Dean Fodor St. Petersburg July 2013 Class 1. Language acquisition theory: Origins and issues.

Slides:



Advertisements
Similar presentations
Intro to NLP - J. Eisner1 Learning in the Limit Golds Theorem.
Advertisements

Natural Language Processing Syntax. Syntactic structure John likes Mary PN VtVt NP VP S DetPNVtVt NP VP S Every man likes Mary Noun.
How Children Acquire Language
Second Language Acquisition
Chapter 4 Key Concepts.
Psych 156A/ Ling 150: Acquisition of Language II Lecture 12 Poverty of the Stimulus I.
Main points of Interlanguage, Krashen, and Universal Grammar
Introduction: The Chomskian Perspective on Language Study.
Statistical Methods and Linguistics - Steven Abney Thur. POSTECH Computer Science NLP Lab Shim Jun-Hyuk.
Module 14 Thought & Language. INTRODUCTION Definitions –Cognitive approach method of studying how we process, store, and use information and how this.
Learning about Literacy: A 30-Year Journey By P
1 Introduction to Computational Natural Language Learning Linguistics (Under: Topics in Natural Language Processing ) Computer Science (Under:
Introduction to Computational Natural Language Learning Linguistics (Under: Topics in Natural Language Processing ) Computer Science (Under:
1 Introduction to Computational Natural Language Learning Linguistics (Under: Topics in Natural Language Processing ) Computer Science (Under:
Normal forms for Context-Free Grammars
Latent Semantic Analysis (LSA). Introduction to LSA Learning Model Uses Singular Value Decomposition (SVD) to simulate human learning of word and passage.
Fundamentals: Linguistic principles
Transformational Grammar p.33 - p.43 Jack October 30 th, 2012.
Formal Learning Theory Michele Friend (Philosophy) and Valentina Harizanov (Mathematics)
Models of Generative Grammar Smriti Singh. Generative Grammar  A Generative Grammar is a set of formal rules that can generate an infinite set of sentences.
Generative Grammar(Part ii)
Second language acquisition
Linguistic Theory Lecture 2 Phrase Structure. What was there before structure? Classical studies: Classical studies: –Languages such as Latin Rich morphology.
Linguistic Theory Lecture 3 Movement. A brief history of movement Movements as ‘special rules’ proposed to capture facts that phrase structure rules cannot.
The problem of sampling error in psychological research We previously noted that sampling error is problematic in psychological research because differences.
Main Branches of Linguistics
Emergence of Syntax. Introduction  One of the most important concerns of theoretical linguistics today represents the study of the acquisition of language.
Lecture 21: Languages and Grammars. Natural Language vs. Formal Language.
Cognitive Development: Language Infants and children face an especially important developmental task with the acquisition of language.
Language PERTEMUAN Communication Psycholinguistics –study of mental processes and structures that underlie our ability to produce and comprehend.
Evolution of Universal Grammar Pia Göser Universität Tübingen Seminar: Sprachevolution Dozent: Prof. Jäger
1 Words and rules Linguistics lecture #2 October 31, 2006.
Grammaticality Judgments Do you want to come with?20% I might could do that.38% The pavements are all wet.60% Y’all come back now.38% What if I were Romeo.
What is linguistics  It is the science of language.  Linguistics is the systematic study of language.  The field of linguistics is concerned with the.
Validity & Practicality
“On the Teaching and Learning of Grammar: Challenging the Myths” By Diane Larsen-Freeman.
THE BIG PICTURE Basic Assumptions Linguistics is the empirical science that studies language (or linguistic behavior) Linguistics proposes theories (models)
TEMPLATE DESIGN © Learning Words and Rules Abstract Knowledge of Word Order in Early Sentence Comprehension Yael Gertner.
Introduction to Language Acquisition Theory Janet Dean Fodor St. Petersburg July 2013 Class 2. From computer science (then) to psycholinguistics (now)
Psycholinguistic Theory
Theories of First Language Acquisition
First Language Acquisition
Language Acquisition By: Marissa Persinger Whitney Lewis Jessica Kline.
The Minimalist Program
1 LIN 1310B Introduction to Linguistics Prof: Nikolay Slavkov TA: Qinghua Tang CLASS 16, March 6, 2007.
First Language Acquisition
PSY 219 – Academic Writing in Psychology Fall Çağ University Faculty of Arts and Sciences Department of Psychology Inst. Nilay Avcı Week 9.
1 Introduction to Language Acquisition Theory Janet Dean Fodor St. Petersburg July 2013 Class 8. Implications and further questions Class 8. Implications.
Communicative Language Teaching CLT
Chapter 7 Linguistics aspect of interlanguage
Variability in Interlanguage Session 6. Variability Variability refers to cases where a second language learner uses two or more linguistic variants to.
Chapter 3 Language Acquisition: A Linguistic Treatment Jang, HaYoung Biointelligence Laborotary Seoul National University.
PSYCHOLINGUISTICS Grupo: Alejandro Ferrer Moya, Blanca Gallego Lopez,María Jose Garcia Ramón, Victoria Herrera Mercader, Soraya Lozano Carrión Grupo: Alejandro.
Interlanguage L. Selinker 2007 년 2 학기 담당교수 : 홍우평 이중언어커뮤니케이 션.
Poverty of Stimulus Poverty of Stimulus Reading Group.
Back to Board Welcome to Jeopardy!. Back to Board Today’s Categories~ ~ Cognitive Psychology ~ Solving Problems ~ Obstacles to Solving Problems ~ Language.
Child Syntax and Morphology
Welcome to Jeopardy!.
Biointelligence Laboratory, Seoul National University
What is Language Acquisition?
Theories of Language Development
William Gregory Sakas City University of New York (CUNY)
Over the past fifty years, three main theoretical positions have been advanced to explain language development from infancy through the early school years:
Over the past fifty years, three main theoretical positions have been advanced to explain language development from infancy through the early school years:
Competence and performance
Over the past fifty years, three main theoretical positions have been advanced to explain language development from infancy through the early school years:
Linguistic aspects of interlanguage
Learning to Communicate
Implementation of Learning Systems
Biointelligence Laboratory, Seoul National University
Presentation transcript:

Introduction to Language Acquisition Theory Janet Dean Fodor St. Petersburg July 2013 Class 1. Language acquisition theory: Origins and issues

Today’s class - Introduction  A broad sweep of ‘learnability’ research, as background.  Ultimate goal of this research : How is it possible for an infant, in just 5 or 6 years, to master the intricate system that is a human language? An extraordinary feat!  Origins and early discoveries  Interdisciplinary methods  Moving toward a closer relation with linguistics and psycholinguistics  In following classes we will go back over this material and consider it more carefully.  Feel free to ask as many questions as you like. But don’t feel pressured to understand everything today. 2

Real children, real language acquisition 3

E-children, models of language acquisition 4 How do children do it?

55 Goals & methods  Ultimate goal = a psychological model of the step by step processes by which all normally developing children create a rich mental grammar, based on what they hear people saying.  For adults, it’s a serious challenge  very mixed outcomes. For toddlers, essentially perfect outcomes guaranteed. How?  Interdisciplinary methods. All of these are needed:  Linguistics – what is a mental grammar like?  Developmental psychology – what stages of knowledge?  Psycholinguistics – how do children parse & comprehend sentences they hear?  Corpus search – what sentences do they hear?  Computational linguistics – simulation expts to test models.  Formal learning theory – what could be learned from input? 5

66 Early days: Mathematical theorems  Mathematical proofs concerning the relation between properties of languages and of the grammars that generate them. The ‘Chomsky hierarchy’. (Chomsky & Miller 1963)  The first learnability studies were also mathematical. No attempt at psychological plausibility. Just: How in principle could grammars be fitted to sets of sentences?  Proofs based on abstractions, far from actual human languages. Grammars designated as g1, g2, g3…; languages as constituted of an arbitrary set of sentences s1, s2, s3….  Gold’s Theorem (1967): A proof that no possible learning algorithm could acquire languages in the Chomsky hierarchy on the basis of ‘text presentation’ only (i.e., a sample of the sentences in the language; no info about non-sentences).  Gold’s Theorem (1967): A proof that no possible learning algorithm could acquire languages in the Chomsky hierarchy on the basis of ‘text presentation’ only (i.e., a sample of the sentences in the language; no info about non-sentences). 6

77 Psychological implications: innate knowledge Psychological implications: innate knowledge  Gold’s Theorem resonated with psychologists who had observed that children receive little or no negative evidence (about ungrammatical sentences in the target language).  Mostly anecdotes at first – examples in Class 3. But later, systematic documentation of the uninformativeness of adult responses to child errors. (Marcus 1993)  Gold’s Theorem thus fueled the conviction among many linguists & psycholinguists that learners must have innate information about what human languages are like, to compensate for missing information in their input.  Chomsky called it Universal Grammar (UG). UG has been a basis for much linguistic theory ever since.  More generally: the Poverty of the Stimulus = relevant evidence of all kinds is absent. (Classes 3 and 4.) 7

88 The Subset Problem – still central to language acquisition theory  Gold noted that without negative evidence: If a language Lx (a set of sentences) is a proper subset of a language Ly, then a learner that hypothesized Ly when the target was in fact Lx, would have made an error not curable by any subsequent input.  Example: Lx = English; Ly = a language like English but with scrambling.  For target English, a child guessing Ly would make word-order errors (e.g. Jim fish likes) detectable to English listeners, but not by the learner herself listening to positive examples by others!  This doesn’t happen! If it did, languages would grow larger and larger over time, as one person’s superset error would be input to later learners. But this is not the case. What prevents it? 8 Ly Lx

99 How to solve the Subset Problem?  Since children rarely if ever make superset errors, our models of learning must include some factor that prevents them.  Gold’s proposal: The possible grammar hypotheses are priority- ordered such that every ‘subset grammar’ takes precedence over all of its ‘superset grammars’.  Learners would need innate knowledge of this ordering. And an innate strategy of systematically testing grammars in order of priority. The learner can move on from Li to Li+1 only if Li fails on an input sentence.  This ‘enumeration’ approach works! But it has since been rejected as linguistically unrealistic. It predicts that some languages are acquired vastly more slowly than others, just because they appear late in the enumeration – regardless of how distinctive their sentences are. (Pinker 1979)  We’ll consider other possible solutions in Class 5. 9

Moving closer to linguistics…  Wexler & Culicover (1980) developed a learning model that incorporated:  linguistic theory ( Extended Standard Theory, Chomsky 1973)  psychological resource limits ( i.e., maximum 3-clause sentences needed for acquisition, “ degree-2 embedding” ).  In EST linguistics, surface structures of sentences were derived from deep structures by long derivations, with cyclic application of many transformational rules.  To achieve their proof of learnability, W&C had to make some very strong assumptions: that deep structures are universal; and that a unique association with meaning provides the learner with the deep structure of each input sentence.  Thus, only the transformational rules needed to be learned. 10

Acquiring T-rules (W&C)  Learner hears a sentence, guesses its meaning from context, represents the DS, applies currently hypothesized T-rules to generate a surface string of words. If it matches the input string, retain the current grammar hypothesis.  Learner hears a sentence, guesses its meaning from context, represents the DS, applies currently hypothesized T-rules to generate a surface string of words. If it matches the input string, retain the current grammar hypothesis.  If the strings mis-match, the learner either: deletes one T-rule from the grammar, at random; or: adds at random any one T-rule that creates a match.  Psychologically implausible! Also, the resulting grammar might be less adequate than the one it replaced (e.g., deletion of a needed T-rule).   Much trial-and-error needed to arrive at the target grammar.  W&C proved (heroically!) that such grammars could be reliably acquired in this fashion, but only if the T-rules were constrained in certain ways. 11

Close convergence with linguistic theory  Very exciting! These constraints proven necessary for learnability had theoretical linguistic credentials.  W&C’s Binary Principle disallowed transformational operations over more than two adjacent clauses. Very like the Subjacency Condition that Chomsky had argued for on purely linguistic grounds.  W&C’s Freezing Principle incorporated several other limits on transformational applications already observed by linguists.  To be useful, learners must innately know these constraints. So these parallels with linguistic theory were welcomed as support for highly specific innate linguistic knowledge.  Though not very psychologically realistic(!), this work paved the way toward linguistically-informed learning models.  But it had a short life. Published in Then in 1981… !! 12

1981: No more transformational rules!  Through the 1970s, linguists had been streamlining the original-style T-rules, each with its own built-in constraints. Instead, general constraints on rule application (as in W&C’s learnability proof). Encouraging for W&C, but…  End result was one single T-rule: Move α (later, Affect α). This means: any possible transformational operation is free to apply at any point in a derivation except where it is blocked by the general constraints (e.g., Subjacency).  Consequence: No way to register variation between lang- uages in terms of different sets of transformational rules.  The newly introduced parameters took on this role.  The syntactic component was now held to be fully innate and universal, except for a finite number of choice points. 13

New look: Syntax acquisition as setting parameters  The Headedness parameter: Are syntactic phrases head- initial (e.g., in VP, the verb precedes its object) or head-final (the verb follows the object)?  The Wh-movement parameter: Does a Wh-phrase move to the top (the Complementizer projection) of a clause or does it remain in situ?  Now, all the learner has to do is detect the correct setting of each of a finite number of parameters, based on the input.  Parameter values were said to be ‘triggered’ by the learner’s encounter with a distinctive property of an input sentence.  This would greatly reduce the workload of data-processing, and would help address the Poverty of Stimulus problem. 14

15 Parameter setting as flipping switches Parameter setting as flipping switches  Chomsky never provided a specific implementation of parametric triggering. He often employed the metaphor of setting switches.  In next class (Wednesday), we’ll see how this metaphor might be cashed out as specific psychological computations.  Today we covered the early days: Applied Math  Linguistics  We’ll move on through: Computer science  Psycholinguistics  Focus will be on how to set syntactic parameters.  READING: Please read for Wednesday (Class 2) The 7-page encyclopedia entry: “Principles and parameters theory and language acquisition” by Snyder & Lillo-Martin. More examples of syntactic parameters. Related to child performance data and learning models. Well-presented.