Why Syntax is Impossible Mike Dowman. Syntax FLanguages have tens of thousands of words FSome combinations of words make valid sentences FOthers don’t.

Slides:



Advertisements
Similar presentations
Writing an Amazing Thesis Statement Developmental Writing.
Advertisements

Linguistic Theory Lecture 11 Explanation.
Chapter Three: Closure Properties for Regular Languages
Chapter 4 Key Concepts.
GETTING TO KNOW THE SAT TIPS AND TRICKS TO IMPROVE YOUR SAT SCORE MR. TORRES 10/02/2013.
Data Mining Methodology 1. Why have a Methodology  Don’t want to learn things that aren’t true May not represent any underlying reality ○ Spurious correlation.
Psych 156A/ Ling 150: Acquisition of Language II Lecture 12 Poverty of the Stimulus I.
Ling 240: Language and Mind Structure Dependence in Grammar Formation.
Introduction: The Chomskian Perspective on Language Study.
Computational Models of Discourse Analysis Carolyn Penstein Rosé Language Technologies Institute/ Human-Computer Interaction Institute.
1 Language and kids Linguistics lecture #8 November 21, 2006.
The Linguistics of SLA.
Second Language Acquisition (SLA)
Language Special form of communication in which we learn complex rules to manipulate symbols that can be used to generate an endless number of meaningful.
Module 14 Thought & Language. INTRODUCTION Definitions –Cognitive approach method of studying how we process, store, and use information and how this.
Week 5a. Binding theory CAS LX 522 Syntax I. Structural ambiguity John said that Bill slipped in the kitchen. John said that Bill slipped in the kitchen.
LEARNING FROM OBSERVATIONS Yılmaz KILIÇASLAN. Definition Learning takes place as the agent observes its interactions with the world and its own decision-making.
Introduction to Computational Natural Language Learning Linguistics (Under: Topics in Natural Language Processing ) Computer Science (Under:
Language, Mind, and Brain by Ewa Dabrowska Chapter 3: Language acquisition: robustness.
COMP305. Part II. Genetic Algorithms. Genetic Algorithms.
1 Introduction to Computational Natural Language Learning Linguistics (Under: Topics in Natural Language Processing ) Computer Science (Under:
Unit 1 Biology Notes Characteristics of Life
Fundamentals: Linguistic principles
Meaning and Language Part 1.
Models of Generative Grammar Smriti Singh. Generative Grammar  A Generative Grammar is a set of formal rules that can generate an infinite set of sentences.
Lecture 1 Introduction: Linguistic Theory and Theories
Generative Grammar(Part ii)
Scot Exec Course Nov/Dec 04 Ambitious title? Confidence intervals, design effects and significance tests for surveys. How to calculate sample numbers when.
How many polar bears? How many fish? What about Plankton?
Eric Cohen Books 2007 Simply Writing - Task to Project to Bagrut What's expected and what you can do! Clarity, Expectation and Format.
Genetic Algorithm.
Functions Definition & purpose Notation Functions on binary / returning binary values Finite automaton model We haven’t completely left the world of counting.
Language and Cognition. Language is the arrangement of sounds to communicate ideas…  1. Language is arbitrary-words don’t sound like the ideas they are.
Process of Science The Scientific Method.
Cognitive Development: Language Infants and children face an especially important developmental task with the acquisition of language.
Populations, Samples, and Probability. Populations and Samples Population – Any complete set of observations (or potential observations) may be characterized.
1 Words and rules Linguistics lecture #2 October 31, 2006.
Modelling Language Evolution Lecture 5: Iterated Learning Simon Kirby University of Edinburgh Language Evolution & Computation Research Unit.
What is linguistics  It is the science of language.  Linguistics is the systematic study of language.  The field of linguistics is concerned with the.
Big Idea 1: The Practice of Science Description A: Scientific inquiry is a multifaceted activity; the processes of science include the formulation of scientifically.
Learning from Observations Chapter 18 Through
An Intelligent Analyzer and Understander of English Yorick Wilks 1975, ACM.
1 The Scientist Game Chris Slaughter, DrPH (courtesy of Scott Emerson) Dept of Biostatistics Vanderbilt University © 2002, 2003, 2006, 2008 Scott S. Emerson,
GENETIC ALGORITHM A biologically inspired model of intelligence and the principles of biological evolution are applied to find solutions to difficult problems.
Copyright © Curt Hill Languages and Grammars This is not English Class. But there is a resemblance.
Universal Grammar Noam Chomsky.
WHAT IS THE NATURE OF SCIENCE?. SCIENTIFIC WORLD VIEW 1.The Universe Is Understandable. 2.The Universe Is a Vast Single System In Which the Basic Rules.
PSY270 Michaela Porubanova. Language  a system of communication using sounds or symbols that enables us to express our feelings, thoughts, ideas, and.
Introduction Chapter 1 Foundations of statistical natural language processing.
Syntax Andrew Carnie. The web page for this textbook.
Inference: Probabilities and Distributions Feb , 2012.
Writing a paragraph. What is a paragraph? A paragraph is a group of about sentences about one topic. Every sentence in a strong paragraph is about.
SYNTAX.
Instructor Availability AIM address: EleBranch Office Hours: Mondays & Thursdays 9:00 p.m. to 10:00 p.m. And by appointment.
Unit 2 The Nature of Learner Language 1. Errors and errors analysis 2. Developmental patterns 3. Variability in learner language.
SIMS 296a-4 Text Data Mining Marti Hearst UC Berkeley SIMS.
Syntax.
Chapter 3 Language Acquisition: A Linguistic Treatment Jang, HaYoung Biointelligence Laborotary Seoul National University.
Asteroid Strike! Research the answers to these questions: What caused the extinction of the dinosaurs? What is the evidence for this theory? What were.
Purpose… …to help you improve your expository and persuasive writing, including conventions, in order to meet standard on the writing portion of the HSPE.
Language is common to all humans; we seem to be “hard-wired” for it
Psych156A/Ling150: Psychology of Language Learning
Theories of Language Development
Areas of Language Development Theories of Language Development
Communicative Competence (Canale and Swain, 1980)
Roger Brown’s (1973) First Language Development Study and MLU
Inference for Distributions of Categorical Data
Development and Language
Presentation transcript:

Why Syntax is Impossible Mike Dowman

Syntax FLanguages have tens of thousands of words FSome combinations of words make valid sentences FOthers don’t FNo one understands the grammar of any language FLanguages have tens of thousands of words FSome combinations of words make valid sentences FOthers don’t FNo one understands the grammar of any language

Syntax is Complicated! I saw Bill with Mary yesterday. You saw WHO with Mary yesterday?! Who did you see with Mary yesterday? I saw Bill with Mary yesterday. You saw WHO with Mary yesterday?! Who did you see with Mary yesterday?

Syntax is Complicated! I saw Bill with Mary yesterday. You saw WHO with Mary yesterday?! Who did you see with Mary yesterday? I saw Bill and Mary yesterday. You saw WHO and Mary yesterday?! I saw Bill with Mary yesterday. You saw WHO with Mary yesterday?! Who did you see with Mary yesterday? I saw Bill and Mary yesterday. You saw WHO and Mary yesterday?!

Syntax is Complicated! I saw Bill with Mary yesterday. You saw WHO with Mary yesterday?! Who did you see with Mary yesterday? I saw Bill and Mary yesterday. You saw WHO and Mary yesterday?! Who did you see and Mary yesterday? I saw Bill with Mary yesterday. You saw WHO with Mary yesterday?! Who did you see with Mary yesterday? I saw Bill and Mary yesterday. You saw WHO and Mary yesterday?! Who did you see and Mary yesterday?

Generative Grammar FAn explicit formal system that defines the set of valid sentences in a language FAnd maybe also explains what each one means FGenerative grammar is the core research topic in linguistics FIncludes strongly nativist theories and theories proposing that languages are primarily learned FAn explicit formal system that defines the set of valid sentences in a language FAnd maybe also explains what each one means FGenerative grammar is the core research topic in linguistics FIncludes strongly nativist theories and theories proposing that languages are primarily learned

Grammar Writing FLinguists take a selection of possible sentences FAnd obtain grammaticality judgments for those sentences FThen they produce a grammar that accounts for all the data FLinguists take a selection of possible sentences FAnd obtain grammaticality judgments for those sentences FThen they produce a grammar that accounts for all the data

Grammar Coverage FLinguists’ grammars only work for selected sentences FThey can’t explain most naturally occurring sentences FThe more data we consider the more surprising quirks of syntax that emerge FLinguists’ grammars only work for selected sentences FThey can’t explain most naturally occurring sentences FThe more data we consider the more surprising quirks of syntax that emerge

Children’s Language Acquisition FKid’s observe a limited number of example sentences FBut quickly internalize a system that correctly characterizes the whole language FKid’s observe a limited number of example sentences FBut quickly internalize a system that correctly characterizes the whole language I-language E-language LAD

How can kids do syntax when linguists can’t? FInnate component of language (provided by genes) FLearned component of language (provided by language data) FInnate component of language (provided by genes) FLearned component of language (provided by language data)

How can kids do syntax when linguists can’t? FInnate component of language (provided by genes) FLearned component of language (provided by language data) FLinguists have to infer both FChildren only the learned component FInnate component of language (provided by genes) FLearned component of language (provided by language data) FLinguists have to infer both FChildren only the learned component

Information Theory FBoth components of language must contain some amount of information FData available to children must provide at least enough information as is in the learned component FThis puts a limit on the complexity of the learned component of language FBoth components of language must contain some amount of information FData available to children must provide at least enough information as is in the learned component FThis puts a limit on the complexity of the learned component of language

Linguists’ Task FLinguists need to have at least as much information as is in the learned and innate components together FCan use data from multiple languages to try to characterize innate components FAnd can use positive and negative data FLinguists need to have at least as much information as is in the learned and innate components together FCan use data from multiple languages to try to characterize innate components FAnd can use positive and negative data

Correspondence to Linguistic Theories Small learned component = parameter setting Large learned component = learned languages Small innate component = general learning mechanism Large innate component = universal grammar Small learned component = parameter setting Large learned component = learned languages Small innate component = general learning mechanism Large innate component = universal grammar

Size of Each Component

Which component is large? FAs we haven’t yet managed to produce a generative grammar, at least one of innate or learned components must be large FChildren learn relatively easily, so the learned component can’t be too big FAs we haven’t yet managed to produce a generative grammar, at least one of innate or learned components must be large FChildren learn relatively easily, so the learned component can’t be too big

Size of Each Component

How big could the innate component be? FGenome contains 3 billion base pairs = 6 billion bits FCell metabolism adds more information FEach base pair can be modified FHuge amount of information! FGenome contains 3 billion base pairs = 6 billion bits FCell metabolism adds more information FEach base pair can be modified FHuge amount of information!

What could be in a huge innate component? FNot words forms - vary from language to language FGrammaticality patterns FRules of syntax would be hugely complex FNot words forms - vary from language to language FGrammaticality patterns FRules of syntax would be hugely complex

Impossibility of Syntax FGrammaticality judgments on average can provide no more than one bit of information each FIf syntax is hugely complex, there will be many grammars that are compatible with any given body of data FBut all but one of these grammars would fail when tested on enough new data FGrammaticality judgments on average can provide no more than one bit of information each FIf syntax is hugely complex, there will be many grammars that are compatible with any given body of data FBut all but one of these grammars would fail when tested on enough new data

A Concrete Example FA multi-agent model FEach agent has: innate component learned component FBoth are bit strings of fixed length FSentences are 100 bit strings FA multi-agent model FEach agent has: innate component learned component FBoth are bit strings of fixed length FSentences are 100 bit strings

Deciding on the Grammaticality of a Sentence 1 FTreat the sentence as a binary number FFind: b i = s mod n i b l = s mod n l b is an index to a bit in the innate ( b i ) or learned ( b l ) component n is the number of bits in the innate ( n i ) or learned ( n l ) component s is the length of the sentences FTreat the sentence as a binary number FFind: b i = s mod n i b l = s mod n l b is an index to a bit in the innate ( b i ) or learned ( b l ) component n is the number of bits in the innate ( n i ) or learned ( n l ) component s is the length of the sentences

Deciding on the Grammaticality of a Sentence 2 FA pseudo-random function maps from the two selected bits plus the sentence to a Boolean grammaticality judgment FIt’s therefore typically necessary to know every bit of the sentence and both the innate and learned bits to predict the grammaticality of the sentence  Every bit counts FUsually about half of sentences are grammatical, half ungrammatical FA pseudo-random function maps from the two selected bits plus the sentence to a Boolean grammaticality judgment FIt’s therefore typically necessary to know every bit of the sentence and both the innate and learned bits to predict the grammaticality of the sentence  Every bit counts FUsually about half of sentences are grammatical, half ungrammatical

4 Kinds of Agent Teacher Innate: Learned: Related Innate: Learned: Unrelated Innate: Learned: Linguist Innate: Learned:

Learning by Related, Unrelated FObserve a sentence from the teacher FWork out if it is grammatical according to current I-language FIf not, invert the relevant bit of the learned component FObserve a sentence from the teacher FWork out if it is grammatical according to current I-language FIf not, invert the relevant bit of the learned component

Grammar Inference by Linguists FChoose random sentences FAsk the teacher if they are grammatical FStore all sentences and grammaticality judgments FSearch for a setting of innate and learned components that assigns the correct grammaticality rating to every sentence FChoose random sentences FAsk the teacher if they are grammatical FStore all sentences and grammaticality judgments FSearch for a setting of innate and learned components that assigns the correct grammaticality rating to every sentence

1,000 Bit Innate and Learned Components

1,000 Bit Innate Component 1,000,000 Bit Learned Component

1,000,000 Bit Innate Component 1,000 Bit Learned Component

Implications of Impossible Syntax FA linguist can write a grammar that will adequately characterize any body of data FBut it will fail when tested on new data FPartial grammars are not a stepping stone to complete generative grammars FA linguist can write a grammar that will adequately characterize any body of data FBut it will fail when tested on new data FPartial grammars are not a stepping stone to complete generative grammars

A Universal Law of Generative Grammar Generative grammar is impossible if: H(learned component) + H(innate component) > H(language data) Unless we can use information from another source (genetic, neuroscientific, psycholinguistic) Generative grammar is impossible if: H(learned component) + H(innate component) > H(language data) Unless we can use information from another source (genetic, neuroscientific, psycholinguistic)

Why do Syntax? FStudying generative grammar may tell us something about the human mind FIt won’t help us build natural language processing systems FIs studying rare and obscure constructions the best way to do syntax? FStudying generative grammar may tell us something about the human mind FIt won’t help us build natural language processing systems FIs studying rare and obscure constructions the best way to do syntax?

Conclusion FThe idea that we can characterize a language by considering enough linguistic data is a hypothesis FIt’s very unlikely that it’s possible to write a complete generative grammar FThe idea that we can characterize a language by considering enough linguistic data is a hypothesis FIt’s very unlikely that it’s possible to write a complete generative grammar