Computational Intelligence 696i Language Lecture 3 Sandiway Fong.

Slides:



Advertisements
Similar presentations
Lecture 2: Constraints on Movement.  Formal movement rules (called Transformations) were first introduced in the late 1950s  During the 1960s a lot.
Advertisements

CS107: Introduction to Computer Science Lecture 2 Jan 29th.
Using the Crosscutting Concepts As conceptual tools when meeting an unfamiliar problem or phenomenon.
Language and Cognition Colombo, June 2011 Day 2 Introduction to Linguistic Theory, Part 4.
Introduction: The Chomskian Perspective on Language Study.
David Evans CS200: Computer Science University of Virginia Computer Science Lecture 3: Rules of Evaluation.
Week 3a. UG and L2A: Background, principles, parameters CAS LX 400 Second Language Acquisition.
Statistical NLP: Lecture 3
ArchE Presented By Samanvitha Ramayanam. TOPICS 1. Introduction 2. Theoretical assumptions 3. ArchE as an expert system 4. Overall flow of ArchE 5. Key.
The theoretical significance of UG in second language acquisition Present by Esther, Karen, Mei.
Linguistic Theory Lecture 8 Meaning and Grammar. A brief history In classical and traditional grammar not much distinction was made between grammar and.
Week 13. Loose ends, minimalism CAS LX 522 Syntax I.
LING 388: Language and Computers Sandiway Fong Lecture 22: 11/10.
LING 364: Introduction to Formal Semantics Lecture 4 January 24th.
Computational Intelligence 696i Language Lecture 2 Sandiway Fong.
1 Introduction to Computational Linguistics Eleni Miltsakaki AUTH Fall 2005-Lecture 2.
Computational Intelligence 696i Language Lecture 4 Sandiway Fong.
1 Introduction to Computational Natural Language Learning Linguistics (Under: Topics in Natural Language Processing ) Computer Science (Under:
Week 13a. QR CAS LX 522 Syntax I. Quantifiers We interpret Bill saw everyone as We interpret Bill saw everyone as For every person x, Bill saw x. For.
Bounding Theory Constraints on Wh-movement. NP islands What i did Bill claim [ CP that he read t i ?] *What did Bill make [ NP the claim [ CP that he.
Syntax Lecture 3: The Subject. The Basic Structure of the Clause Recall that our theory of structure says that all structures follow this pattern: It.
Models of Generative Grammar Smriti Singh. Generative Grammar  A Generative Grammar is a set of formal rules that can generate an infinite set of sentences.
Lecture 1 Introduction: Linguistic Theory and Theories
Linguistic Theory Lecture 3 Movement. A brief history of movement Movements as ‘special rules’ proposed to capture facts that phrase structure rules cannot.
Albert Gatt LIN 3098 Corpus Linguistics. In this lecture Some more on corpora and grammar Construction Grammar as a theoretical framework Collostructional.
Syntax Lecture 8: Verb Types 1. Introduction We have seen: – The subject starts off close to the verb, but moves to specifier of IP – The verb starts.
LING 696G Computational Linguistics Seminar Lecture 2 2/2/04.
1 What does “meaning” mean? Linguistics lecture #3 November 2, 2006.
1 Syntax 1 Homework to be handed in on 19th or 26th December: Fromkin morphology exercises 6, 7, & 14. Also, do either 8 or 9, and only one of 15, 16 or.
THE BIG PICTURE Basic Assumptions Linguistics is the empirical science that studies language (or linguistic behavior) Linguistics proposes theories (models)
Linguistic Theory Lecture 10 Grammaticality. How do grammars determine what is grammatical? 1 st idea (traditional – 1970): 1 st idea (traditional – 1970):
October 15, 2007 Non-finite clauses and control : Grammars and Lexicons Lori Levin.
Context Free Grammars Reading: Chap 9, Jurafsky & Martin This slide set was adapted from J. Martin, U. Colorado Instructor: Rada Mihalcea.
David Evans CS200: Computer Science University of Virginia Computer Science Lecture 3: Rules of Evaluation.
Models of Linguistic Choice Christopher Manning. 2 Explaining more: How do people choose to express things? What people do say has two parts: Contingent.
Culture , Language and Communication
Albert Gatt LIN3021 Formal Semantics Lecture 4. In this lecture Compositionality in Natural Langauge revisited: The role of types The typed lambda calculus.
Revision.  Movements leave behind a phonologically null trace in all their extraction sites.
Programming Languages and Design Lecture 3 Semantic Specifications of Programming Languages Instructor: Li Ma Department of Computer Science Texas Southern.
Rules, Movement, Ambiguity
The Minimalist Program
Linguistic Theory Lecture 5 Filters. The Structure of the Grammar 1960s (Standard Theory) LexiconPhrase Structure Rules Deep Structure Transformations.
1 LIN 1310B Introduction to Linguistics Prof: Nikolay Slavkov TA: Qinghua Tang CLASS 16, March 6, 2007.
Ian Roberts  Generate well-formed structural descriptions  “create” trees/labelled bracketings  More (X’) or less (PS-rules) abstract.
1 Introduction to Computational Linguistics Eleni Miltsakaki AUTH Spring 2006-Lecture 2.
1 Syntax 1. 2 In your free time Look at the diagram again, and try to understand it. Phonetics Phonology Sounds of language Linguistics Grammar MorphologySyntax.
SYNTAX.
Lecture 1: Trace Theory.  We have seen that things move :  Arguments move out of the VP into subject position  Wh-phrases move out of IP into CP 
1 Introduction to Language Acquisition Theory Janet Dean Fodor St. Petersburg July 2013 Class 8. Implications and further questions Class 8. Implications.
1 Principles & Parameters Approach in Linguistics II Bibhuti Bhusan Mahapatra.
Piaget, Vygotsky, Bruner, Krashen, Chomsky
Lecture 3: Rules of Evaluation CS150: Computer Science
Chapter 7 Linguistics aspect of interlanguage
Principles and Parameters (II) Rajat Kumar Mohanty Department of Computer Science and Engineering Indian Institute of Technology Bombay.
LING/C SC/PSYC 438/538 Lecture 19 Sandiway Fong 1.
Computational Intelligence 696i Language Homework 1 Answers Sandiway Fong.
CSE 311 Foundations of Computing I Lecture 19 Recursive Definitions: Context-Free Grammars and Languages Autumn 2012 CSE
MENTAL GRAMMAR Language and mind. First half of 20 th cent. – What the main goal of linguistics should be? Behaviorism – Bloomfield: goal of linguistics.
Chapter 3 Language Acquisition: A Linguistic Treatment Jang, HaYoung Biointelligence Laborotary Seoul National University.
Lec. 10.  In this section we explain which constituents of a sentence are minimally required, and why. We first provide an informal discussion and then.
Poverty of Stimulus Poverty of Stimulus Reading Group.
Syntax 1.
Syntax Lecture 9: Verb Types 1.
Statistical NLP: Lecture 3
Introduction to Linguistics
Language Variations: Japanese and English
CSCI 5832 Natural Language Processing
Traditional Grammar VS. Generative Grammar
Principles and Parameters (I)
Structure of a Lexicon Debasri Chakrabarti 13-May-19.
Presentation transcript:

Computational Intelligence 696i Language Lecture 3 Sandiway Fong

Administriva Has every group managed to install PAPPI? –(see instructions from last Thursday) –You’ll need it to do homework 1

Last Time –we talked about the problem of gap filling, a necessary component in the recovery of meaning Examples: –Which report did you file without reading? –Which report did you file [the report] without [you] reading [the report]? –*Which book did you file the report without reading –These papers are easy to file without reading

Last Time speakers assume hearers share the same rules or decoding mechanisms –allow gaps to exist in input the gap decoding mechanism is pretty complicated –more soon poverty of stimulus: is it real? –if mechanism is really complicated –is the mechanism acquired (at all)? –is there enough data? the decoding mechanism –might be part of our genetic endowment –or a consequence of the pre-wiring possibly part of Universal Grammar (UG) –reduce the burden of the language learner

Today the very idea of UG is super-cool –very provocative hypothesis –all languages obey the same rules or have the same structure at some level of abstraction but attractive if true –e.g. important consequences for machine translation What might such a theory look like? Can it even be formalized? Can we compute with UG? –e.g. use it to parse sentences Given UG, can we come up with a mechanism for setting the parameters –are the parameters learnable? PAPPI

Today’s Lecture goal is to take a concrete look at one possible instantiation of UG –the principles-and-parameters (P&P) framework parameter: language-specific {0,1} –e.g. determines things like word-order SVO, SOV, VSO etc. principle: e.g. a constraint or architectural feature –a primitive of the UG –may be parameterized aka –Government-Binding (GB) Theory best known instantiation being Lectures on Government and Binding (Chomsky 1981)

The “Rules” Minimalist Program (MP) current linguistic technology (research area) language is a computational system even fewer mechanisms Principles-and-Parameters Framework (GB) reduction of construction-specific rules to fundamental principles (the atoms of theory) explanatory adequacy Rule-based systems construction-based monostratal, e.g. context-free grammars multiple levels. e.g. transformational grammars

Rule-Based Approach –to understand the totally revolutionary nature of the paradigm shift in linguistics that occurred around 1980 –we have to first look at pre-existing approaches Examples: –Which report did you file without reading? –Which report did you file [the report] without [you] reading [the report]? –*Which book did you file the report without reading –These papers are easy to file without reading

Rule-Based Approach Gap filling: –Which report did you file without reading? –Which report did you file [the report] without [you] reading [the report]? Assume some phrase structure: –[ S subject [ VP V object]]for transitive V We have: –wh-phrase did [ S you [ VP [ VP file e][ PP without [ S e [ VP reading e]]]]] Construction-specific rule: –if [ S subject [ VP [ VP V e-object 1 ][ PP P [ S e-subject [ VP V+ing e-object 2 ]]]]] –then e-subject = subject, e-object 2 = e-object 1

Rule-Based Approach Construction-specific rule: –if [ S subject [ VP [ VP V e-object 1 ][ PP P [ S e-subject [ VP V+ing e-object 2 ]]]]] –then e-subject = subject, e-object 2 = e-object 1 Where does this rule come from? How does anyone manage to learn this rule? Can we generalize this rule to other examples? –These papers are easy to file without reading –[ S e-subject [ VP [ VP file e-object 1 ][ PP without [ S e-subject [ VP reading e-object 2 ]]]]] Revised rule: –if [ S (e-)subject [ VP [ VP V e-object 1 ][ PP P [ S e-subject [ VP V+ing e-object 2 ]]]]] –then e-subject = (e-)subject, e-object 2 = e-object 1

Rule-Based Approach Revised rule: –if [ S (e-)subject [ VP [ VP V e-object 1 ][ PP P [ S e-subject [ VP V+ing e- object 2 ]]]]] –then e-subject = (e-)subject, e-object 2 = e-object 1 BTW, e-object 2 has to be linked with an e-object 1 (not an overt one) on the basis of examples like: –*you filed the report without reading –(cf. you filed the report without reading it) –[ S you [ VP [ VP filed the report][ PP without [ S e [ VP reading e]]]]] Have to know: –if [ S subject [ VP [ VP V object][ PP P [ S e-subject [ VP V+ing e-object]]]]] –then gap filling fails

Rule-Based Approach Have to know: –if [ S subject [ VP [ VP V object][ PP P [ S e-subject [ VP V+ing e-object]]]]] –then gap filling fails Note: –I’m using negative data to refine my rule Also works for: –*Which book did you file the report without reading –[ S you [ VP [ VP filed the report][ PP without [ S e [ VP reading e]]]]] Generalization (simplified): –(final) e-object requires another e-object to be present –e-object is a parasitic gap

Rule-Based Approach Consider: –*Which book did you file the report without reading How to say it in English? –assuming underlying structure is –you filed the report without reading which book

Rule-Based Approach –repeat process for all constructions in the language –end up with a huge number of complex rules –(repeat for next language...) (Some) linguists found such construction-specific “rule-based” systems unsatisfactory –too many rules –rules seem somewhat arbitrary (rule systems too powerful) –can’t possibly be learned (maybe) –lack of conceptual elegance –is there a better way? –can the apparent complexity be derived from more fundamental (and simpler) systems?

The “Rules” Minimalist Program (MP) current linguistic technology (research area) language is a computational system even fewer mechanisms Principles-and-Parameters Framework (GB) reduction of construction-specific rules to fundamental principles (the atoms of theory) explanatory adequacy Rule-based systems construction-based huge number of rules

Principles-and-Parameters No construction-specific pattern-matching rules –if [ S (e-)subject [ VP [ VP V e-object 1 ][ PP P [ S e-subject [ VP V+ing e-object 2 ]]]]] –then e-subject = (e-)subject, e-object 2 = e-object 1 Not even generalizations like –(final) e-object requires another e-object to be present Won’t find any principle in the system directly resembling these rules These are all lemmas (or generalizations) derivable from more primitive properties of grammar

Principles-and-Parameters –can’t explain all the details in a few lectures –there are dozens of principles –... give just a sketch of the general system the P&P framework is a modular system –of simple sub-theories, i.e. modules –each module is responsible for constraining or licensing of some class of element(s), e.g. e- elements –these simple modules interact in complex ways to explain complex behavior

Principles-and-Parameters which report did you file? ECP Theta Move-α Binding

Principles-and-Parameters Some modules –X-bar theory universal phrase structure [X” specifier [X’ X complement]] binary-branching only X ranging over {N,V,A,P,I,C,neg and a few others } order of specifier and complement vary for particular languages parameters: head-complement order etc.

Principles-and-Parameters Some modules –Move-alpha (Move-α) universal displacement property what did John see John see what what did John see trace principle: move any phrase anywhere don’t worry about cases where we can’t displace a phrase (other modules will take care of that) *what does Bill wonder who saw? who did you mention that Bill believes that you saw? *who did you mention Bill's belief that you saw?

Principles-and-Parameters Some modules –Subjacency locality of displacement: things can’t move too far in one hop interaction with X-bar theory (phrase structure) *what does Bill wonder who saw? who did you mention that Bill believes that you saw? *who did you mention Bill's belief that you saw? parameter: bounding node IP (English), CP (Italian)

Principles-and-Parameters Some modules –Theta theory who did what to whom file: (filer,filed) read:(reader,read) theta-roles: filer/reader => agent.. –(patient, theme, experiencer) arguments: the report, you [V” specifier [V’ V complement]] principle: theta-criterion –every arguments needs one theta-role –every theta-role needs to be expressed don’t worry about e-elements (other modules’ responsibility)

Principles-and-Parameters Some modules –Case Theory John is likely to be here It is likely that John is here *It is likely John to be here (cf. I believe John to be here) –Empty Category Principle (ECP) subject/object asymmetry for e-elements who do you think (that) John saw? who do you think saw John? *who do you think that saw John? –Binding Theory (anaphors and pronouns) interaction of displacement and binding theory who that John knows does he like? (ambiguous) He likes everyone that John knows (not ambiguous)

House of Cards Analogy a system of modules delicate hard to build rely on each other interact in complex ways independent justification principles affect many different kinds of constructions theoretically more satisfying