Introduction and Jurafsky Model Resource: A Probabilistic Model of Lexical and Syntactic Access and Disambiguation, Jurafsky 1996.

Slides:



Advertisements
Similar presentations
Computational language: week 10 Lexical Knowledge Representation concluded Syntax-based computational language Sentence structure: syntax Context free.
Advertisements

Natural Language Processing - Parsing 1 - Language, Syntax, Parsing Problems in Parsing Ambiguity, Attachment / Binding Bottom vs. Top Down Parsing.
PARSING WITH CONTEXT-FREE GRAMMARS
Statistical NLP: Lecture 3
Chapter 12 Lexicalized and Probabilistic Parsing Guoqiang Shan University of Arizona November 30, 2006.
March 1, 2009 Dr. Muhammed Al-Mulhem 1 ICS 482 Natural Language Processing Probabilistic Context Free Grammars (Chapter 14) Muhammed Al-Mulhem March 1,
For Monday Read Chapter 23, sections 3-4 Homework –Chapter 23, exercises 1, 6, 14, 19 –Do them in order. Do NOT read ahead.
Artificial Intelligence 2004 Natural Language Processing - Syntax and Parsing - Language, Syntax, Parsing Problems in Parsing Ambiguity, Attachment.
PCFG Parsing, Evaluation, & Improvements Ling 571 Deep Processing Techniques for NLP January 24, 2011.
Parsing context-free grammars Context-free grammars specify structure, not process. There are many different ways to parse input in accordance with a given.
Albert Gatt LIN3022 Natural Language Processing Lecture 8.
Amirkabir University of Technology Computer Engineering Faculty AILAB Efficient Parsing Ahmad Abdollahzadeh Barfouroush Aban 1381 Natural Language Processing.
The Neural Basis of Thought and Language Week 15 The End is near...
Day 2: Pruning continued; begin competition models
Syntax LING October 11, 2006 Joshua Tauberer.
1 CONTEXT-FREE GRAMMARS. NLE 2 Syntactic analysis (Parsing) S NPVP ATNNSVBD NP AT NNthechildrenate thecake.
Context Free Grammar S -> NP VP NP -> det (adj) N
Artificial Intelligence 2004 Natural Language Processing - Syntax and Parsing - Language Syntax Parsing.
Intro to Psycholinguistics What its experiments are teaching us about language processing and production.
Context-Free Grammar CSCI-GA.2590 – Lecture 3 Ralph Grishman NYU.
1 Basic Parsing with Context Free Grammars Chapter 13 September/October 2012 Lecture 6.
Syntax Nuha AlWadaani.
11 CS 388: Natural Language Processing: Syntactic Parsing Raymond J. Mooney University of Texas at Austin.
Context Free Grammars Reading: Chap 12-13, Jurafsky & Martin This slide set was adapted from J. Martin, U. Colorado Instructor: Paul Tarau, based on Rada.
PARSING David Kauchak CS457 – Fall 2011 some slides adapted from Ray Mooney.
BİL711 Natural Language Processing1 Statistical Parse Disambiguation Problem: –How do we disambiguate among a set of parses of a given sentence? –We want.
1 Statistical Parsing Chapter 14 October 2012 Lecture #9.
1 Natural Language Processing Lecture 11 Efficient Parsing Reading: James Allen NLU (Chapter 6)
GRAMMARS David Kauchak CS159 – Fall 2014 some slides adapted from Ray Mooney.
Chapter 10. Parsing with CFGs From: Chapter 10 of An Introduction to Natural Language Processing, Computational Linguistics, and Speech Recognition, by.
LINGUISTICA GENERALE E COMPUTAZIONALE ANALISI SINTATTICA (PARSING)
PS: Introduction to Psycholinguistics Winter Term 2005/06 Instructor: Daniel Wiechmann Office hours: Mon 2-3 pm Phone:
Parsing I: Earley Parser CMSC Natural Language Processing May 1, 2003.
A Cascaded Finite-State Parser for German Michael Schiehlen Institut für Maschinelle Sprachverarbeitung Universität Stuttgart
PARSING David Kauchak CS159 – Spring 2011 some slides adapted from Ray Mooney.
Context Free Grammars Reading: Chap 9, Jurafsky & Martin This slide set was adapted from J. Martin, U. Colorado Instructor: Rada Mihalcea.
11 Chapter 14 Part 1 Statistical Parsing Based on slides by Ray Mooney.
Albert Gatt Corpora and Statistical Methods Lecture 11.
For Wednesday Read chapter 23 Homework: –Chapter 22, exercises 1,4, 7, and 14.
Linguistic Essentials
CSA2050 Introduction to Computational Linguistics Parsing I.
Natural Language - General
PARSING 2 David Kauchak CS159 – Spring 2011 some slides adapted from Ray Mooney.
NLP. Introduction to NLP Motivation –A lot of the work is repeated –Caching intermediate results improves the complexity Dynamic programming –Building.
Section 11.3 Features structures in the Grammar ─ Jin Wang.
LING 001 Introduction to Linguistics Spring 2010 Syntactic parsing Part-Of-Speech tagging Apr. 5 Computational linguistics.
Artificial Intelligence 2004
Dec 11, Human Parsing Do people use probabilities for parsing?! Sentence processing Study of Human Parsing.
CPSC 422, Lecture 27Slide 1 Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 27 Nov, 16, 2015.
GRAMMARS David Kauchak CS457 – Spring 2011 some slides adapted from Ray Mooney.
Entropy Reduction Model Resource: The Information Conveyed by Words in Sentences, John Hale.
PARSING David Kauchak CS159 – Fall Admin Assignment 3 Quiz #1  High: 36  Average: 33 (92%)  Median: 33.5 (93%)
Syntactic Parsing: Part I Niranjan Balasubramanian Stony Brook University March 8 th and Mar 22 nd, 2016 Many slides adapted from: Ray Mooney, Michael.
Roadmap Probabilistic CFGs –Handling ambiguity – more likely analyses –Adding probabilities Grammar Parsing: probabilistic CYK Learning probabilities:
Statistical Parsing IP disclosure: Content borrowed from J&M 3 rd edition and Raymond Mooney.
Natural Language Processing Vasile Rus
COSC 6336 Natural Language Processing Statistical Parsing
Statistical NLP: Lecture 3
Basic Parsing with Context Free Grammars Chapter 13
CKY Parser 0Book 1 the 2 flight 3 through 4 Houston5 6/19/2018
Probabilistic CKY Parser
CS 388: Natural Language Processing: Statistical Parsing
Probabilistic and Lexicalized Parsing
CKY Parser 0Book 1 the 2 flight 3 through 4 Houston5 11/16/2018
CS 388: Natural Language Processing: Syntactic Parsing
CSCI 5832 Natural Language Processing
Natural Language - General
Linguistic Essentials
David Kauchak CS159 – Spring 2019
David Kauchak CS159 – Spring 2019
Presentation transcript:

Introduction and Jurafsky Model Resource: A Probabilistic Model of Lexical and Syntactic Access and Disambiguation, Jurafsky 1996

 Process of accessing words? or  What happens after all the words are recognized?  Analyze it  Evaluate it  Creating new knowledge  Thinking guided by print?  Psycholinguistics

 What is constructed when we comprehend a sentence?  Propositional representation  What role do words play?  Mental lexicon  How does the process of constructing representation occur?  How is the ability to construct a representation acquired?

 Reading time: total time by a reader to read a peace of text  Reading time experiments  Word by word reading  Sentence by sentence reading  Eye tracking experiment

Source: Eye Tracking While Reading (ETWR) Experiment

Reading time is higher for ambiguous sentences

 Probabilistic Context Free Grammar (PCFG) model  Entropy Reduction model  Competition-Integration model  Connectionist model

 Probabilistic Context Free Grammar (PCFG) model  Entropy Reduction model  Competition-Integration model  Connectionist model

 Assumptions by Jurafsky (1996)  Observed preferences in interpretation of ambiguous sentences reflect probabilities of different syntactic structures.  Processing difficulty is a continuum ▪ Slight preferences at one end ▪ Garden path constructs at another end  Several types of ambiguities  Lexical category ambiguity  Attachment ambiguity  Unexpected thematic fit  Main clause vs. reduced relative clause ambiguity

 Garden path: to be led down to the garden path  To be misled or deceived  A garden path sentence [Wikipedia]  Grammatically correct sentence  Starts in such a way that reader’s most likely interpretation will be incorrect  Reader will be lured to a dead-end parse  Garden path sentences will be marked with # now on

(1) #The old man the boat. (2) #The horse raced (3) #The complex houses married and single soldiers and their families. (4) #The government plans to raise taxes were defeated. past the barnfell.

 Ambiguity resolved without trouble (fires = N or V): 1a. The warehouse fires destroyed all the buildings 1b. The warehouse fires a dozen employees each year.  Ambiguity leads to garden path (complex= N or Adj, houses= N or V, etc.): 2a. #The complex houses married and single students. 2b. #The old man the boats

(1) The spy saw the policeman with binocular (2) The spy saw the policeman with a revolver (3) The bird saw the birdwatcher with binocular

 Prepositional phrase can attach to NP or VP. 1. I saw the man with the glasses

 The arguments required by a verb are its subcategorization frame or valence.  Primary arguments  Secondary arguments  Attachment preferences vary between verbs (1)The women discussed the dogs on the beach. a.The women discussed the dogs which were on the beach. (90%) b.The women discussed them (the dogs) while on the beach. (10%) (2)The women kept the dogs on the beach. a.The women kept the dogs which were on the beach. (5%) b.The women kept them (the dogs) on the beach. (95%)

(1) The cop arrested by the detective was guilty of taking bribes (2) The crook arrested by the detective was guilty of taking bribes  (1) introduces more disambiguation difficulty as the initial noun phrase (the cop) is a good agent of the first verb (arrested)

 Reduced relative clause: that-clause without the that. a. #The horse raced past the barn fell. a’. #The horse (that) raced past the barn fell. b. The horse found in the woods died b’. The horse (that was) found in the woods died  Another case of different subcategorization preferences:  X raced >> X raced Y (intransitive preferred over transitive)  X found Y >> X found (transitive preferred over intransitive)

 if multiple rules can apply, choose one based on a selection rule  determinism  example selection rule: minimal attachment (choose the tree with the fewest nodes).  if parse fails, backtrack to choice point and reparse  backtracking occurs, causes increased processing times.

 if multiple rules can apply, pursue all possibilities in parallel  non-determinism  if any parse fails, discard it;  problem: number of parse trees can grow exponentially  solution: only pursue a limited number of possibilities (bounded parallelism).  Prune some of the unpromising parses  garden path means correct tree was pruned from search space;  backtracking occurs, causes increased processing times.

 How to model non-determinism  Probabilistic parsing  Parsing model [Jurafsky (1996)]  Each full or partial parse is assigned a probability.  Parses are pruned from the search space if their probability is a factor of α below the most probable parse (beam search).  How are parse probabilities determined?

 Grammar for a language is a set of rewrite rules Non-terminal Symbol Terminal Symbol Context-Free Context-Sensitive

Simple PCFG S → NP VP S → Aux NP VP S → VP NP → Pronoun NP → Proper-Noun NP → Det Nominal Nominal → Noun Nominal → Nominal Noun Nominal → Nominal PP VP → Verb VP → Verb NP VP → VP PP PP → Prep NP Grammar Prob Det → the | a | that | this Noun → book | flight | meal | money Verb → book | include | prefer Pronoun → I | he | she | me Proper-Noun → Houston | NWA Aux → does 1.0 Prep → from | to | on | near | through Lexicon

Probabilistic Grammar Conversion S → NP VP S → Aux NP VP S → VP NP → Pronoun NP → Proper-Noun NP → Det Nominal Nominal → Noun Nominal → Nominal Noun Nominal → Nominal PP VP → Verb VP → Verb NP VP → VP PP PP → Prep NP Original Grammar Chomsky Normal Form S → NP VP S → X1 VP X1 → Aux NP S → book | include | prefer S → Verb NP S → VP PP NP → I | he | she | me NP → Houston | NWA NP → Det Nominal Nominal → book | flight | meal | money Nominal → Nominal Noun Nominal → Nominal PP VP → book | include | prefer VP → Verb NP VP → VP PP PP → Prep NP

 Subcategorization frames of the verb keep:  Valence probabilities tell us how likely each of these frames is.

 Like PCFG probabilities, valence probabilities are estimated from a treebank.

Garden path caused by construction probabilities

Disambiguation using construction probabilities, no garden path “The warehouse fires destroyed all the buildings”

Disambiguation using construction probabilities, no garden path “The warehouse fires a dozen employees each year.”

Disambiguation using valence probabilities, no garden path: “keep the dogs on the beach”

Disambiguation using valence probabilities, no garden path: “keep the dogs on the beach”

Disambiguation using valence probabilities, no garden path: “discuss the dogs on the beach”

Disambiguation using valence probabilities, no garden path: “discuss the dogs on the beach”

Garden path caused by construction probabilities and valence probabilities: (main verb interpretation) “the horse raced past……”

Garden path caused by construction probabilities and valence probabilities: (reduced relative interpretation) “the horse raced past……”

Disambiguation using construction probabilities and valence probabilities, no garden path: (main verb) “The bird found in the room died”

Disambiguation using construction probabilities and valence probabilities, no garden path: (reduced relative) “The bird found in the room died”

Crucial assumption: if the relative probability of a tree falls below a certain value, then it will be pruned. Assumption: a garden path occurs if the probability ratio is higher than 5:1.