CSCI 5832 Natural Language Processing

Slides:



Advertisements
Similar presentations
School of something FACULTY OF OTHER School of Computing FACULTY OF ENGINEERING Parsing: computing the grammatical structure of English sentences COMP3310.
Advertisements

Mrach 1, 2009Dr. Muhammed Al-Mulhem1 ICS482 Formal Grammars Chapter 12 Muhammed Al-Mulhem March 1, 2009.
Syntactic analysis using Context Free Grammars. Analysis of language Morphological analysis – Chairs, Part Of Speech (POS) tagging – The/DT man/NN left/VBD.
Lecture 1, 7/21/2005Natural Language Processing1 CS60057 Speech &Natural Language Processing Autumn 2007 Lecture August 2007.
Parsing with Context Free Grammars Reading: Chap 13, Jurafsky & Martin
1 LING 180 Autumn 2007 LINGUIST 180: Introduction to Computational Linguistics Dan Jurafsky, Marie-Catherine de Marneffe Lecture 9: Grammar and Parsing.
Syntax and Context-Free Grammars CMSC 723: Computational Linguistics I ― Session #6 Jimmy Lin The iSchool University of Maryland Wednesday, October 7,
Syntactic Parsing with CFGs CMSC 723: Computational Linguistics I ― Session #7 Jimmy Lin The iSchool University of Maryland Wednesday, October 14, 2009.
Parsing SLP Chapter 13. 7/2/2015 Speech and Language Processing - Jurafsky and Martin 2 Outline  Parsing with CFGs  Bottom-up, top-down  CKY parsing.
CPSC 503 Computational Linguistics
CS 4705 Lecture 11 Feature Structures and Unification Parsing.
1 Basic Parsing with Context Free Grammars Chapter 13 September/October 2012 Lecture 6.
11 CS 388: Natural Language Processing: Syntactic Parsing Raymond J. Mooney University of Texas at Austin.
Context Free Grammars Reading: Chap 12-13, Jurafsky & Martin This slide set was adapted from J. Martin, U. Colorado Instructor: Paul Tarau, based on Rada.
School of something FACULTY OF OTHER School of Computing FACULTY OF ENGINEERING COMP3310 Natural Language Processing Eric Atwell, Language Research Group.
1 Features and Unification Chapter 15 October 2012 Lecture #10.
Context Free Grammars Reading: Chap 12-13, Jurafsky & Martin This slide set was adapted from J. Martin and Rada Mihalcea.
Speech and Language Processing Lecture 12—02/24/2015 Susan W. Brown.
Session 13 Context-Free Grammars and Language Syntax Introduction to Speech and Natural Language Processing (KOM422 ) Credits: 3(3-0)
1 Syntax Sudeshna Sarkar 25 Aug Sentence-Types Declaratives: A plane left S -> NP VP Imperatives: Leave! S -> VP Yes-No Questions: Did the plane.
1 CPE 480 Natural Language Processing Lecture 5: Parser Asst. Prof. Nuttanart Facundes, Ph.D.
CS 4705 Parsing More Efficiently and Accurately. Review Top-Down vs. Bottom-Up Parsers Left-corner table provides more efficient look- ahead Left recursion.
10/3/2015CPSC503 Winter CPSC 503 Computational Linguistics Lecture 8 Giuseppe Carenini.
Parsing with Context Free Grammars CSC 9010 Natural Language Processing Paula Matuszek and Mary-Angela Papalaskari This slide set was adapted from: Jim.
Context Free Grammars Reading: Chap 9, Jurafsky & Martin This slide set was adapted from J. Martin, U. Colorado Instructor: Rada Mihalcea.
6/2/2016CPSC503 Winter CPSC 503 Computational Linguistics Lecture 9 Giuseppe Carenini.
1 LIN6932 Spring 2007 LIN6932 Topics in Computational Linguistics Lecture 6: Grammar and Parsing (I) February 15, 2007 Hana Filip.
Rules, Movement, Ambiguity
PARSING 2 David Kauchak CS159 – Spring 2011 some slides adapted from Ray Mooney.
Natural Language Processing Lecture 15—10/15/2015 Jim Martin.
Natural Language Processing Lecture 14—10/13/2015 Jim Martin.
1/22/2016CPSC503 Winter CPSC 503 Computational Linguistics Lecture 8 Giuseppe Carenini.
PARSING David Kauchak CS159 – Fall Admin Assignment 3 Quiz #1  High: 36  Average: 33 (92%)  Median: 33.5 (93%)
Speech and Language Processing Formal Grammars Chapter 12.
Speech and Language Processing SLP Chapter 13 Parsing.
Parsing with Context Free Grammars. Slide 1 Outline Why should you care? Parsing Top-Down Parsing Bottom-Up Parsing Bottom-Up Space (an example) Top -
Context Free Grammars. Slide 1 Syntax Syntax = rules describing how words can connect to each other * that and after year last I saw you yesterday colorless.
Natural Language Processing Vasile Rus
CSC 594 Topics in AI – Natural Language Processing
Parsing Recommended Reading: Ch th Jurafsky & Martin 2nd edition
CSC 594 Topics in AI – Natural Language Processing
Heng Ji September 13, 2016 SYNATCTIC PARSING Heng Ji September 13, 2016.
Speech and Language Processing
Basic Parsing with Context Free Grammars Chapter 13
Parsing Recommended Reading: Ch th Jurafsky & Martin 2nd edition
CSC NLP -Context-Free Grammars
CPSC 503 Computational Linguistics
Lecture 13: Grammar and Parsing (I) November 9, 2004 Dan Jurafsky
CSCI 5832 Natural Language Processing
CSC 594 Topics in AI – Natural Language Processing
CSCI 5832 Natural Language Processing
CPSC 503 Computational Linguistics
CPSC 503 Computational Linguistics
CSCI 5832 Natural Language Processing
CSCI 5832 Natural Language Processing
CSCI 5832 Natural Language Processing
CS 388: Natural Language Processing: Syntactic Parsing
CSCI 5832 Natural Language Processing
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 25
CSCI 5832 Natural Language Processing
Probabilistic and Lexicalized Parsing
CSCI 5832 Natural Language Processing
Natural Language - General
CSCI 5832 Natural Language Processing
Heng Ji January 17, 2019 SYNATCTIC PARSING Heng Ji January 17, 2019.
CPSC 503 Computational Linguistics
CPSC 503 Computational Linguistics
CPSC 503 Computational Linguistics
David Kauchak CS159 – Spring 2019
David Kauchak CS159 – Spring 2019
Presentation transcript:

CSCI 5832 Natural Language Processing Lecture 11 Jim Martin 1/15/2019 CSCI 5832 Spring 2007

Today: 2/22 More on CFGs and English grammar facts Break Parsing with CFGs Project discussions 1/15/2019 CSCI 5832 Spring 2007

Context-Free Grammars Capture constituency and ordering Ordering is easy (well not really) What are the rules that govern the ordering of words and bigger units in the language What’s constituency? How words group into units 1/15/2019 CSCI 5832 Spring 2007

CFG Examples S -> NP VP NP -> Det NOMINAL NOMINAL -> Noun VP -> Verb Det -> a Noun -> flight Verb -> left 1/15/2019 CSCI 5832 Spring 2007

Problems for CFGs Agreement Subcategorization Movement 1/15/2019 CSCI 5832 Spring 2007

Agreement This dog Those dogs This dog eats Those dogs eat *This dogs *Those dogs eats 1/15/2019 CSCI 5832 Spring 2007

Agreement In English, subjects and verbs have to agree in person and number Determiners and nouns have to agree in number Many languages have agreement systems that are far more complex than this. 1/15/2019 CSCI 5832 Spring 2007

Possible CFG Solution S -> NP VP NP -> Det Nominal VP -> V NP … SgS -> SgNP SgVP PlS -> PlNp PlVP SgNP -> SgDet SgNom PlNP -> PlDet PlNom PlVP -> PlV NP SgVP ->SgV Np … 1/15/2019 CSCI 5832 Spring 2007

CFG Solution for Agreement It works and stays within the power of CFGs But its ugly And it doesn’t scale all that well 1/15/2019 CSCI 5832 Spring 2007

Subcategorization Sneeze: John sneezed Find: Please find [a flight to NY]NP Give: Give [me]NP[a cheaper fare]NP Help: Can you help [me]NP[with a flight]PP Prefer: I prefer [to leave earlier]TO-VP Told: I was told [United has a flight]S … 1/15/2019 CSCI 5832 Spring 2007

Forward Pointer It turns out that verb subcategorization facts will provide a key element for semantic analysis (determining who did what to who in an event). 1/15/2019 CSCI 5832 Spring 2007

Subcategorization *John sneezed the book *I prefer United has a flight *Give with a flight Subcat expresses the constraints that a predicate (verb for now) places on the number and type of the argument it wants to take 1/15/2019 CSCI 5832 Spring 2007

So? So the various rules for VPs overgenerate. They permit the presence of strings containing verbs and arguments that don’t go together For example VP -> V NP therefore Sneezed the book is a VP since “sneeze” is a verb and “the book” is a valid NP 1/15/2019 CSCI 5832 Spring 2007

Possible CFG Solution VP -> V VP -> V NP VP -> V NP PP … VP -> IntransV VP -> TransV NP VP -> TransPP NP PP … 1/15/2019 CSCI 5832 Spring 2007

Movement Core (canonical) example My travel agent booked the flight 1/15/2019 CSCI 5832 Spring 2007

Movement Core example [[My travel agent]NP [booked [the flight]NP]VP]S I.e. “book” is a straightforward transitive verb. It expects a single NP arg within the VP as an argument, and a single NP arg as the subject. 1/15/2019 CSCI 5832 Spring 2007

Movement What about? Which flight do you want me to have the travel agent book? The direct object argument to “book” isn’t appearing in the right place. It is in fact a long way from where its supposed to appear. And note that its separated from its verb by 2 other verbs. 1/15/2019 CSCI 5832 Spring 2007

The Point CFGs appear to be just about what we need to account for a lot of basic syntactic structure in English. But there are problems That can be dealt with adequately, although not elegantly, by staying within the CFG framework. There are simpler, more elegant, solutions that take us out of the CFG framework (beyond its formal power) 1/15/2019 CSCI 5832 Spring 2007

Break 1/15/2019 CSCI 5832 Spring 2007

Break Quiz Average was X True or False Distributional and Morphological Evidence aaa*b* or aa+b* or aa*ab* They have a non-zero intersection: The machine in 4 accepts some of strings in L1 but not all. It accepts some strings not in L1 as well. That is, with respect to L1 it makes two kinds of errors false positives and false negatives. 1/15/2019 CSCI 5832 Spring 2007

Break 5 aaaabbb a) b)  a b 3 1 2 a b 3 2 1 1/15/2019 2 a b 3 2 1 1/15/2019 CSCI 5832 Spring 2007

Quiz c* = (c + 1) N_c+1/Nc c*_1 = (1 + 1) N_2/N_1 = 2 * (1/1) = 2 1/15/2019 CSCI 5832 Spring 2007

Colloquium Today’s CS colloquium talk is on NLP. That’s at 3:30 in ECCR 265. There’s also an interesting talk tomorrow at noon in Muenzinger E214 on issues related to ML/NLP. 1/15/2019 CSCI 5832 Spring 2007

Parsing Parsing with CFGs refers to the task of assigning correct trees to input strings Correct here means a tree that covers all and only the elements of the input and has an S at the top It doesn’t actually mean that the system can select the correct tree from among all the possible trees 1/15/2019 CSCI 5832 Spring 2007

Parsing As with everything of interest, parsing involves a search which involves the making of choices We’ll start with some basic (meaning bad) methods before moving on to the one or two that you need to know 1/15/2019 CSCI 5832 Spring 2007

For Now Assume… You have all the words already in some buffer The input isn’t POS tagged We won’t worry about morphological analysis All the words are known 1/15/2019 CSCI 5832 Spring 2007

Top-Down Parsing Since we’re trying to find trees rooted with an S (Sentences) start with the rules that give us an S. Then work your way down from there to the words. 1/15/2019 CSCI 5832 Spring 2007

Top Down Space 1/15/2019 CSCI 5832 Spring 2007

Bottom-Up Parsing Of course, we also want trees that cover the input words. So start with trees that link up with the words in the right way. Then work your way up from there. 1/15/2019 CSCI 5832 Spring 2007

Bottom-Up Space 1/15/2019 CSCI 5832 Spring 2007

Bottom Up Space 1/15/2019 CSCI 5832 Spring 2007

Control Of course, in both cases we left out how to keep track of the search space and how to make choices Which node to try to expand next Which grammar rule to use to expand a node 1/15/2019 CSCI 5832 Spring 2007

Top-Down and Bottom-Up Only searches for trees that can be answers (i.e. S’s) But also suggests trees that are not consistent with any of the words Bottom-up Only forms trees consistent with the words But suggest trees that make no sense globally 1/15/2019 CSCI 5832 Spring 2007

Problems Even with the best filtering, backtracking methods are doomed if they don’t address certain problems Ambiguity Shared subproblems 1/15/2019 CSCI 5832 Spring 2007

Ambiguity 1/15/2019 CSCI 5832 Spring 2007

Shared Sub-Problems No matter what kind of search (top-down or bottom-up or mixed) that we choose. We don’t want to unnecessarily redo work we’ve already done. 1/15/2019 CSCI 5832 Spring 2007

Shared Sub-Problems Consider A flight from Indianapolis to Houston on TWA 1/15/2019 CSCI 5832 Spring 2007

Shared Sub-Problems Assume a top-down parse making bad initial choices on the Nominal rule. In particular… Nominal -> Nominal Noun Nominal -> Nominal PP 1/15/2019 CSCI 5832 Spring 2007

Shared Sub-Problems 1/15/2019 CSCI 5832 Spring 2007

Shared Sub-Problems 1/15/2019 CSCI 5832 Spring 2007

Shared Sub-Problems 1/15/2019 CSCI 5832 Spring 2007

Shared Sub-Problems 1/15/2019 CSCI 5832 Spring 2007