CS344: Introduction to Artificial Intelligence Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture 20-21– Natural Language Parsing.

Slides:



Advertisements
Similar presentations
Linguistics Lecture-12: X-bar Theory; Adjuncts and Complements
Advertisements

CS460/626 : Natural Language Processing/Speech, NLP and the Web (Lecture 20– Parsing) Pushpak Bhattacharyya CSE Dept., IIT Bombay 28 th Feb, 2011.
Computational language: week 10 Lexical Knowledge Representation concluded Syntax-based computational language Sentence structure: syntax Context free.
CS460/IT632 Natural Language Processing/Language Technology for the Web Lecture 2 (06/01/06) Prof. Pushpak Bhattacharyya IIT Bombay Part of Speech (PoS)
Grammars, constituency and order A grammar describes the legal strings of a language in terms of constituency and order. For example, a grammar for a fragment.
May 2006CLINT-LN Parsing1 Computational Linguistics Introduction Approaches to Parsing.
Properties of X-bar Complements, Adjuncts, & Specifiers.
Dr. Bibhuti Mahapatra, KReSIT, IIT BombayIntroduction to Linguistics1 Principle of structure dependency Bibhuti Bhusan Mahapatra.
Artificial Intelligence 2004 Natural Language Processing - Syntax and Parsing - Language Syntax Parsing.
Phrase Structure The formal means of representing constituency.
Basic Parsing with Context- Free Grammars 1 Some slides adapted from Julia Hirschberg and Dan Jurafsky.
Dr. Ansa Hameed Syntax (4).
Constituency Tests Phrase Structure Rules
Syntax Nuha AlWadaani.
11 CS 388: Natural Language Processing: Syntactic Parsing Raymond J. Mooney University of Texas at Austin.
LING/C SC/PSYC 438/538 Lecture 19 Sandiway Fong 1.
CS626: NLP, Speech and the Web Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture 12,13: Parsing 23 rd and 27 th August, 2012.
PARSING David Kauchak CS457 – Fall 2011 some slides adapted from Ray Mooney.
LI 2013 NATHALIE F. MARTIN S YNTAX. Grammatical vs Ungrammatical.
1 LIN 1310B Introduction to Linguistics Prof: Nikolay Slavkov TA: Qinghua Tang CLASS 14, Feb 27, 2007.
Probabilistic Parsing Reading: Chap 14, Jurafsky & Martin This slide set was adapted from J. Martin, U. Colorado Instructor: Paul Tarau, based on Rada.
CS : Speech, Natural Language Processing and the Web/Topics in Artificial Intelligence Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture 12: Deeper.
1 Statistical Parsing Chapter 14 October 2012 Lecture #9.
CS : Language Technology for the Web/Natural Language Processing Pushpak Bhattacharyya CSE Dept., IIT Bombay Constituent Parsing and Algorithms (with.
May 2006CLINT-LN Parsing1 Computational Linguistics Introduction Parsing with Context Free Grammars.
NLP. Introduction to NLP Is language more than just a “bag of words”? Grammatical rules apply to categories and groups of words, not individual words.
October 2005csa3180: Parsing Algorithms 11 CSA350: NLP Algorithms Sentence Parsing I The Parsing Problem Parsing as Search Top Down/Bottom Up Parsing Strategies.
Parsing with Context Free Grammars CSC 9010 Natural Language Processing Paula Matuszek and Mary-Angela Papalaskari This slide set was adapted from: Jim.
Today Phrase structure rules, trees Constituents Recursion Conjunction
CS626: NLP, Speech and the Web Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture 15, 17: Parsing Ambiguity, Probabilistic Parsing, sample seminar 17.
PARSING David Kauchak CS159 – Spring 2011 some slides adapted from Ray Mooney.
Lecture 1, 7/21/2005Natural Language Processing1 CS60057 Speech &Natural Language Processing Autumn 2007 Lecture August 2007.
Syntax Why is the structure of language (syntax) important? How do we represent syntax? What does an example grammar for English look like? What strategies.
For Wednesday Read chapter 23 Homework: –Chapter 22, exercises 1,4, 7, and 14.
LING 388: Language and Computers Sandiway Fong Lecture 12.
Rules, Movement, Ambiguity
CS460/IT632 Natural Language Processing/Language Technology for the Web Lecture 17 (14/03/06) Prof. Pushpak Bhattacharyya IIT Bombay Formulation of Grammar.
CSA2050 Introduction to Computational Linguistics Parsing I.
Natural Language - General
PARSING 2 David Kauchak CS159 – Spring 2011 some slides adapted from Ray Mooney.
Natural Language Processing
CS 4705 Lecture 10 The Earley Algorithm. Review Top-Down vs. Bottom-Up Parsers –Both generate too many useless trees –Combine the two to avoid over-generation:
csa3050: Parsing Algorithms 11 CSA350: NLP Algorithms Parsing Algorithms 1 Top Down Bottom-Up Left Corner.
CS460/IT632 Natural Language Processing/Language Technology for the Web Lecture 6 (14/02/06) Prof. Pushpak Bhattacharyya IIT Bombay Top-Down and Bottom-Up.
LING/C SC/PSYC 438/538 Lecture 20 Sandiway Fong 1.
Parsing and Code Generation Set 24. Parser Construction Most of the work involved in constructing a parser is carried out automatically by a program,
CS : Speech, NLP and the Web/Topics in AI Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture 13: Deeper Adjective and PP Structure; Structural Ambiguity.
CS460/IT632 Natural Language Processing/Language Technology for the Web Lecture 13 (17/02/06) Prof. Pushpak Bhattacharyya IIT Bombay Top-Down Bottom-Up.
GRAMMARS & PARSING. Parser Construction Most of the work involved in constructing a parser is carried out automatically by a program, referred to as a.
NLP. Introduction to NLP #include int main() { int n, reverse = 0; printf("Enter a number to reverse\n"); scanf("%d",&n); while (n != 0) { reverse =
CS : Language Technology for the Web/Natural Language Processing Pushpak Bhattacharyya CSE Dept., IIT Bombay Parsing Algos.
PARSING David Kauchak CS159 – Fall Admin Assignment 3 Quiz #1  High: 36  Average: 33 (92%)  Median: 33.5 (93%)
Week 3. Clauses and Trees English Syntax. Trees and constituency A sentence has a hierarchical structure Constituents can have constituents of their own.
CS : Speech, NLP and the Web/Topics in AI Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture 11: Evidence for Deeper Structure; Top Down Parsing.
SYNTAX.
King Faisal University جامعة الملك فيصل Deanship of E-Learning and Distance Education عمادة التعلم الإلكتروني والتعليم عن بعد [ ] 1 King Faisal University.
4.3 The Generative Approach
Chapter Eight Syntax.
Part I: Basics and Constituency
CS : Speech, NLP and the Web/Topics in AI
LING/C SC/PSYC 438/538 Lecture 21 Sandiway Fong.
Probabilistic and Lexicalized Parsing
CS460/626 : Natural Language Processing/Speech, NLP and the Web (Lecture 23, 24– Parsing Algorithms; Parsing in case of Ambiguity; Probabilistic Parsing)
Probabilistic and Lexicalized Parsing
Chapter Eight Syntax.
Natural Language - General
Parsing and More Parsing
CS : Language Technology for the Web/Natural Language Processing
David Kauchak CS159 – Spring 2019
Presentation transcript:

CS344: Introduction to Artificial Intelligence Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture 20-21– Natural Language Parsing

Parsing of Sentences

Are sentences flat linear structures? Why tree? Is there a principle in branching When should the constituent give rise to children? What is the hierarchy building principle?

Structure Dependency: A Case Study  Interrogative Inversion (1)John will solve the problem. Will John solve the problem? DeclarativeInterrogative (2) a.Susan must leave. Must Susan leave? b.Harry can swim.Can Harry swim? c.Mary has read the book.Has Mary read the book? d. Bill is sleeping.Is Bill sleeping? ………………………………………………………. The section, “Structure dependency a case study” here is adopted from a talk given by Howard Lasnik (2003) in Delhi university.

Interrogative inversion Structure Independent (1 st attempt) (3)Interrogative inversion process Beginning with a declarative, invert the first and second words to construct an interrogative. DeclarativeInterrogative (4) a. The woman must leave.*Woman the must leave? b. A sailor can swim.*Sailor a can swim? c. No boy has read the book. *Boy no has read the book? d. My friend is sleeping. *Friend my is sleeping?

Interrogative inversion correct pairings Compare the incorrect pairings in (4) with the correct pairings in (5): DeclarativeInterrogative (5) a. The woman must leave.Must the woman leave? b. A sailor can swim.Can a sailor swim? c. No boy has read the book. Has no boy read the book? d. My friend is sleeping.Is my friend sleeping?

Interrogative inversion Structure Independent (2 nd attempt) (6) Interrogative inversion process: Beginning with a declarative, move the auxiliary verb to the front to construct an interrogative. DeclarativeInterrogative (7) a. Bill could be sleeping.*Be Bill could sleeping? Could Bill be sleeping? b. Mary has been reading.*Been Mary has reading? Has Mary been reading? c. Susan should have left.*Have Susan should left? Should Susan have left?

Structure independent (3 rd attempt): (8) Interrogative inversion process Beginning with a declarative, move the first auxiliary verb to the front to construct an interrogative. Declarative Interrogative (9) a. The man who is here can swim.*Is the man who here can swim? b. The boy who will play has left.*Will the boy who play has left?

Structure Dependent Correct Pairings For the above examples, fronting the second auxiliary verb gives the correct form: Declarative Interrogative (10) a.The man who is here can swim. Can the man who is here swim? b.The boy who will play has left. Has the boy who will play left?

Natural transformations are structure dependent (11) Does the child acquiring English learn these properties? (12) We are not dealing with a peculiarity of English. No known human language has a transformational process that would produce pairings like those in (4), (7) and (9), repeated below: (4) a. The woman must leave.*Woman the must leave? (7) a. Bill could be sleeping.*Be Bill could sleeping? (9) a. The man who is here can swim. *Is the man who here can swim?

Deeper trees needed for capturing sentence structure NP PPAP big The of poems with the blue cover [The big book of poems with the Blue cover] is on the table. book This wont do! Flat structure! PP

Other languages NP PPAP big The of poems with the blue cover [niil jilda vaalii kavita kii kitaab] book English NP PP AP niil jilda vaalii kavita kii kitaab PP badii Hindi PP

Other languages: contd NP PPAP big The of poems with the blue cover [niil malaat deovaa kavitar bai ti] book English NP PP AP niil malaat deovaa kavitar bai PP motaa Bengali PP ti

PPs are at the same level: flat with respect to the head word “book” NP PPAP big The of poems with the blue cover [The big book of poems with the Blue cover] is on the table. book No distinction in terms of dominance or c-command PP

“Constituency test of Replacement” runs into problems One-replacement: I bought the big [book of poems with the blue cover] not the small [one] One-replacement targets book of poems with the blue cover Another one-replacement: I bought the big [book of poems] with the blue cover not the small [one] with the red cover One-replacement targets book of poems

More deeply embedded structure NP PP AP big The of poems with the blue cover N’ 1 N book PP N’ 2 N’ 3

To target N 1 ’ I want [ NP this [ N’ big book of poems with the red cover] and not [ N that [ N one]]

Bar-level projections Add intermediate structures NP  (D) N’ N’  (AP) N’ | N’ (PP) | N (PP) () indicates optionality

New rules produce this tree NP PP AP big The of poems with the blue cover N’ 1 N book PP N’ 2 N’ 3 N-bar

As opposed to this tree NP PPAP big The of poems with the blue coverbook PP

V-bar What is the element in verbs corresponding to one-replacement for nouns do-so or did-so

As opposed to this tree NP PPAP big The of poems with the blue coverbook PP

I [eat beans with a fork] VP NP beans eat with a fork PP No constituent that groups together V and NP and excludes PP

Need for intermediate constituents I [eat beans] with a fork but Ram [does so] with a spoon V2’V2’ NP beans eat with a fork PP VP V1’V1’ V VP  V’ V’  V’ (PP) V’  V (NP)

How to target V 1 ’ I [eat beans with a fork], and Ram [does so] too. V2’V2’ NP beans eat with a fork PP VP V1’V1’ V VP  V’ V’  V’ (PP) V’  V (NP)

Parsing Algorithms

A simplified grammar S  NP VP NP  DT N | N VP  V ADV | V

A segment of English Grammar S’  (C) S S  {NP/S’} VP VP  (AP+) (VAUX) V (AP+) ({NP/S’}) (AP+) (PP+) (AP+) NP  (D) (AP+) N (PP+) PP  P NP AP  (AP) A

Example Sentence People laugh Lexicon: People - N, V Laugh - N, V These are positions This indicate that both Noun and Verb is possible for the word “People”

Top-Down Parsing State Backup State Action ((S) 1) ((NP VP)1) - - 3a. ((DT N VP)1) ((N VP) 1) - 3b. ((N VP)1) ((VP)2) - Consume “People” 5a. ((V ADV)2) ((V)2) - 6. ((ADV)3) ((V)2) Consume “laugh” 5b. ((V)2) ((.)3) - Consume “laugh” Termination Condition : All inputs over. No symbols remaining. Note: Input symbols can be pushed back. Position of input pointer

Discussion for Top-Down Parsing This kind of searching is goal driven. Gives importance to textual precedence (rule precedence). No regard for data, a priori (useless expansions made).

Bottom-Up Parsing Some conventions: N 12 S 1? -> NP 12 ° VP 2? Represents positions End position unknown Work on the LHS done, while the work on RHS remaining

Bottom-Up Parsing (pictorial representation) S -> NP 12 VP 23 ° People Laugh N 12 N 23 V 12 V 23 NP 12 -> N 12 ° NP 23 -> N 23 ° VP 12 -> V 12 ° VP 23 -> V 23 ° S 1? -> NP 12 ° VP 2?

Problem with Top-Down Parsing Left Recursion Suppose you have A-> AB rule. Then we will have the expansion as follows: ((A)K) -> ((AB)K) -> ((ABB)K) ……..

Combining top-down and bottom-up strategies

Top-Down Bottom-Up Chart Parsing Combines advantages of top-down & bottom- up parsing. Does not work in case of left recursion. e.g. – “People laugh” People – noun, verb Laugh – noun, verb Grammar – S  NP VP NP  DT N | N VP  V ADV | V

Transitive Closure People laugh 123 S  NP VPNP  N  VP  V  NP  DT NS  NP  VPS  NP VP  NP  NVP  V ADVsuccess VP  V

Arcs in Parsing Each arc represents a chart which records Completed work (left of  ) Expected work (right of  )

Example People laughloudly 1234 S  NP VPNP  N  VP  V  VP  V ADV  NP  DT NS  NP  VPVP  V  ADVS  NP VP  NP  NVP   V ADVS  NP VP  VP   V

Dealing With Structural Ambiguity Multiple parses for a sentence The man saw the boy with a telescope. The man saw the mountain with a telescope. The man saw the boy with the ponytail. At the level of syntax, all these sentences are ambiguous. But semantics can disambiguate 2 nd & 3 rd sentence.

Prepositional Phrase (PP) Attachment Problem V – NP 1 – P – NP 2 (Here P means preposition) NP 2 attaches to NP 1 ? or NP 2 attaches to V ?

Parse Trees for a Structurally Ambiguous Sentence Let the grammar be – S  NP VP NP  DT N | DT N PP PP  P NP VP  V NP PP | V NP For the sentence, “I saw a boy with a telescope”

Parse Tree - 1 S NPVP NVNP DetNPP PNP DetN I saw a boy with atelescope

Parse Tree -2 S NPVP NVNP DetN PP PNP DetN I saw a boy with atelescope