Presentation is loading. Please wait.

Presentation is loading. Please wait.

CS460/626 : Natural Language Processing/Speech, NLP and the Web (Lecture 23, 24– Parsing Algorithms; Parsing in case of Ambiguity; Probabilistic Parsing)

Similar presentations


Presentation on theme: "CS460/626 : Natural Language Processing/Speech, NLP and the Web (Lecture 23, 24– Parsing Algorithms; Parsing in case of Ambiguity; Probabilistic Parsing)"— Presentation transcript:

1 CS460/626 : Natural Language Processing/Speech, NLP and the Web (Lecture 23, 24– Parsing Algorithms; Parsing in case of Ambiguity; Probabilistic Parsing) Pushpak Bhattacharyya CSE Dept., IIT Bombay 8th, 10th March, 2011 (Lectures 21 and 22 were on Sentiment Analysis by Aditya Joshi)

2 A note on Language Modeling
Example sentence “ ^ The tortoise beat the hare in the race.” Guided Guided by Guided by by frequency Language world Knowledge Knowledge N-gram (n=3) CFG Probabilistic CFG Dependency Grammar Prob. DG ^ the tortoise 5*10-3 S-> NP VP S->NP VP Semantic Roles agt, obj, sen, etc. Semantic Rules are always between “Heads” Semantic Roles with probabilities the tortoise beat *10-2 NP->DT N NP->DT N tortoise beat the 7*10-5 VP->V NP PP VP->V NP PP 0.4 beat the hare 5*10-6 PP-> P NP PP-> P NP

3 Parse Tree DT N V NP PP The Tortoise beat DT N P NP S NP VP
the hare in DT N the race

4 UNL Expression beat@past scn (scene) agt obj race@def tortoise@def

5 Purpose of LM Prediction of next word (Speech Processing)
Language Identification (for same script) Belongingness check (parsing) P(NP->DT N) means what is the probability that the ‘YIELD’ of the non terminal NP is DT N

6 Need for Deep Parsing Sentences are linear structures
But there is a hierarchy- a tree- hidden behind the linear structure There are constituents and branches

7 PPs are at the same level: flat with respect to the head word “book”
No distinction in terms of dominance or c-command NP PP The AP PP book with the blue cover of poems big [The big book of poems with the Blue cover] is on the table.

8 “Constituency test of Replacement” runs into problems
One-replacement: I bought the big [book of poems with the blue cover] not the small [one] One-replacement targets book of poems with the blue cover Another one-replacement: I bought the big [book of poems] with the blue cover not the small [one] with the red cover One-replacement targets book of poems

9 More deeply embedded structure
NP N’1 The AP N’2 N’3 big PP N book with the blue cover PP of poems

10 Grammar and Parsing Algorithms

11 A simplified grammar S  NP VP NP  DT N | N VP  V ADV | V

12 Example Sentence People laugh 2 3 Lexicon: People - N, V Laugh - N, V
Lexicon: People - N, V Laugh - N, V These are positions This indicate that both Noun and Verb is possible for the word “People”

13 Position of input pointer
Top-Down Parsing State Backup State Action 1. ((S) 1) 2. ((NP VP)1) 3a. ((DT N VP)1) ((N VP) 1) 3b. ((N VP)1) 4. ((VP)2) Consume “People” 5a. ((V ADV)2) ((V)2) 6. ((ADV)3) ((V)2) Consume “laugh” 5b. ((V)2) 6. ((.)3) Consume “laugh” Termination Condition : All inputs over. No symbols remaining. Note: Input symbols can be pushed back. Position of input pointer

14 Discussion for Top-Down Parsing
This kind of searching is goal driven. Gives importance to textual precedence (rule precedence). No regard for data, a priori (useless expansions made).

15 Work on the LHS done, while the work on RHS remaining
Bottom-Up Parsing Some conventions: N12 S1? -> NP12 ° VP2? Represents positions Work on the LHS done, while the work on RHS remaining End position unknown

16 Bottom-Up Parsing (pictorial representation)
S -> NP12 VP23 ° People Laugh N N23 V V23 NP12 -> N12 ° NP23 -> N23 ° VP12 -> V12 ° VP23 -> V23 ° S1? -> NP12 ° VP2?

17 Problem with Top-Down Parsing
Left Recursion Suppose you have A-> AB rule. Then we will have the expansion as follows: ((A)K) -> ((AB)K) -> ((ABB)K) ……..

18 Combining top-down and bottom-up strategies

19 Top-Down Bottom-Up Chart Parsing
Combines advantages of top-down & bottom-up parsing. Does not work in case of left recursion. e.g. – “People laugh” People – noun, verb Laugh – noun, verb Grammar – S  NP VP NP  DT N | N VP  V ADV | V

20 Transitive Closure People laugh 1 2 3 S NP VP NP N VP  V 
S NP VP NP N VP  V  NP DT N S  NPVP S  NP VP  NP N VP V ADV success VP V

21 Arcs in Parsing Each arc represents a chart which records
Completed work (left of ) Expected work (right of )

22 Example People laugh loudly 1 2 3 4
S  NP VP NP  N VP  V VP  V ADV NP  DT N S  NPVP VP  VADV S  NP VP NP  N VP  V ADV S  NP VP VP  V

23 Dealing With Structural Ambiguity
Multiple parses for a sentence The man saw the boy with a telescope. The man saw the mountain with a telescope. The man saw the boy with the ponytail. At the level of syntax, all these sentences are ambiguous. But semantics can disambiguate 2nd & 3rd sentence.

24 Prepositional Phrase (PP) Attachment Problem
V – NP1 – P – NP2 (Here P means preposition) NP2 attaches to NP1 ? or NP2 attaches to V ?

25 Parse Trees for a Structurally Ambiguous Sentence
Let the grammar be – S  NP VP NP  DT N | DT N PP PP  P NP VP  V NP PP | V NP For the sentence, “I saw a boy with a telescope”

26 Parse Tree - 1 S NP VP N V NP Det N PP P NP Det N saw I a boy with a
telescope

27 Parse Tree -2 S NP VP PP N V NP P NP Det N Det N saw I a with boy a
telescope

28 Parsing Structural Ambiguity

29 Parsing for Structurally Ambiguous Sentences
Sentence “I saw a boy with a telescope” Grammar: S  NP VP NP  ART N | ART N PP | PRON VP  V NP PP | V NP ART  a | an | the N  boy | telescope PRON  I V  saw

30 Ambiguous Parses Two possible parses:
PP attached with Verb (i.e. I used a telescope to see) ( S ( NP ( PRON “I” ) ) ( VP ( V “saw” ) ( NP ( (ART “a”) ( N “boy”)) ( PP (P “with”) (NP ( ART “a” ) ( N “telescope”))))) PP attached with Noun (i.e. boy had a telescope) ( NP ( (ART “a”) ( N “boy”) (PP (P “with”) (NP ( ART “a” ) ( N “telescope”))))))

31 Top Down Parse State Backup State 1 ( ( S ) 1 ) − Action Comments
Use S  NP VP

32 Use NP  ART N | ART N PP | PRON
Top Down Parse State Backup State Action Comments 1 ( ( S ) 1 ) Use S  NP VP 2 ( ( NP VP ) 1 ) Use NP  ART N | ART N PP | PRON

33 Top Down Parse State Backup State 1 ( ( S ) 1 ) − 2 ( ( NP VP ) 1 ) 3
Action Comments 1 ( ( S ) 1 ) Use S  NP VP 2 ( ( NP VP ) 1 ) Use NP  ART N | ART N PP | PRON 3 ( ( ART N VP ) 1 ) ( ( ART N PP VP ) 1 ) ( ( PRON VP ) 1) ART does not match “I”, backup state (b) used

34 Top Down Parse State Backup State 1 ( ( S ) 1 ) − 2 ( ( NP VP ) 1 ) 3
Action Comments 1 ( ( S ) 1 ) Use S  NP VP 2 ( ( NP VP ) 1 ) Use NP  ART N | ART N PP | PRON 3 ( ( ART N VP ) 1 ) ( ( ART N PP VP ) 1 ) ( ( PRON VP ) 1) ART does not match “I”, backup state (b) used 3B ( ( PRON VP ) 1 )

35 Top Down Parse State Backup State 1 ( ( S ) 1 ) − 2 ( ( NP VP ) 1 ) 3
Action Comments 1 ( ( S ) 1 ) Use S  NP VP 2 ( ( NP VP ) 1 ) Use NP  ART N | ART N PP | PRON 3 ( ( ART N VP ) 1 ) ( ( ART N PP VP ) 1 ) ( ( PRON VP ) 1) ART does not match “I”, backup state (b) used 3B ( ( PRON VP ) 1 ) 4 ( ( VP ) 2 ) Consumed “I”

36 Top Down Parse State Backup State 1 ( ( S ) 1 ) − 2 ( ( NP VP ) 1 ) 3
Action Comments 1 ( ( S ) 1 ) Use S  NP VP 2 ( ( NP VP ) 1 ) Use NP  ART N | ART N PP | PRON 3 ( ( ART N VP ) 1 ) ( ( ART N PP VP ) 1 ) ( ( PRON VP ) 1) ART does not match “I”, backup state (b) used 3B ( ( PRON VP ) 1 ) 4 ( ( VP ) 2 ) Consumed “I” 5 ( ( V NP PP ) 2 ) ( ( V NP ) 2 ) Verb Attachment Rule used

37 Top Down Parse State Backup State 1 ( ( S ) 1 ) − 2 ( ( NP VP ) 1 ) 3
Action Comments 1 ( ( S ) 1 ) Use S  NP VP 2 ( ( NP VP ) 1 ) Use NP  ART N | ART N PP | PRON 3 ( ( ART N VP ) 1 ) ( ( ART N PP VP ) 1 ) ( ( PRON VP ) 1) ART does not match “I”, backup state (b) used 3B ( ( PRON VP ) 1 ) 4 ( ( VP ) 2 ) Consumed “I” 5 ( ( V NP PP ) 2 ) ( ( V NP ) 2 ) Verb Attachment Rule used 6 ( ( NP PP ) 3 ) Consumed “saw”

38 Top Down Parse State Backup State 1 ( ( S ) 1 ) − 2 ( ( NP VP ) 1 ) 3
Action Comments 1 ( ( S ) 1 ) Use S  NP VP 2 ( ( NP VP ) 1 ) Use NP  ART N | ART N PP | PRON 3 ( ( ART N VP ) 1 ) ( ( ART N PP VP ) 1 ) ( ( PRON VP ) 1) ART does not match “I”, backup state (b) used 3B ( ( PRON VP ) 1 ) 4 ( ( VP ) 2 ) Consumed “I” 5 ( ( V NP PP ) 2 ) ( ( V NP ) 2 ) Verb Attachment Rule used 6 ( ( NP PP ) 3 ) Consumed “saw” 7 ( ( ART N PP ) 3 ) ( ( ART N PP PP ) 3 ) ( ( PRON PP ) 3 )

39 Top Down Parse State Backup State 1 ( ( S ) 1 ) − 2 ( ( NP VP ) 1 ) 3
Action Comments 1 ( ( S ) 1 ) Use S  NP VP 2 ( ( NP VP ) 1 ) Use NP  ART N | ART N PP | PRON 3 ( ( ART N VP ) 1 ) ( ( ART N PP VP ) 1 ) ( ( PRON VP ) 1) ART does not match “I”, backup state (b) used 3B ( ( PRON VP ) 1 ) 4 ( ( VP ) 2 ) Consumed “I” 5 ( ( V NP PP ) 2 ) ( ( V NP ) 2 ) Verb Attachment Rule used 6 ( ( NP PP ) 3 ) Consumed “saw” 7 ( ( ART N PP ) 3 ) ( ( ART N PP PP ) 3 ) ( ( PRON PP ) 3 ) 8 ( ( N PP) 4 ) Consumed “a”

40 Top Down Parse State Backup State 1 ( ( S ) 1 ) − 2 ( ( NP VP ) 1 ) 3
Action Comments 1 ( ( S ) 1 ) Use S  NP VP 2 ( ( NP VP ) 1 ) Use NP  ART N | ART N PP | PRON 3 ( ( ART N VP ) 1 ) ( ( ART N PP VP ) 1 ) ( ( PRON VP ) 1) ART does not match “I”, backup state (b) used 3B ( ( PRON VP ) 1 ) 4 ( ( VP ) 2 ) Consumed “I” 5 ( ( V NP PP ) 2 ) ( ( V NP ) 2 ) Verb Attachment Rule used 6 ( ( NP PP ) 3 ) Consumed “saw” 7 ( ( ART N PP ) 3 ) ( ( ART N PP PP ) 3 ) ( ( PRON PP ) 3 ) 8 ( ( N PP) 4 ) Consumed “a” 9 ( ( PP ) 5 ) Consumed “boy”

41 Top Down Parse State Backup State 1 ( ( S ) 1 ) − 2 ( ( NP VP ) 1 ) 3
Action Comments 1 ( ( S ) 1 ) Use S  NP VP 2 ( ( NP VP ) 1 ) Use NP  ART N | ART N PP | PRON 3 ( ( ART N VP ) 1 ) ( ( ART N PP VP ) 1 ) ( ( PRON VP ) 1) ART does not match “I”, backup state (b) used 3B ( ( PRON VP ) 1 ) 4 ( ( VP ) 2 ) Consumed “I” 5 ( ( V NP PP ) 2 ) ( ( V NP ) 2 ) Verb Attachment Rule used 6 ( ( NP PP ) 3 ) Consumed “saw” 7 ( ( ART N PP ) 3 ) ( ( ART N PP PP ) 3 ) ( ( PRON PP ) 3 ) 8 ( ( N PP) 4 ) Consumed “a” 9 ( ( PP ) 5 ) Consumed “boy” 10 ( ( P NP ) )

42 Top Down Parse State Backup State 1 ( ( S ) 1 ) − … 7
Action Comments 1 ( ( S ) 1 ) Use S  NP VP 7 ( ( ART N PP ) 3 ) ( ( ART N PP PP ) 3 ) ( ( PRON PP ) 3 ) 8 ( ( N PP) 4 ) Consumed “a” 9 ( ( PP ) 5 ) Consumed “boy” 10 ( ( P NP ) 5 ) 11 ( ( NP ) 6 ) Consumed “with”

43 Top Down Parse State Backup State 1 ( ( S ) 1 ) − … 7
Action Comments 1 ( ( S ) 1 ) Use S  NP VP 7 ( ( ART N PP ) 3 ) ( ( ART N PP PP ) 3 ) ( ( PRON PP ) 3 ) 8 ( ( N PP) 4 ) Consumed “a” 9 ( ( PP ) 5 ) Consumed “boy” 10 ( ( P NP ) 5 ) 11 ( ( NP ) 6 ) Consumed “with” 12 ( ( ART N ) 6 ) ( ( ART N PP ) 6 ) ( ( PRON ) 6)

44 Top Down Parse State Backup State 1 ( ( S ) 1 ) − … 7
Action Comments 1 ( ( S ) 1 ) Use S  NP VP 7 ( ( ART N PP ) 3 ) ( ( ART N PP PP ) 3 ) ( ( PRON PP ) 3 ) 8 ( ( N PP) 4 ) Consumed “a” 9 ( ( PP ) 5 ) Consumed “boy” 10 ( ( P NP ) 5 ) 11 ( ( NP ) 6 ) Consumed “with” 12 ( ( ART N ) 6 ) ( ( ART N PP ) 6 ) ( ( PRON ) 6) 13 ( ( N ) 7 )

45 Top Down Parse State Backup State 1 ( ( S ) 1 ) − … 7
Action Comments 1 ( ( S ) 1 ) Use S  NP VP 7 ( ( ART N PP ) 3 ) ( ( ART N PP PP ) 3 ) ( ( PRON PP ) 3 ) 8 ( ( N PP) 4 ) Consumed “a” 9 ( ( PP ) 5 ) Consumed “boy” 10 ( ( P NP ) 5 ) 11 ( ( NP ) 6 ) Consumed “with” 12 ( ( ART N ) 6 ) ( ( ART N PP ) 6 ) ( ( PRON ) 6) 13 ( ( N ) 7 ) 14 ( ( − ) 8 ) Consume “telescope” Finish Parsing

46 Top Down Parsing - Observations
Top down parsing gave us the Verb Attachment Parse Tree (i.e., I used a telescope) To obtain the alternate parse tree, the backup state in step 5 will have to be invoked Is there an efficient way to obtain all parses ?

47 Bottom Up Parse I saw a boy with a telescope 1 2 3 4 5 6 7 8
Colour Scheme : Blue for Normal Parse Green for Verb Attachment Parse Purple for Noun Attachment Parse Red for Invalid Parse

48 Bottom Up Parse I saw a boy with a telescope 1 2 3 4 5 6 7 8
NP12  PRON12 S1?NP12VP2?

49 Bottom Up Parse I saw a boy with a telescope 1 2 3 4 5 6 7 8
NP12  PRON12 VP2?V23NP3? S1?NP12VP2? VP2?V23NP3?PP??

50 Bottom Up Parse I saw a boy with a telescope 1 2 3 4 5 6 7 8
NP12  PRON12 VP2?V23NP3? NP35 ART34N45 S1?NP12VP2? VP2?V23NP3?PP?? NP3?ART34N45PP5?

51 Bottom Up Parse I saw a boy with a telescope 1 2 3 4 5 6 7 8
NP12  PRON12 VP2?V23NP3? NP35 ART34N45 NP35ART34N45 S1?NP12VP2? VP2?V23NP3?PP?? NP3?ART34N45PP5? NP3?ART34N45PP5?

52 Bottom Up Parse I saw a boy with a telescope 1 2 3 4 5 6 7 8
NP12  PRON12 VP2?V23NP3? NP35 ART34N45 NP35ART34N45 S1?NP12VP2? VP2?V23NP3?PP?? NP3?ART34N45PP5? NP3?ART34N45PP5? VP25V23NP35 VP2?V23NP35PP5? S15NP12VP25

53 Bottom Up Parse I saw a boy with a telescope 1 2 3 4 5 6 7 8
NP12  PRON12 VP2?V23NP3? NP35 ART34N45 NP35ART34N45 PP5?P56NP6? S1?NP12VP2? VP2?V23NP3?PP?? NP3?ART34N45PP5? NP3?ART34N45PP5? VP25V23NP35 VP2?V23NP35PP5? S15NP12VP25

54 Bottom Up Parse I saw a boy with a telescope 1 2 3 4 5 6 7 8
NP12  PRON12 VP2?V23NP3? NP35 ART34N45 NP35ART34N45 PP5?P56NP6? NP68ART67N7? S1?NP12VP2? VP2?V23NP3?PP?? NP3?ART34N45PP5? NP3?ART34N45PP5? NP6?ART67N78PP8? VP25V23NP35 VP2?V23NP35PP5? S15NP12VP25

55 Bottom Up Parse I saw a boy with a telescope 1 2 3 4 5 6 7 8
NP12  PRON12 VP2?V23NP3? NP35 ART34N45 NP35ART34N45 PP5?P56NP6? NP68ART67N7? NP68ART67N78 S1?NP12VP2? VP2?V23NP3?PP?? NP3?ART34N45PP5? NP3?ART34N45PP5? NP6?ART67N78PP8? VP25V23NP35 VP2?V23NP35PP5? S15NP12VP25

56 Bottom Up Parse I saw a boy with a telescope 1 2 3 4 5 6 7 8
NP12  PRON12 VP2?V23NP3? NP35 ART34N45 NP35ART34N45 PP5?P56NP6? NP68ART67N7? NP68ART67N78 S1?NP12VP2? VP2?V23NP3?PP?? NP3?ART34N45PP5? NP3?ART34N45PP5? NP6?ART67N78PP8? VP25V23NP35 VP2?V23NP35PP5? PP58P56NP68 S15NP12VP25

57 Bottom Up Parse I saw a boy with a telescope 1 2 3 4 5 6 7 8
NP12  PRON12 VP2?V23NP3? NP35 ART34N45 NP35ART34N45 PP5?P56NP6? NP68ART67N7? NP68ART67N78 S1?NP12VP2? VP2?V23NP3?PP?? NP3?ART34N45PP5? NP3?ART34N45PP5? NP6?ART67N78PP8? VP25V23NP35 VP2?V23NP35PP5? PP58P56NP68 S15NP12VP25 NP38ART34N45PP58

58 Bottom Up Parse I saw a boy with a telescope 1 2 3 4 5 6 7 8
NP12  PRON12 VP2?V23NP3? NP35 ART34N45 NP35ART34N45 PP5?P56NP6? NP68ART67N7? NP68ART67N78 S1?NP12VP2? VP2?V23NP3?PP?? NP3?ART34N45PP5? NP3?ART34N45PP5? NP6?ART67N78PP8? VP25V23NP35 VP2?V23NP35PP5? PP58P56NP68 S15NP12VP25 NP38ART34N45PP58 VP28V23NP35PP58 VP28V23NP38

59 Bottom Up Parse I saw a boy with a telescope 1 2 3 4 5 6 7 8
NP12  PRON12 VP2?V23NP3? NP35 ART34N45 NP35ART34N45 PP5?P56NP6? NP68ART67N7? NP68ART67N78 S1?NP12VP2? VP2?V23NP3?PP?? NP3?ART34N45PP5? NP3?ART34N45PP5? NP6?ART67N78PP8? VP25V23NP35 VP2?V23NP35PP5? PP58P56NP68 S15NP12VP25 NP38ART34N45PP58 VP28V23NP35PP58 VP28V23NP38 S18NP12VP28

60 Bottom Up Parsing - Observations
Both Noun Attachment and Verb Attachment Parses obtained by simply systematically applying the rules Numbers in subscript help in verifying the parse and getting chunks from the parse

61 Exercise For the sentence,
“The man saw the boy with a telescope” & the grammar given previously, compare the performance of top-down, bottom-up & top-down chart parsing.

62 Start of Probabilistic Parsing

63 Example of Sentence labeling: Parsing
[S1[S[S[VP[VBCome][NP[NNPJuly]]]] [,,] [CC and] [S [NP [DT the] [JJ IIT] [NN campus]] [VP [AUX is] [ADJP [JJ abuzz] [PP[IN with] [NP[ADJP [JJ new] [CC and] [ VBG returning]] [NNS students]]]]]] [..]]]

64 Noisy Channel Modeling
Source sentence Noisy Channel Target parse T*= argmax [P(T|S)] T = argmax [P(T).P(S|T)] = argmax [P(T)], since given the parse the T sentence is completely determined and P(S|T)=1

65 Corpus A collection of text called corpus, is used for collecting various language data With annotation: more information, but manual labor intensive Practice: label automatically; correct manually The famous Brown Corpus contains 1 million tagged words. Switchboard: very famous corpora 2400 conversations, 543 speakers, many US dialects, annotated with orthography and phonetics

66 Discriminative vs. Generative Model
W* = argmax (P(W|SS)) W Generative Model Discriminative Model Compute directly from P(W|SS) Compute from P(W).P(SS|W)

67 Language Models N-grams: sequence of n consecutive words/characters
Probabilistic / Stochastic Context Free Grammars: Simple probabilistic models capable of handling recursion A CFG with probabilities attached to rules Rule probabilities  how likely is it that a particular rewrite rule is used?

68 PCFGs Why PCFGs? Intuitive probabilistic models for tree-structured languages Algorithms are extensions of HMM algorithms Better than the n-gram model for language modeling.

69 Formal Definition of PCFG
A PCFG consists of A set of terminals {wk}, k = 1,….,V {wk} = { child, teddy, bear, played…} A set of non-terminals {Ni}, i = 1,…,n {Ni} = { NP, VP, DT…} A designated start symbol N1 A set of rules {Ni  j}, where j is a sequence of terminals & non-terminals NP  DT NN A corresponding set of rule probabilities

70 Rule Probabilities Rule probabilities are such that
E.g., P( NP  DT NN) = 0.2 P( NP  NN) = 0.5 P( NP  NP PP) = 0.3 P( NP  DT NN) = 0.2 Means 20 % of the training data parses use the rule NP  DT NN

71 Probabilistic Context Free Grammars
S  NP VP 1.0 NP  DT NN 0.5 NP  NNS 0.3 NP  NP PP 0.2 PP  P NP 1.0 VP  VP PP 0.6 VP  VBD NP 0.4 DT  the 1.0 NN  gunman 0.5 NN  building 0.5 VBD  sprayed 1.0 NNS  bullets 1.0

72 Example Parse t1` The gunman sprayed the building with bullets. S1.0
P (t1) = 1.0 * 0.5 * 1.0 * 0.5 * 0.6 * 0.4 * 1.0 * 0.5 * 1.0 * 0.5 * 1.0 * 1.0 * 0.3 * = NP0.5 VP0.6 NN0.5 DT1.0 VP0.4 PP1.0 P1.0 NP0.3 NP0.5 The gunman VBD1.0 DT1.0 NN0.5 with NNS1.0 sprayed the building bullets

73 Another Parse t2 The gunman sprayed the building with bullets. S1.0
P (t2) = 1.0 * 0.5 * 1.0 * 0.5 * 0.4 * 1.0 * 0.2 * 0.5 * 1.0 * 0.5 * 1.0 * 1.0 * 0.3 * = NP0.5 VP0.4 NN0.5 DT1.0 VBD1.0 NP0.2 The gunman sprayed NP0.5 PP1.0 DT1.0 NN0.5 P1.0 NP0.3 NNS1.0 the building with bullets

74 Probability of a sentence
Notation : wab – subsequence wa….wb Nj dominates wa….wb or yield(Nj) = wa….wb Nj NP wa……………..wb the..sweet..teddy ..bear Probability of a sentence = P(w1m) Where t is a parse tree of the sentence If t is a parse tree for the sentence w1m, this will be 1 !!

75 Assumptions of the PCFG model
Place invariance : P(NP  DT NN) is same in locations 1 and 2 Context-free : P(NP  DT NN | anything outside “The child”) = P(NP  DT NN) Ancestor free : At 2, P(NP  DT NN|its ancestor is VP) = P(NP DT NN) S NP VP 1 The child NP 2 The toy

76 Probability of a parse tree
Domination :We say Nj dominates from k to l, symbolized as , if Wk,l is derived from Nj P (tree |sentence) = P (tree | S1,l ) where S1,l means that the start symbol S dominates the word sequence W1,l P (t |s) approximately equals joint probability of constituent non-terminals dominating the sentence fragments (next slide)

77 Probability of a parse tree (cont.)
S1,l NP1,2 VP3,l N2 V3,3 PP4,l P4,4 NP5,l w2 w4 DT1 w1 w3 w5 wl P ( t|s ) = P (t | S1,l ) = P ( NP1,2, DT1,1 , w1, N2,2, w2, VP3,l, V3,3 , w3, PP4,l, P4,4 , w4, NP5,l, w5…l | S1,l ) = P ( NP1,2 , VP3,l | S1,l) * P ( DT1,1 , N2,2 | NP1,2) * D(w1 | DT1,1) * P (w2 | N2,2) * P (V3,3, PP4,l | VP3,l) * P(w3 | V3,3) * P( P4,4, NP5,l | PP4,l ) * P(w4|P4,4) * P (w5…l | NP5,l) (Using Chain Rule, Context Freeness and Ancestor Freeness )

78 HMM ↔ PCFG O observed sequence ↔ w1m sentence
X state sequence ↔ t parse tree  model ↔ G grammar Three fundamental questions

79 HMM ↔ PCFG How likely is a certain observation given the model? ↔ How likely is a sentence given the grammar? How to choose a state sequence which best explains the observations? ↔ How to choose a parse which best supports the sentence?

80 HMM ↔ PCFG How to choose the model parameters that best explain the observed data? ↔ How to choose rule probabilities which maximize the probabilities of the observed sentences?

81 Interesting Probabilities
What is the probability of having a NP at this position such that it will derive “the building” ? - Inside Probabilities NP The gunman sprayed the building with bullets Outside Probabilities What is the probability of starting from N1 and deriving “The gunman sprayed”, a NP and “with bullets” ? -

82 Interesting Probabilities
Random variables to be considered The non-terminal being expanded E.g., NP The word-span covered by the non-terminal. E.g., (4,5) refers to words “the building” While calculating probabilities, consider: The rule to be used for expansion : E.g., NP  DT NN The probabilities associated with the RHS non-terminals : E.g., DT subtree’s inside/outside probabilities & NN subtree’s inside/outside probabilities

83 Outside Probability j(p,q) : The probability of beginning with N1 & generating the non-terminal Njpq and all words outside wp..wq N1 Nj w1 ………wp-1 wp…wqwq+1 ……… wm

84 Inside Probabilities j(p,q) : The probability of generating the words wp..wq starting with the non-terminal Njpq. N1 Nj w1 ………wp-1 wp…wqwq+1 ……… wm

85 Outside & Inside Probabilities: example
NP The gunman sprayed the building with bullets


Download ppt "CS460/626 : Natural Language Processing/Speech, NLP and the Web (Lecture 23, 24– Parsing Algorithms; Parsing in case of Ambiguity; Probabilistic Parsing)"

Similar presentations


Ads by Google