Chapter 16: Features and Unification Heshaam Faili University of Tehran.

Slides:



Advertisements
Similar presentations
Feature-based Grammar Ling 571 Deep Techniques for NLP February 2, 2001.
Advertisements

 Christel Kemke 2007/08 COMP 4060 Natural Language Processing Feature Structures and Unification.
Feature Structures and Parsing Unification Grammars Algorithms for NLP 18 November 2014.
Lexical Functional Grammar : Grammar Formalisms Spring Term 2004.
1 Unification Grammars Allen ’ s Chapter 4 J&M ’ s Chapter 11.
BİL711 Natural Language Processing1 Problems with CFGs We know that CFGs cannot handle certain things which are available in natural languages. In particular,
1 Natural Language Processing Lecture 7 Unification Grammars Reading: James Allen NLU (Chapter 4)
Syntactic analysis using Context Free Grammars. Analysis of language Morphological analysis – Chairs, Part Of Speech (POS) tagging – The/DT man/NN left/VBD.
Grammars, constituency and order A grammar describes the legal strings of a language in terms of constituency and order. For example, a grammar for a fragment.
Grammatical Relations and Lexical Functional Grammar Grammar Formalisms Spring Term 2004.
Parsing with Context Free Grammars Reading: Chap 13, Jurafsky & Martin
1 Features and Augmented Grammars Allen ’ s Chapter 4 J&M ’ s Chapter 11.
Natural Language Processing - Feature Structures - Feature Structures and Unification.
Features & Unification Ling 571 Deep Processing Techniques for NLP January 26, 2011.
PCFG Parsing, Evaluation, & Improvements Ling 571 Deep Processing Techniques for NLP January 24, 2011.
Features & Unification Ling 571 Deep Processing Techniques for NLP January 31, 2011.
Parsing context-free grammars Context-free grammars specify structure, not process. There are many different ways to parse input in accordance with a given.
Parsing with CFG Ling 571 Fei Xia Week 2: 10/4-10/6/05.
Syllabus Text Books Classes Reading Material Assignments Grades Links Forum Text Books עיבוד שפות טבעיות - שיעור עשר Chart Parsing (cont) Features.
Sag et al., Chapter 4 Complex Feature Values 10/7/04 Michael Mulyar.
Issues in Computational Linguistics: Parsing and Generation Dick Crouch and Tracy King.
Earley’s algorithm Earley’s algorithm employs the dynamic programming technique to address the weaknesses of general top-down parsing. Dynamic programming.
 Christel Kemke 2007/08 COMP 4060 Natural Language Processing Feature Structures and Unification.
Features and Unification
Syntax and Context-Free Grammars CMSC 723: Computational Linguistics I ― Session #6 Jimmy Lin The iSchool University of Maryland Wednesday, October 7,
1 CONTEXT-FREE GRAMMARS. NLE 2 Syntactic analysis (Parsing) S NPVP ATNNSVBD NP AT NNthechildrenate thecake.
Ch.11 Features and Unification Ling 538, 2006 Fall Jiyoung Kim.
Unification-based algorithms We have seen that treating features, such as person-number agreement, as part of the category of a phrase results in a large.
CS 4705 Lecture 11 Feature Structures and Unification Parsing.
Stochastic POS tagging Stochastic taggers choose tags that result in the highest probability: P(word | tag) * P(tag | previous n tags) Stochastic taggers.
Features and Unification Read J & M Chapter 11.. Solving the Agreement Problem Number agreement: S  NP VP * Mary walk. [NP NUMBER] [VP NUMBER] NP  det.
Context-Free Parsing Part 2 Features and Unification.
1 Basic Parsing with Context Free Grammars Chapter 13 September/October 2012 Lecture 6.
11 CS 388: Natural Language Processing: Syntactic Parsing Raymond J. Mooney University of Texas at Austin.
Context Free Grammars Reading: Chap 12-13, Jurafsky & Martin This slide set was adapted from J. Martin, U. Colorado Instructor: Paul Tarau, based on Rada.
Feature structures and unification Attributes and values.
1 Features and Unification Chapter 15 October 2012 Lecture #10.
1 CPE 480 Natural Language Processing Lecture 5: Parser Asst. Prof. Nuttanart Facundes, Ph.D.
CS 4705 Parsing More Efficiently and Accurately. Review Top-Down vs. Bottom-Up Parsers Left-corner table provides more efficient look- ahead Left recursion.
What it’s ? “parsing” Parsing or syntactic analysis is the process of analysing a string of symbols, either in natural language or in computer languages,
10. Parsing with Context-free Grammars -Speech and Language Processing- 발표자 : 정영임 발표일 :
Parsing with Context Free Grammars CSC 9010 Natural Language Processing Paula Matuszek and Mary-Angela Papalaskari This slide set was adapted from: Jim.
Parsing I: Earley Parser CMSC Natural Language Processing May 1, 2003.
7. Parsing in functional unification grammar Han gi-deuc.
Context Free Grammars Reading: Chap 9, Jurafsky & Martin This slide set was adapted from J. Martin, U. Colorado Instructor: Rada Mihalcea.
Page 1 Probabilistic Parsing and Treebanks L545 Spring 2000.
6/2/2016CPSC503 Winter CPSC 503 Computational Linguistics Lecture 9 Giuseppe Carenini.
Chapter 12: Context-Free Grammars Heshaam Faili University of Tehran.
Lecture 1, 7/21/2005Natural Language Processing1 CS60057 Speech &Natural Language Processing Autumn 2007 Lecture August 2007.
2007CLINT-LIN-FEATSTR1 Computational Linguistics for Linguists Feature Structures.
Head-driven Phrase Structure Grammar (HPSG)
Rules, Movement, Ambiguity
CSA2050 Introduction to Computational Linguistics Parsing I.
Chart Parsing and Augmenting Grammars CSE-391: Artificial Intelligence University of Pennsylvania Matt Huenerfauth March 2005.
Section 11.3 Features structures in the Grammar ─ Jin Wang.
CS 4705 Lecture 10 The Earley Algorithm. Review Top-Down vs. Bottom-Up Parsers –Both generate too many useless trees –Combine the two to avoid over-generation:
Artificial Intelligence 2004
Instructor: Nick Cercone CSEB - 1 Parsing and Context Free Grammars Parsers, Top Down, Bottom Up, Left Corner, Earley.
Chapter 11: Parsing with Unification Grammars Heshaam Faili University of Tehran.
Natural Language Processing Vasile Rus
Natural Language Processing Vasile Rus
Lecture – VIII Monojit Choudhury RS, CSE, IIT Kharagpur
Basic Parsing with Context Free Grammars Chapter 13
Dependency Parsing & Feature-based Parsing
Instructor: Nick Cercone CSEB -
Earley’s Algorithm (1970) Nice combo of our parsing ideas so far:
Parsing and More Parsing
CSCI 5832 Natural Language Processing
CPSC 503 Computational Linguistics
Parsing I: CFGs & the Earley Parser
Presentation transcript:

Chapter 16: Features and Unification Heshaam Faili University of Tehran

2 Overview Feature Structures and Unification Unification-Based Grammars Chart Parsing with Unification-Based Grammars Type Hierarchies

3 Agenda we introduce the idea that grammatical categories like VPto, Sthat, Non3sgAux, or 3sgNP, as well as the grammatical rules like S → NP VP that make use of them, should be thought of as objects that can have complex sets of properties associated with them The information in these properties is represented by constraints, and so these kinds of models are often called constraint based formalisms Use these constraints for Agreement (this flights*) Subcategorization () Adding properties (features) to words Adding some operation to rules in order to test equalities

4 Feature structures We had a problem adding agreement to CFGs. What we needed were features, e.g., a way to say: [number sg person 3 ] A structure like this allows us to state properties, e.g., about a noun phrase [cat NP number sg person 3 ] Each feature (e.g., ‘number’) is paired with a value (e.g., ‘sg’) A bundle of feature-value pairs can be put into an attribute- value matrix (AVM)

5 examples

6 Feature paths Values can be atomic (e.g. ‘sg’ or ‘NP’ or ‘3’), or can be complex, and thus we can define feature paths [cat NP agreement [number sg person 3]] The value of the path [agreement number] is ‘sg’ A grammar with only atomic feature values can be converted to a CFG e.g. AVM on previous page  NP 3,sg However, when the values are complex, it is more expressive than a CFG  can represent more linguistic phenomena

7 An Example for FS

8 Reentrancy (structure-sharing) Feature structures embedded in feature structures can share the same values That is, two features have the exact same value— they share precisely the same object as their value we’ll indicate this with a tag like *1 In this example, the agreement features of both the matrix sentence and the embedded subject are identical This is referred to as reentrancy

9 FS with shared value (Reentrant)

10 Feature structures as graphs Technically, feature structures are directed acyclic graphs (DAGs) So, the feature structure represented by the attribute-value matrix (AVM): [cat NP agreement [number sg person 3]] is really the graph: CAT AGR    NUM PER   sg 3 np

11 Unification Unification (U) = a basic operation to merge two feature structures into a resultant feature structure (FS) The two feature structures must be compatible, i.e., have no values that conflict Identical FSs: [number sg] U [number sg] = [number sg] Conflicting FSs: [number sg] U [number pl] = Fail Merging with an unspecified FS: [number sg] U [number []] = [number sg]

12 Unification (cont.) Merging FSs with different features specified: [number sg] U [person 3] = [number sg person 3 ] More examples: [cat NP] U [agreement [number sg]] = [cat NP agreement [number sg]] [agr [num sg] subj [agr [num sg]]] U [subj [agr [num sg]]]= [agr [num sg] subj [agr [num sg]]]

13 Unification with Reentrancies Remember that structure-sharing means they are the same object:

14 Unification with Reentrancies When unification takes place, shared values are copied over:

15 example

16 Subsumption We can see that a more general feature structure (less values specified) subsumes a more specific feature structure (1) [num sg] (2) [per 3] (3) [num sg per 3] So, we have the following subsumption relations, where (1) subsumes (3) (2) subsumes (3) (1) does not subsume (2), and (2) does not subsume (1)

17 Subsumption F ⊑ G if and only if 1. For every feature x in F, F(x) ⊑ G(x) (where F(x) means “the value of the feature x of feature structure F”). 2. For all paths p and q in F such that F(p) = F(q), it is also the case that G(p) = G(q).

18 Subsumption

19 Subsumption (partial order)

20 Semilattice, unification  F [] ⊑ F, so we can model the sumsumption as a lattice, which [] at the top of it Unification F ⊔ G = most general H such that F ⊑ H and G ⊑ H

21 Overview Feature Structures and Unification Unification-Based Grammars Chart Parsing with Unification-Based Grammars Type Hierarchies

22 Grammars with Feature Structures CFG skeleton augmented with feature structure path equations, i.e., each category has a feature structure CFG skeleton S  NP VP Path equations = 1. There can be zero or more path equations for each rule skeleton  no longer atomic 2. When a path equation references constituents, they can only be constituents from the CFG rule e.g., = is an illegal equation for the above rule! (But it would be fine for NP  Det Nom)

23 FEATURES STRUCTURES IN THE GRAMMAR

24 agreement subject-verb agreement and determiner nominal agreement.

25 Agreement in Feature-Based Grammars S  NP VP = VP  V NP = NP  Det Nom = Nom  Noun = Noun  flights = pl Compare with the CFG case: S  3sgNP 3sgVP S  PluralNP PluralVP 3sgVP  3sgVerb 3sgVP  3sgVerb NP 3sgVP  3sgVerb NP PP 3sgVP  3sgVerb PP etc.

26 Percolating Agreement Features S  NP VP = VP  V NP = NP  Det Nom = Nom  Noun =

27 agreement

28 Head features in the grammar An important concept shown in the previous rules is that heads of grammar rules share properties with their mothers VP  V NP = Knowing the head will tell you about the whole phrase This is important for many parsing techniques

29 Sub-categorization We could specify subcategorization like so: VP  V = intrans VP  V NP = trans VP  V NP NP = ditrans But values like ‘intrans’ do not correspond to anything that the rules actually look like To make SUBCAT better match the rules, we can make it a list of a verb’s arguments, e.g.

30 Handling Subcategorization VP  V NP PP = V  leaves = sg = There is also a longer, more formal way to specify lists: is equivalent to: [FIRST NP REST [FIRST= PP REST = <>]] VP V PP NP [head: *1[subcat ]] [cat *2][cat *3] leaves [head: *1[agr [num sg] subcat < [cat np], [cat pp] >

31 Subcategorization

32 Subcategorization frames Subcategorization, or valency, or dependency is a very important notion in capturing syntactic regularity And there is a wide variety of arguments that a verb (or noun or adjective) can take. Some subcategorization frames for ask: He asked [ Q “What was it like?”] He asked [ S[wh] what it was like] He asked [ NP her] [ S[wh] what it was like] He asked [ VP[to] to see you] He asked [ NP her] [ VP[to] to tell you] He asked [ NP a question] He asked [ NP her] [ NP a question]

33 Subcategorization frame

34 Long Distance Dependencies What cities does Continental service? What flights do you have from Boston to Baltimore? What time does that flight leave Atlanta? wh-non-subject-question: S → Wh-NP Aux NP VP Should WH-NP agreed with NP Gap-list : implemented by feature “GAP” are passed through different trees Fillers: (what cities) are put on the top of gap list and should be unified to the subcategorization frame of the verb

35 Long-Distance Dependencies TOP (fill gap): S  WH-word Be-copula NP = MIDDLE (pass gap): NP  D Nom = Nom  Nom RelClause = RelClause  RelPro NP VP = BOTTOM (identify gap): VP  V = What is the earliest flight that you have _? S Wh-wdBe-copulaNP DNom RelClause RelProNPVP V [gap *1] [head *1] [gap *1] [subcat ]

36 Overview Feature Structures and Unification Unification-Based Grammars Chart Parsing with Unification-Based Grammars Type Hierarchies

37 Implementing Unification How do we implement a check on unification? i.e., given feature structures F1 and F2, return F, the unification of F1 and F2 Unification is a recursive operation: If a feature has an atomic value, see if the other FS has that feature with the same value [F a] unifies with [], [F ], and [F a] If a feature has a complex value, follow the paths to see if they’re compatible and have the same values at bottom Does [F G1] unify with [F G2]? We have to inspect G1 and G2 to find out. To avoid cycles, we have to do an occur check to see if we’ve seen a FS before or not

38

39

40 Base case One or both of the arguments has a null value. The arguments are identical. The arguments are non-complex and non-identical.

41 An Example

42

43 original arguments are neither identical, nor null, nor atomic, These arguments are also non-identical, non-null, and non-atomic so the loop is entered again leading to a recursive check of the values of the AGREEMENT features

44 f1 and f2 after the recursion adds the value of the new PERSON feature

45 The final structures of f1 and f2 at the end

46 Modifying a Early Parser to handle Unification Our grammar still has a context-free backbone, so we could just parse a sentence with a CFG and use the features to filter out the ungrammatical sentences But by utilizing unification as we parse, we can eliminate parses that won’t work in the end e.g., we’ll eliminate NPs that don’t match in agreement features with their VPs as we parse, instead of ruling them out later

47 Changes to the Chart Representation Each state will be extended to include the LHS DAG (which can get augmented as it goes along). i.e., Add a feature structure (in DAG form) to each state So, S   NP VP, [0,0] Becomes S   NP VP, [0,0], DagS The predictor, scanner, and completer have to pass in the DAG, so all three operations have to be altered

48 Earley Parser with Unification

49 Predictor, Scanner, Completer

50 Unify States

51 Example (That flight )

52 Change to ENQUEUE The enqueue procedure should also be changed to use a subsumption test: Do not add a state to the chart if an equivalent or more general state is already there. So, if Enqueue wants to add a singular determiner state at [x, y], and the chart already has a determiner state at [x, y] unspecified for number, then Enqueue will not add it.

53 Why a Subsumption Test? If we don't impose a subsumption restriction, enqueue will add two states at [x, y], one expecting to see a singular determiner, the other just a determiner. On seeing a singular determiner, the parser will advance the dot on both rules, creating two edges (since singular will unify with both singular and with unspecified). As a result, we would get duplicate edges. If we impose the restriction, and we see either a single or plural determiner, and we advance the dot, only one edge (singular or plural) gets created at [x, y].

54 The Need for Copying show me morning flights

55 Overview Feature Structures and Unification Unification-Based Grammars Chart Parsing with Unification-Based Grammars Type Hierarchies

56 Using Type Hierarchies Instead of simple feature structures, formalisms like Head-Driven Phrase Structure Grammar (HPSG) use typed feature structures Two problems right now: What prevents us right now from specifying the following? How can we capture the fact that all values of NUMBER are the same sort of thing, i.e., make a generalization? Solution: use types

57 Type Systems 1. Each feature structure is labeled by a type. [noun CASE case] 2. Each type has appropriateness conditions specifying what features are appropriate for it. noun  [CASE case] verb  [VFORM vform] 3. Types are organized into a type hierarchy. 4. Unification is modified to allow two different types to unify.

58 Simple Type Hierarchy

59 Type Hierarchy So, if CASE is appropriate for noun, and the value of CASE is case, and we have the following type hierarchy case nom acc dat Then, the following are possible feature structures: [noun[noun[noun CASE nom] CASE acc][CASE dat]

60 Unification of types Now, when we unify feature structures, we have to unify types, too: [CASE case] U [CASE nom] = [CASE nom] [CASE nom] U [CASE acc] = fail Let’s also assume that acc and dat have a common subtype, obj acc dat obj Then, we have the following unification: [CASE acc] U [CASE dat] = [CASE obj]

61 Practices 16.2, 16.3, 16.6, 16.7,