For Friday Finish chapter 23 Homework: –Chapter 22, exercise 9.

Slides:



Advertisements
Similar presentations
GRAMMAR & PARSING (Syntactic Analysis) NLP- WEEK 4.
Advertisements

For Friday No reading Homework –Chapter 23, exercises 1, 13, 14, 19 –Not as bad as it sounds –Do them IN ORDER – do not read ahead here.
10. Lexicalized and Probabilistic Parsing -Speech and Language Processing- 발표자 : 정영임 발표일 :
March 1, 2009 Dr. Muhammed Al-Mulhem 1 ICS 482 Natural Language Processing Probabilistic Context Free Grammars (Chapter 14) Muhammed Al-Mulhem March 1,
For Monday Read Chapter 23, sections 3-4 Homework –Chapter 23, exercises 1, 6, 14, 19 –Do them in order. Do NOT read ahead.
NLP and Speech Course Review. Morphological Analyzer Lexicon Part-of-Speech (POS) Tagging Grammar Rules Parser thethe – determiner Det NP → Det.
Amirkabir University of Technology Computer Engineering Faculty AILAB Efficient Parsing Ahmad Abdollahzadeh Barfouroush Aban 1381 Natural Language Processing.
Big Ideas in Cmput366. Search Blind Search Iterative deepening Heuristic Search A* Local and Stochastic Search Randomized algorithm Constraint satisfaction.
Features and Unification
Introduction to Syntax, with Part-of-Speech Tagging Owen Rambow September 17 & 19.
Natural Language Query Interface Mostafa Karkache & Bryce Wenninger.
تمرين شماره 1 درس NLP سيلابس درس NLP در دانشگاه هاي ديگر ___________________________ راحله مکي استاد درس: دکتر عبدالله زاده پاييز 85.
Parsing SLP Chapter 13. 7/2/2015 Speech and Language Processing - Jurafsky and Martin 2 Outline  Parsing with CFGs  Bottom-up, top-down  CKY parsing.
Models of Generative Grammar Smriti Singh. Generative Grammar  A Generative Grammar is a set of formal rules that can generate an infinite set of sentences.
1 Basic Parsing with Context Free Grammars Chapter 13 September/October 2012 Lecture 6.
March 1, 2009 Dr. Muhammed Al-Mulhem 1 ICS 482 Natural Language Processing INTRODUCTION Muhammed Al-Mulhem March 1, 2009.
11 CS 388: Natural Language Processing: Syntactic Parsing Raymond J. Mooney University of Texas at Austin.
9/8/20151 Natural Language Processing Lecture Notes 1.
For Wednesday Read chapter 23, sections 1-2 Homework: –Chapter 22, exercises 1, 8, 14.
Lecture 12: 22/6/1435 Natural language processing Lecturer/ Kawther Abas 363CS – Artificial Intelligence.
For Monday Read chapter 23, sections 1-2 FOIL exercise due.
1 Statistical Parsing Chapter 14 October 2012 Lecture #9.
1 Natural Language Processing Lecture Notes 11 Chapter 15 (part 1)
Natural Language Processing Artificial Intelligence CMSC February 28, 2002.
Introduction to CL & NLP CMSC April 1, 2003.
Parsing I: Earley Parser CMSC Natural Language Processing May 1, 2003.
PARSING David Kauchak CS159 – Spring 2011 some slides adapted from Ray Mooney.
11 Chapter 14 Part 1 Statistical Parsing Based on slides by Ray Mooney.
Notes on Pinker ch.7 Grammar, parsing, meaning. What is a grammar? A grammar is a code or function that is a database specifying what kind of sounds correspond.
For Wednesday Finish Chapter 22 Program 4 due. Program 4 Any questions?
Lecture 1, 7/21/2005Natural Language Processing1 CS60057 Speech &Natural Language Processing Autumn 2007 Lecture August 2007.
For Wednesday Read chapter 23 Homework: –Chapter 22, exercises 1,4, 7, and 14.
CSA2050 Introduction to Computational Linguistics Lecture 1 Overview.
CPE 480 Natural Language Processing Lecture 4: Syntax Adapted from Owen Rambow’s slides for CSc Fall 2006.
CSA2050 Introduction to Computational Linguistics Lecture 1 What is Computational Linguistics?
For Wednesday Read chapter 22, sections 4-6 Homework: –Chapter 18, exercise 7.
Rules, Movement, Ambiguity
Artificial Intelligence: Natural Language
CSA2050 Introduction to Computational Linguistics Parsing I.
PARSING 2 David Kauchak CS159 – Spring 2011 some slides adapted from Ray Mooney.
For Monday Read chapter 24, sections 1-3 Homework: –Chapter 23, exercise 8.
For Friday Finish chapter 24 No written homework.
For Monday Read chapter 26 Last Homework –Chapter 23, exercise 7.
Chapter 22 - Communication April 8, – Semantic Interpretation Uses First Order Logic as the representation language Compositional Semantics.
For Friday No reading Program 4 due. Program 4 Any questions?
CS460/IT632 Natural Language Processing/Language Technology for the Web Lecture 1 (03/01/06) Prof. Pushpak Bhattacharyya IIT Bombay Introduction to Natural.
For Friday Finish chapter 23 Homework –Chapter 23, exercise 15.
CS 4705 Lecture 7 Parsing with Context-Free Grammars.
CPSC 422, Lecture 27Slide 1 Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 27 Nov, 16, 2015.
For Monday Read chapter 26 Homework: –Chapter 23, exercises 8 and 9.
Overview of Statistical NLP IR Group Meeting March 7, 2006.
NATURAL LANGUAGE PROCESSING
PARSING David Kauchak CS159 – Fall Admin Assignment 3 Quiz #1  High: 36  Average: 33 (92%)  Median: 33.5 (93%)
Speech and Language Processing SLP Chapter 13 Parsing.
Introduction to Parsing
Natural Language Processing Vasile Rus
Basic Parsing with Context Free Grammars Chapter 13
Syntax Analysis Chapter 4.
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 27
CS 388: Natural Language Processing: Syntactic Parsing
CSCI 5832 Natural Language Processing
CSCI 5832 Natural Language Processing
Natural Language - General
CSA2050 Introduction to Computational Linguistics
PRESENTATION: GROUP # 5 Roll No: 14,17,25,36,37 TOPIC: STATISTICAL PARSING AND HIDDEN MARKOV MODEL.
David Kauchak CS159 – Spring 2019
Artificial Intelligence 2004 Speech & Natural Language Processing
Information Retrieval
Presentation transcript:

For Friday Finish chapter 23 Homework: –Chapter 22, exercise 9

Program 5 Any questions?

Syntactic Parsing Given a string of words, determine if it is grammatical, i.e. if it can be derived from a particular grammar. The derivation itself may also be of interest. Normally want to determine all possible parse trees and then use semantics and pragmatics to eliminate spurious parses and build a semantic representation.

Parsing Complexity Problem: Many sentences have many parses. An English sentence with n prepositional phrases at the end has at least 2 n parses. I saw the man on the hill with a telescope on Tuesday in Austin... The actual number of parses is given by the Catalan numbers: 1, 2, 5, 14, 42, 132, 429, 1430, 4862,

Parsing Algorithms Top Down: Search the space of possible derivations of S (e.g.depth­first) for one that matches the input sentence. I saw the man. S ­> NP VP NP ­> Det Adj* N Det ­> the Det ­> a Det ­> an NP ­> ProN ProN ­> I VP ­> V NP V ­> hit V ­> took V ­> saw NP ­> Det Adj* N Det ­> the Adj* ­> e N ­> man

Parsing Algorithms (cont.) Bottom Up: Search upward from words finding larger and larger phrases until a sentence is found. I saw the man. ProN saw the man ProN ­> I NP saw the man NP ­> ProN NP N the man N ­> saw (dead end) NP V the man V ­> saw NP V Det man Det ­> the NP V Det Adj* man Adj* ­> e NP V Det Adj* N N ­> man NP V NP NP ­> Det Adj* N NP VP VP ­> V NP S S ­> NP VP

Bottom­up Parsing Algorithm function BOTTOM­UP­PARSE(words, grammar) returns a parse tree forest  words loop do if LENGTH(forest) = 1 and CATEGORY(forest[1]) = START(grammar) then return forest[1] else i  choose from {1...LENGTH(forest)} rule  choose from RULES(grammar) n  LENGTH(RULE­RHS(rule)) subsequence  SUBSEQUENCE(forest, i, i+n­1) if MATCH(subsequence, RULE­RHS(rule)) then forest[i...i+n­1] / [MAKE­NODE(RULE­LHS(rule), subsequence)] else fail end

Augmented Grammars Simple CFGs generally insufficient: “The dogs bites the girl.” Could deal with this by adding rules. –What’s the problem with that approach? Could also “augment” the rules: add constraints to the rules that say number and person must match.

Verb Subcategorization

Semantics Need a semantic representation Need a way to translate a sentence into that representation. Issues: –Knowledge representation still a somewhat open question –Composition “He kicked the bucket.” –Effect of syntax on semantics

Dealing with Ambiguity Types: –Lexical –Syntactic ambiguity –Modifier meanings –Figures of speech Metonymy Metaphor

Resolving Ambiguity Use what you know about the world, the current situation, and language to determine the most likely parse, using techniques for uncertain reasoning.

Discourse More text = more issues Reference resolution Ellipsis Coherence/focus

Survey of Some Natural Language Processing Research

Speech Recognition Two major approaches –Neural Networks –Hidden Markov Models A statistical technique Tries to determine the probability of a certain string of words producing a certain string of sounds Choose the most probable string of words Both approaches are “learning” approaches

Syntax Both hand-constructed approaches and data- driven or learning approaches Multiple levels of processing and goals of processing Most active area of work in NLP (maybe the easiest because we understand syntax much better than we understand semantics and pragmatics)

POS Tagging Statistical approaches--based on probability of sequences of tags and of words having particular tags Symbolic learning approaches –One of these: transformation-based learning developed by Eric Brill is perhaps the best known tagger Approaches data-driven

Developing Parsers Hand-crafted grammars Usually some variation on CFG Definite Clause Grammars (DCG) –A variation on CFGs that allow extensions like agreement checking –Built-in handling of these in most Prologs Hand-crafted grammars follow the different types of grammars popular in linguistics Since linguistics hasn’t produced a perfect grammar, we can’t code one

Efficient Parsing Top down and bottom up both have issues Also common is chart parsing –Basic idea is we’re going to locate and store info about every string that matches a grammar rule One area of research is producing more efficient parsing

Data-Driven Parsing PCFG - Probabilistic Context Free Grammars Constructed from data Parse by determining all parses (or many parses) and selecting the most probable Fairly successful, but requires a LOT of work to create the data

Applying Learning to Parsing Basic problem is the lack of negative examples Also, mapping complete string to parse seems not the right approach Look at the operations of the parse and learn rules for the operations, not for the complete parse at once

Syntax Demos ch_overviewhttp://nlp.cs.berkeley.edu/Main.html#resear ch_overview

Language Identification

Semantics Most work probably hand-constructed systems Some more interested in developing the semantics than the mappings Basic question: what constitutes a semantic representation? Answer may depend on application???

Possible Semantic Representations Logical representation Database query Case grammar

Distinguishing Word Senses Use context to determine which sense of a word is meant Probabilistic approaches Rules Issues –Obtaining sense-tagged corpora –What senses do we want to distinguish?

Semantic Demos mlhttp:// ml