Natural Language Processing Artificial Intelligence CMSC 25000 February 28, 2002.

Slides:



Advertisements
Similar presentations
 Christel Kemke 2007/08 COMP 4060 Natural Language Processing Feature Structures and Unification.
Advertisements

Introduction and Jurafsky Model Resource: A Probabilistic Model of Lexical and Syntactic Access and Disambiguation, Jurafsky 1996.
March 1, 2009 Dr. Muhammed Al-Mulhem 1 ICS 482 Natural Language Processing Probabilistic Context Free Grammars (Chapter 14) Muhammed Al-Mulhem March 1,
LING NLP 1 Introduction to Computational Linguistics Martha Palmer April 19, 2006.
For Monday Read Chapter 23, sections 3-4 Homework –Chapter 23, exercises 1, 6, 14, 19 –Do them in order. Do NOT read ahead.
Natural Language Processing - Feature Structures - Feature Structures and Unification.
Natural Language and Speech Processing Creation of computational models of the understanding and the generation of natural language. Different fields coming.
NLP and Speech Course Review. Morphological Analyzer Lexicon Part-of-Speech (POS) Tagging Grammar Rules Parser thethe – determiner Det NP → Det.
Cognitive Processes PSY 334 Chapter 11 – Language Structure.
Introduction to Deep Processing Techniques for NLP Deep Processing Techniques for Natural Language Processing Ling 571 January 3, 2011 Gina-Anne Levow.
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 27
6/9/2015CPSC503 Winter CPSC 503 Computational Linguistics Lecture 11 Giuseppe Carenini.
Introduction to Computational Linguistics Lecture 2.
1 Chapter 20 Understanding Language. 2 Chapter 20 Contents (1) l Natural Language Processing l Morphologic Analysis l BNF l Rewrite Rules l Regular Languages.
 Christel Kemke 2007/08 COMP 4060 Natural Language Processing Feature Structures and Unification.
تمرين شماره 1 درس NLP سيلابس درس NLP در دانشگاه هاي ديگر ___________________________ راحله مکي استاد درس: دکتر عبدالله زاده پاييز 85.
Machine Learning in Natural Language Processing Noriko Tomuro November 16, 2006.
Artificial Intelligence 2004 Natural Language Processing - Syntax and Parsing - Language Syntax Parsing.
Context-Free Grammar CSCI-GA.2590 – Lecture 3 Ralph Grishman NYU.
March 1, 2009 Dr. Muhammed Al-Mulhem 1 ICS 482 Natural Language Processing INTRODUCTION Muhammed Al-Mulhem March 1, 2009.
11 CS 388: Natural Language Processing: Syntactic Parsing Raymond J. Mooney University of Texas at Austin.
Lecture 1, 7/21/2005Natural Language Processing1 CS60057 Speech &Natural Language Processing Autumn 2005 Lecture 1 21 July 2005.
Context Free Grammars Reading: Chap 12-13, Jurafsky & Martin This slide set was adapted from J. Martin, U. Colorado Instructor: Paul Tarau, based on Rada.
PARSING David Kauchak CS457 – Fall 2011 some slides adapted from Ray Mooney.
9/8/20151 Natural Language Processing Lecture Notes 1.
CCSB354 ARTIFICIAL INTELLIGENCE (AI)
Lecture 12: 22/6/1435 Natural language processing Lecturer/ Kawther Abas 363CS – Artificial Intelligence.
For Monday Read chapter 23, sections 1-2 FOIL exercise due.
For Friday Finish chapter 23 Homework: –Chapter 22, exercise 9.
10/12/2015CPSC503 Winter CPSC 503 Computational Linguistics Lecture 10 Giuseppe Carenini.
Finite State Parsing & Information Extraction CMSC Intro to NLP January 10, 2006.
Chapter 15 Natural Language Processing (cont)
Syntax: 10/18/2015IT 3271 Semantics: Describe the structures of programs Describe the meaning of programs Programming Languages (formal languages) -- How.
Introduction to CL & NLP CMSC April 1, 2003.
Parsing I: Earley Parser CMSC Natural Language Processing May 1, 2003.
PARSING David Kauchak CS159 – Spring 2011 some slides adapted from Ray Mooney.
11 Chapter 14 Part 1 Statistical Parsing Based on slides by Ray Mooney.
1 CSI 5180: Topics in AI: Natural Language Processing, A Statistical Approach Instructor: Nathalie Japkowicz Objectives of.
For Wednesday Read chapter 23 Homework: –Chapter 22, exercises 1,4, 7, and 14.
October 2005CSA3180 NLP1 CSA3180 Natural Language Processing Introduction and Course Overview.
Linguistic Essentials
Artificial Intelligence: Natural Language
CSA2050 Introduction to Computational Linguistics Parsing I.
Basic Parsing Algorithms: Earley Parser and Left Corner Parsing
PARSING 2 David Kauchak CS159 – Spring 2011 some slides adapted from Ray Mooney.
1 Context Free Grammars October Syntactic Grammaticality Doesn’t depend on Having heard the sentence before The sentence being true –I saw a unicorn.
Quick Speech Synthesis CMSC Natural Language Processing April 29, 2003.
For Friday No reading Program 4 due. Program 4 Any questions?
LING 001 Introduction to Linguistics Spring 2010 Syntactic parsing Part-Of-Speech tagging Apr. 5 Computational linguistics.
CS460/IT632 Natural Language Processing/Language Technology for the Web Lecture 1 (03/01/06) Prof. Pushpak Bhattacharyya IIT Bombay Introduction to Natural.
Supertagging CMSC Natural Language Processing January 31, 2006.
CPSC 422, Lecture 27Slide 1 Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 27 Nov, 16, 2015.
Natural Language Processing Slides adapted from Pedro Domingos
English Syntax Read J & M Chapter 9.. Two Kinds of Issues Linguistic – what are the facts about language? The rules of syntax (grammar) Algorithmic –
Overview of Statistical NLP IR Group Meeting March 7, 2006.
NATURAL LANGUAGE PROCESSING
PARSING David Kauchak CS159 – Fall Admin Assignment 3 Quiz #1  High: 36  Average: 33 (92%)  Median: 33.5 (93%)
Introduction to Deep Processing Techniques for NLP Deep Processing Techniques for NLP Ling 571 January 4, 2016 Gina-Anne Levow.
Roadmap Probabilistic CFGs –Handling ambiguity – more likely analyses –Adding probabilities Grammar Parsing: probabilistic CYK Learning probabilities:
SYNTAX.
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 27
Machine Learning in Natural Language Processing
Natural Language - General
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 27
Parsing and More Parsing
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 26
Parsing I: CFGs & the Earley Parser
David Kauchak CS159 – Spring 2019
Artificial Intelligence 2004 Speech & Natural Language Processing
Presentation transcript:

Natural Language Processing Artificial Intelligence CMSC February 28, 2002

Agenda Why NLP? –Goals & Applications Challenges: Knowledge & Ambiguity –Key types of knowledge Morphology, Syntax, Semantics, Pragmatics, Discourse –Handling Ambiguity Syntactic Ambiguity: Probabilistic Parsing Semantic Ambiguity: Word Sense Disambiguation Conclusions

Why Language? Natural Language in Artificial Intelligence –Language use as distinctive feature of human intelligence –Infinite utterances: Diverse languages with fundamental similarities “Computational linguistics” –Communicative acts Inform, request,...

Why Language? Applications Machine Translation Question-Answering –Database queries to web search Spoken language systems Intelligent tutoring

Knowledge of Language What does it mean to know a language? –Know the words (lexicon) Pronunciation, Formation, Conjugation –Know how the words form sentences Sentence structure, Compositional meaning –Know how to interpret the sentence Statement, question,.. –Know how to group sentences Narrative coherence, dialogue

Word-level Knowledge Lexicon: –List of legal words in a language –Part of speech: noun, verb, adjective, determiner Example: –Noun -> cat | dog | mouse | ball | rock –Verb -> chase | bite | fetch | bat –Adjective -> black | brown | furry | striped | heavy –Determiner -> the | that | a | an

Word-level Knowledge: Issues Issue 1: Lexicon Size –Potentially HUGE! –Controlling factor: morphology Store base forms (roots/stems) –Use morphologic process to generate / analyze –E.g. Dog: dog(s); sing: sings, sang, sung, singing, singer,.. Issue 2: Lexical ambiguity –rock: N/V; dog: N/V; –“Time flies like a banana”

Sentence-level Knowledge: Syntax Language models –More than just words: “banana a flies time like” –Formal vs natural: Grammar defines language Chomsky Hierarchy Recursively Enumerable =Any Context = AB->BA Sensitive Context A-> aBc Free Regular S->aS Expression a*b*

Syntactic Analysis: Grammars Natural vs Formal languages –Natural languages have degrees of acceptability ‘It ain’t hard’; ‘You gave what to whom?’ Grammar combines words into phrases –S-> NP VP –NP -> {Det} {Adj} N –VP -> V | V NP | V NP PP

Syntactic Analysis: Parsing Recover phrase structure from sentence –Based on grammar S NP VP Det Adj N V NP Det Adj N The black cat chased the furry mouse

Syntactic Analysis: Parsing Issue 1: Complexity Solution 1: Chart parser - dynamic programming –O( ) Issue 2: Structural ambiguity –‘I saw the man on the hill with the telescope’ Is the telescope on the hill?’ Solution 2 (partial): Probabilistic parsing

Semantic Analysis Grammatical = Meaningful –“Colorless green ideas sleep furiously” Compositional Semantics –Meaning of a sentence is meaning of subparts –Associate semantic interpretation with syntactic –E.g. Nouns are variables (themselves): cat,mouse Adjectives: unary predicates: Black(cat), Furry(mouse) Verbs: multi-place: VP: x chased(x,Furry(mouse)) Sentence ( x chased(x, Furry(mouse))Black(cat) –chased(Black(cat),Furry(mouse))

Semantic Ambiguity Examples: –I went to the bank- of the river to deposit some money –He banked at First Union the plane Interpretation depends on –Sentence (or larger) topic context –Syntactic structure

Pragmatics & Discourse Interpretation in context –Act accomplished by utterance “Do you have the time?”, “Can you pass the salt?” Requests with non-literal meaning –Also, includes politeness, performatives, etc Interpretation of multiple utterances –“The cat chased the mouse. It got away.” –Resolve referring expressions

Natural Language Understanding Input Tokenization/ Morphology Parsing Semantic Analysis Pragmatics/ Discourse Meaning Key issues: –Knowledge How acquire this knowledge of language? –Hand-coded? Automatically acquired? –Ambiguity How determine appropriate interpretation? –Pervasive, preference-based

Handling Syntactic Ambiguity Natural language syntax Varied, has DEGREES of acceptability Ambiguous Probability: framework for preferences –Augment original context-free rules: PCFG –Add probabilities to transitions NP -> N NP -> Det N NP -> Det Adj N NP -> NP PP VP -> V VP -> V NP VP -> V NP PP S -> NP VP S -> S conj S PP -> P NP 1.0

PCFGs Learning probabilities –Strategy 1: Write (manual) CFG, Use treebank (collection of parse trees) to find probabilities –Strategy 2: Use larger treebank (+ linguistic constraint) Learn rules & probabilities (inside-outside algorithm) Parsing with PCFGs –Rank parse trees based on probability –Provides graceful degradation Can get some parse even for unusual constructions - low value

Parse Ambiguity Two parse trees S NP VP N V NP PP Det N P NP Det N I saw the man with the telescope S NP VP N V NP NP PP Det N P NP Det N I saw the man with the telescope

Parse Probabilities –T(ree),S(entence),n(ode),R(ule) –T1 = 0.85*0.2*0.1*0.65*1*0.65 = –T2 = 0.85*0.2*0.45*0.05*0.65*1*0.65 = Select T1 Best systems achieve 92-93% accuracy

Semantic Ambiguity “Plant” ambiguity –Botanical vs Manufacturing senses Two types of context –Local: 1-2 words away –Global: several sentence window Two observations (Yarowsky 1995) –One sense per collocation (local) –One sense per discourse (global)

Learn Disambiguators Initialize small set of “seed” cases Collect local context information –“collocations” E.g. 2 words away from “production”, 1 word from “seed” Contexts = rules Make decision list= rules ranked by mutual info Iterate: Labeling via DL, collecting contexts Label all entries in discourse with majority sense –Repeat

Disambiguate For each new unlabeled case, –Use decision list to label > 95% accurate on set of highly ambiguous –Also used for accent restoration in

Natural Language Processing Goals: Understand and imitate distinctive human capacity Myriad applications: MT, Q&A, SLS Key Issues: –Capturing knowledge of language Automatic acquisition current focus: linguistics+ML –Resolving ambiguity, managing preference Apply (probabilistic) knowledge Effective in constrained environment