1 *Textbooks you need Manning, C. D., Sch ü tze, H.: Foundations of Statistical Natural Language Processing. The MIT Press. 1999. ISBN 0-262-13360-1. [required]

Slides:



Advertisements
Similar presentations
Adverbs and Adjectives
Advertisements

Coordination and Parallel Structure
INTERNATIONAL CONFERENCE ON NATURAL LANGUAGE PROCESSING NLP-AI IIIT-Hyderabad CIIL, Mysore ICON DECEMBER, 2003.
Introduction to NLP What is Natural Language Processing?
For Monday Read Chapter 23, sections 3-4 Homework –Chapter 23, exercises 1, 6, 14, 19 –Do them in order. Do NOT read ahead.
Natural Language and Speech Processing Creation of computational models of the understanding and the generation of natural language. Different fields coming.
Shallow Processing: Summary Shallow Processing Techniques for NLP Ling570 December 7, 2011.
1 Natural Language Processing for the Web Prof. Kathleen McKeown 722 CEPSR, Office Hours: Wed, 1-2; Mon 3-4 TA: Fadi Biadsy 702 CEPSR,
1 Wen-Hsiang Lu ( 盧文祥 ) Department of Computer Science and Information Engineering National Cheng Kung University 2014/02/17 Multilingual and Crosslingual.
Introduction to CL Session 1: 7/08/2011. What is computational linguistics? Processing natural language text by computers  for practical applications.
تمرين شماره 1 درس NLP سيلابس درس NLP در دانشگاه هاي ديگر ___________________________ راحله مکي استاد درس: دکتر عبدالله زاده پاييز 85.
Machine Learning in Natural Language Processing Noriko Tomuro November 16, 2006.
1/23 Applications of NLP. 2/23 Applications Text-to-speech, speech-to-text Dialogues sytems / conversation machines NL interfaces to –QA systems –IR systems.
Natural Language Processing Prof: Jason Eisner Webpage: syllabus, announcements, slides, homeworks.
To the Community Language Program Intermediate Level 3! Instructor: Tina Song.
Natural Language Processing Ellen Back, LIS489, Spring 2015.
SI485i : NLP Set 12 Features and Prediction. What is NLP, really? Many of our tasks boil down to finding intelligent features of language. We do lots.
Reported Speech Roll No Presented By:- Class: Ixth “A”
SI485i : NLP Day 1 Intro to NLP. Assumptions about You You know… how to program Java basic UNIX usage basic probability and statistics (we’ll also review)
Lecture 1, 7/21/2005Natural Language Processing1 CS60057 Speech &Natural Language Processing Autumn 2005 Lecture 1 21 July 2005.
9/8/20151 Natural Language Processing Lecture Notes 1.
1 Ling 569: Introduction to Computational Linguistics Jason Eisner Johns Hopkins University Tu/Th 1:30-3:20 (also this Fri 1-5)
Lemmatization Tagging LELA /20 Lemmatization Basic form of annotation involving identification of underlying lemmas (lexemes) of the words in.
Lecture 12: 22/6/1435 Natural language processing Lecturer/ Kawther Abas 363CS – Artificial Intelligence.
1 Natural Language Processing Gholamreza Ghassem-Sani.
For Friday Finish chapter 23 Homework: –Chapter 22, exercise 9.
NLP.
1 Natural Language Processing Gholamreza Ghassem-Sani Fall 1383.
To-infinitive GERUND.
1 Computational Linguistics Ling 200 Spring 2006.
Natural Language Processing Introduction. 2 Natural Language Processing We’re going to study what goes into getting computers to perform useful and interesting.
Introduction to NLP ch1 What is Natural Language Processing?
9/13/1999 JHU CS /Jan Hajic 1 Introduction to Natural Language Processing AI-Lab
Unit 4 What would you do? Self Check Language Goal: 1. learn how to do before some accidents and problems. 2. grasp the following phrase: let... down;
Natural Language Processing Artificial Intelligence CMSC February 28, 2002.
Introduction to CL & NLP CMSC April 1, 2003.
S1: Chapter 1 Mathematical Models Dr J Frost Last modified: 6 th September 2015.
PARSING David Kauchak CS159 – Spring 2011 some slides adapted from Ray Mooney.
11 Chapter 14 Part 1 Statistical Parsing Based on slides by Ray Mooney.
1 CSI 5180: Topics in AI: Natural Language Processing, A Statistical Approach Instructor: Nathalie Japkowicz Objectives of.
CS 124/LINGUIST 180 From Languages to Information
October 2005CSA3180 NLP1 CSA3180 Natural Language Processing Introduction and Course Overview.
NACLO 2008 North American Computation Linguistics Olympiad Brandeis CL Olympiad Team James Pustejovsky Tai Sassen-Liang Sharone Horowit-Hendler Noam Sienna.
ICS 482: Natural language Processing Pre-introduction
For Friday Finish chapter 24 No written homework.
For Monday Read chapter 26 Last Homework –Chapter 23, exercise 7.
ACE TESOL Diploma Program – London Language Institute OBJECTIVES You will understand: 1. The terminology and concepts of semantics, pragmatics and discourse.
Auckland 2012Kilgarriff: NLP and Corpus Processing1 The contribution of NLP: corpus processing.
CSE467/567 Computational Linguistics Carl Alphonce Computer Science & Engineering University at Buffalo.
9/13/1999JHU CS / Intro to NLP/Jan Hajic1 Textbooks you need Manning, C. D., Sch ü tze, H.: Foundations of Statistical Natural Language Processing.
Natural Language Processing Chapter 1 : Introduction.
CS 188: Artificial Intelligence Spring 2009 Natural Language Processing Dan Klein – UC Berkeley 1.
CS460/IT632 Natural Language Processing/Language Technology for the Web Lecture 1 (03/01/06) Prof. Pushpak Bhattacharyya IIT Bombay Introduction to Natural.
For Friday Finish chapter 23 Homework –Chapter 23, exercise 15.
CPSC 422, Lecture 27Slide 1 Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 27 Nov, 16, 2015.
SYNTAX.
1 An Introduction to Computational Linguistics Mohammad Bahrani.
For Monday Read chapter 26 Homework: –Chapter 23, exercises 8 and 9.
Overview of Statistical NLP IR Group Meeting March 7, 2006.
Lets eat the same porridge with a new taste. I have threads with me but, do not know how to knit the fabric !!!
/665 Natural Language Processing
Natural Language Processing [05 hours/week, 09 Credits] [Theory]
CSC 594 Topics in AI – Natural Language Processing
Statistical NLP: Lecture 3
Introduction to NLP What is Natural Language Processing?
Machine Learning in Natural Language Processing
CS246: Information Retrieval
From Language to Information
Presentation transcript:

1 *Textbooks you need Manning, C. D., Sch ü tze, H.: Foundations of Statistical Natural Language Processing. The MIT Press ISBN [required] Allen, J.: Natural Language Understanding. The Benjamins/Cummins Publishing Co nd edition Jurafsky, D. and J. H. Martin: Speech and Language Processing. Prentice-Hall nd editionSpeech and Language Processing

2 Other reading Charniak, E: –Statistical Language Learning. The MIT Press ISBN Cover, T. M., Thomas, J. A.: –Elements of Information Theory. Wiley ISBN Jelinek, F.: –Statistical Methods for Speech Recognition. The MIT Press ISBN Proceedings of major conferences: –ACL (Assoc. of Computational Linguistics) –NAACL HLT (North American Chapter of ACL) –COLING (Intl. Committee of Computational Linguistics) –ACM SIGIR –Interspeech/ASRU/SLT

3 Course segments Intro & Probability & Information Theory –The very basics: definitions, formulas, examples. Language Modeling –n-gram models, parameter estimation –smoothing (EM algorithm) A Bit of Linguistics –phonology, morphology, syntax, semantics, discourse Words and the Lexicon –word classes, mutual information, bit of lexicography.

4 Course segments (cont.) Hidden Markov Models –background, algorithms, parameter estimation Tagging: Methods, Algorithms, Evaluation –tagsets, morphology, lemmatization –HMM tagging, Transformation-based, Feature-based NL Grammars and Parsing: Data, Algorithms –Grammars and Automata, Deterministic Parsing –Statistical parsing. Algorithms, parameterization, evaluation Applications (MT, ASR, IR, Q&A,...)

5 Goals of the HLT Computers would be a lot more useful if they could handle our , do our library research, talk to us … But they are fazed by natural human language. How can we make computers have abilities to handle human language? (Or help them learn it as kids do?)

6 A few applications of HLT Spelling correction, grammar checking …(language learning and evaluation e.g. TOEFL essay score) Better search engines Information extraction, gisting Psychotherapy; Harlequin romances; etc. New interfaces: –Speech recognition (and text-to-speech) –Dialogue systems (USS Enterprise onboard computer) –Machine translation; speech translation (the Babel tower??) Trans-lingual summarization, detection, extraction …

Dan Jurafsky Question Answering: IBM’s Watson Won Jeopardy on February 16, 2011! 7 WILLIAM WILKINSON’S “AN ACCOUNT OF THE PRINCIPALITIES OF WALLACHIA AND MOLDOVIA” INSPIRED THIS AUTHOR’S MOST FAMOUS NOVEL Bram Stoker

Dan Jurafsky Information Extraction Subject: curriculum meeting Date: January 15, 2012 To: Dan Jurafsky Hi Dan, we’ve now scheduled the curriculum meeting. It will be in Gates 159 tomorrow from 10:00-11:30. -Chris 8 Create new Calendar entry Event: Curriculum mtg Date: Jan Start: 10:00am End: 11:30am Where: Gates 159

Dan Jurafsky Information Extraction & Sentiment Analysis nice and compact to carry! since the camera is small and light, I won't need to carry around those heavy, bulky professional cameras either! the camera feels flimsy, is plastic and very light in weight you have to be very delicate in the handling of this camera 9 Size and weight Attributes: zoom affordability size and weight flash ease of use ✓ ✗ ✓

Dan Jurafsky Machine Translation Fully automatic 10 Helping human translators Enter Source Text: Translation from Stanford’s Phrasal: 这 不过 是 一 个 时间 的 问题. This is only a matter of time.

Dan Jurafsky Language Technology Coreference resolution Question answering (QA) Part-of-speech (POS) tagging Word sense disambiguation (WSD) Paraphrase Named entity recognition (NER) Parsing Summarization Information extraction (IE) Machine translation (MT) Dialog Sentiment analysis mostly solved making good progress still really hard Spam detection Let’s go to Agra! Buy V1AGRA … ✓ ✗ Colorless green ideas sleep furiously. ADJ ADJ NOUN VERB ADV Einstein met with UN officials in Princeton PERSON ORG LOC You’re invited to our dinner party, Friday May 27 at 8:30 Party May 27 add Best roast chicken in San Francisco! The waiter ignored us for 20 minutes. Carter told Mubarak he shouldn’t run again. I need new batteries for my mouse. The 13 th Shanghai International Film Festival… 第 13 届上海国际电影节开幕 … The Dow Jones is up Housing prices rose Economy is good Q. How effective is ibuprofen in reducing fever in patients with acute febrile illness? I can see Alcatraz from the window! XYZ acquired ABC yesterday ABC has been taken over by XYZ Where is Citizen Kane playing in SF? Castro Theatre at 7:30. Do you want a ticket? The S&P500 jumped

Dan Jurafsky Ambiguity makes NLP hard: “Crash blossoms” Violinist Linked to JAL Crash Blossoms Teacher Strikes Idle Kids Red Tape Holds Up New Bridges Hospitals Are Sued by 7 Foot Doctors Juvenile Court to Try Shooting Defendant Local High School Dropouts Cut in Half 100% REAL

Dan Jurafsky Ambiguity is pervasive Fed raises interest rates New York Times headline (17 May 2000) Fed raises interest rates Fed raises interest rates 0.5%

Dan Jurafsky non-standard English Great Were SOO PROUD of what youve accomplished! U taught us 2 #neversaynever & you yourself should never give up either ♥ segmentation issues idioms dark horse get cold feet lose face throw in the towel neologisms unfriend Retweet bromance tricky entity names Where is A Bug’s Life playing … Let It Be was recorded … … a mutation on the for gene … world knowledge Mary and Sue are sisters. Mary and Sue are mothers. But that’s what makes it fun! the New York-New Haven Railroad Why else is natural language understanding difficult?

Dan Jurafsky Making progress on this problem… The task is difficult! What tools do we need? Knowledge about language Knowledge about the world A way to combine knowledge sources How we generally do this: probabilistic models built from language data P(“maison”  “house”) high P(“L’avocat général”  “the general avocado”) low Luckily, rough text features can often do half the job.

Dan Jurafsky This class Teaches key theory and methods for statistical NLP: Viterbi Naïve Bayes, Maxent classifiers N-gram language modeling Statistical Parsing Inverted index, tf-idf, vector models of meaning For practical, robust real-world applications Information extraction Spelling correction Information retrieval Sentiment analysis

17 Levels of Language Phonetics/phonology/morphology: what words (or subwords) are we dealing with? Syntax: What phrases are we dealing with? Which words modify one another? Semantics: What’s the literal meaning? Pragmatics: What should you conclude from the fact that I said something? How should you react?

18 What’s hard – ambiguities, ambiguities, all different levels of ambiguities John stopped at the donut store on his way home from work. He thought a coffee was good every few hours. But it turned out to be too expensive there. [from J. Eisner] - donut: To get a donut (doughnut; spare tire) for his car? - Donut store: store where donuts shop? or is run by donuts? or looks like a big donut? or made of donut? - From work: Well, actually, he stopped there from hunger and exhaustion, not just from work. - Every few hours: That’s how often he thought it? Or that’s for coffee? - it: the particular coffee that was good every few hours? the donut store? the situation - Too expensive: too expensive for what? what are we supposed to conclude about what John did?

19 NLP: The Main Issues Why is NLP difficult? –many “ words ”, many “ phenomena ” --> many “ rules ” OED: 400k words; Finnish lexicon (of forms): ~ sentences, clauses, phrases, constituents, coordination, negation, imperatives/questions, inflections, parts of speech, pronunciation, topic/focus, and much more! irregularity (exceptions, exceptions to the exceptions,...) potato -> potato es (tomato, hero,...); photo -> photo s, and even: both mango -> mango s or -> mango es Adjective / Noun order: new book, electrical engineering, general regulations, flower garden, garden flower,...: but Governor General

20 Difficulties in NLP (cont.) –ambiguity books: NOUN or VERB? –you need many books vs. she books her flights online No left turn weekdays 4-6 pm / except transit vehicles (Charles Street at Cold Spring) –when may transit vehicles turn: Always? Never? Thank you for not smoking, drinking, eating or playing radios without earphones. (MTA bus) –Thank you for not eating without earphones?? –or even: Thank you for not drinking without earphones!? My neighbor ’ s hat was taken by wind. He tried to catch it. –...catch the wind or...catch the hat ?

21 (Categorical) Rules or Statistics? Preferences: –clear cases: context clues: she books --> books is a verb –rule: if an ambiguous word (verb/nonverb) is preceded by a matching personal pronoun -> word is a verb – less clear cases: pronoun reference –she/he/it refers to the most recent noun or pronoun (?) (but maybe we can specify exceptions) –selectional: –catching hat >> catching wind (but why not?) –semantic: –never thank for drinking in a bus! (but what about the earphones?)

22 Solutions Don ’ t guess if you know: morphology (inflections) lexicons (lists of words) unambiguous names perhaps some (really) fixed phrases syntactic rules? Use statistics (based on real-world data) for preferences (only?) No doubt: but this is the big question!

23 Statistical NLP Imagine: –Each sentence W = { w 1, w 2,..., w n } gets a probability P(W|X) in a context X (think of it in the intuitive sense for now) –For every possible context X, sort all the imaginable sentences W according to P(W|X): –Ideal situation: best sentence (most probable in context X) NB: same for interpretation P(W) “ ungrammatical ” sentences

24 Real World Situation Unable to specify set of grammatical sentences today using fixed “ categorical ” rules (maybe never) Use statistical “ model ” based on REAL WORLD DATA and care about the best sentence only (disregarding the “ grammaticality ” issue) best sentence P(W) W best W worst