Natural Language Processing

Slides:



Advertisements
Similar presentations
For Monday Read Chapter 23, sections 3-4 Homework –Chapter 23, exercises 1, 6, 14, 19 –Do them in order. Do NOT read ahead.
Advertisements

January 12, Statistical NLP: Lecture 2 Introduction to Statistical NLP.
Natural Language and Speech Processing Creation of computational models of the understanding and the generation of natural language. Different fields coming.
Advanced AI - Part II Luc De Raedt University of Freiburg WS 2004/2005 Many slides taken from Helmut Schmid.
Part II. Statistical NLP Advanced Artificial Intelligence Part of Speech Tagging Wolfram Burgard, Luc De Raedt, Bernhard Nebel, Lars Schmidt-Thieme Most.
Tom Griffiths CogSci C131/Psych C123 Computational Models of Cognition.
Introduction to CL Session 1: 7/08/2011. What is computational linguistics? Processing natural language text by computers  for practical applications.
تمرين شماره 1 درس NLP سيلابس درس NLP در دانشگاه هاي ديگر ___________________________ راحله مکي استاد درس: دکتر عبدالله زاده پاييز 85.
Machine Learning in Natural Language Processing Noriko Tomuro November 16, 2006.
Statistical Natural Language Processing Advanced AI - Part II Luc De Raedt University of Freiburg WS 2005/2006 Many slides taken from Helmut Schmid.
COP4020 Programming Languages
Fundamentals: Linguistic principles
Models of Generative Grammar Smriti Singh. Generative Grammar  A Generative Grammar is a set of formal rules that can generate an infinite set of sentences.
Statistical Natural Language Processing. What is NLP?  Natural Language Processing (NLP), or Computational Linguistics, is concerned with theoretical.
Lecture 1, 7/21/2005Natural Language Processing1 CS60057 Speech &Natural Language Processing Autumn 2005 Lecture 1 21 July 2005.
9/8/20151 Natural Language Processing Lecture Notes 1.
Part II. Statistical NLP Advanced Artificial Intelligence Applications of HMMs and PCFGs in NLP Wolfram Burgard, Luc De Raedt, Bernhard Nebel, Lars Schmidt-Thieme.
BTANT 129 w5 Introduction to corpus linguistics. BTANT 129 w5 Corpus The old school concept – A collection of texts especially if complete and self-contained:
Lecture 12: 22/6/1435 Natural language processing Lecturer/ Kawther Abas 363CS – Artificial Intelligence.
For Friday Finish chapter 23 Homework: –Chapter 22, exercise 9.
Chapter 10: Compilers and Language Translation Invitation to Computer Science, Java Version, Third Edition.
Some Probability Theory and Computational models A short overview.
Introduction to CL & NLP CMSC April 1, 2003.
Artificial Intelligence By Michelle Witcofsky And Evan Flanagan.
인공지능 연구실 황명진 FSNLP Introduction. 2 The beginning Linguistic science 의 4 부분 –Cognitive side of how human acquire, produce, and understand.
1 CSI 5180: Topics in AI: Natural Language Processing, A Statistical Approach Instructor: Nathalie Japkowicz Objectives of.
Copyright © Curt Hill Languages and Grammars This is not English Class. But there is a resemblance.
Introduction Chapter 1 Foundations of statistical natural language processing.
For Friday Finish chapter 23 Homework –Chapter 23, exercise 15.
1 An Introduction to Computational Linguistics Mohammad Bahrani.
Exploiting Named Entity Taggers in a Second Language Thamar Solorio Computer Science Department National Institute of Astrophysics, Optics and Electronics.
Overview of Statistical NLP IR Group Meeting March 7, 2006.
Reinforcement Learning for Mapping Instructions to Actions S.R.K. Branavan, Harr Chen, Luke S. Zettlemoyer, Regina Barzilay Computer Science and Artificial.
CS416 Compiler Design1. 2 Course Information Instructor : Dr. Ilyas Cicekli –Office: EA504, –Phone: , – Course Web.
Part-Of-Speech Tagging Radhika Mamidi. POS tagging Tagging means automatic assignment of descriptors, or tags, to input tokens. Example: “Computational.
Speech Recognition
What is a Hidden Markov Model?
Approaches to Machine Translation
Sentiment analysis algorithms and applications: A survey
Natural Language Processing with Qt
Programming Languages Translator
Introduction to Parsing (adapted from CS 164 at Berkeley)
What does it mean? Notes from Robert Sebesta Programming Languages
Natural Language Processing (NLP)
Statistical NLP: Lecture 13
CS416 Compiler Design lec00-outline September 19, 2018
CSE 3302 Programming Languages
--Mengxue Zhang, Qingyang Li
Introduction CI612 Compiler Design CI612 Compiler Design.
Machine Learning in Natural Language Processing
Implementing AI solutions using the cognitive services in Azure
KNOWLEDGE REPRESENTATION
Approaches to Machine Translation
CS4705 Natural Language Processing
Natural Language Processing
Language and Learning Introduction to Artificial Intelligence COS302
CS416 Compiler Design lec00-outline February 23, 2019
Teori Bahasa dan Automata Lecture 9: Contex-Free Grammars
CPSC 503 Computational Linguistics
Natural Language Processing (NLP)
Chapter 10: Compilers and Language Translation
Lec00-outline May 18, 2019 Compiler Design CS416 Compiler Design.
Discrete Maths 13. Grammars Objectives
The Winograd Schema Challenge Hector J. Levesque AAAI, 2011
Artificial Intelligence 2004 Speech & Natural Language Processing
Information Retrieval
COMPILER CONSTRUCTION
CS249: Neural Language Model
Natural Language Processing (NLP)
Presentation transcript:

Natural Language Processing Chris O’Neill Very brief overview of NLP – very large field of study

Overview What is NLP Main challenges with NLP Rule-based vs Statistical NLP Hidden Markov Models Modern Applications of NLP Overview

What is Natural Language Processing? Goal: “analyze and process human language, broadly, robustly, and accurately” Computers being able to understand how we speak and write The goal of Natural Language Processing, as stated during the 2017 Lecture Video entitled “Advanced Topics NLP”, (I think Adam Lopez??)

3 Main challenges of NLP Ambiguous Grammar Ambiguous Pronunciation Ambiguous Words

Ambiguous Grammar “Enraged Cow Injures Farmer With Ax” Who is holding the ax? How do we determine the answer of this question? Context is key in understanding the meaning of the sentence

Ambiguous Pronunciation “Put your heart, mind, and soul into even your smallest acts. This is the secret of success.” Swami Sivananda Words that have same pronunciation – hard to capture the semantics (meaning of words) By reading this, it’s easy to know that soul is not referring to sole (right foot), because we can see the spelling (and bottom of foot would make no sense) Imagine if you simply heard this quote? How do we determine the meaning of soul? How do we differentiate soul from sole? Probabilistic approaches – given the surrounding words, how often does the bottom of foot occur?

Ambiguous Words The bear crossed the river The tree failed to bear fruit Problem in speech recognition and in understanding written language - solve this with POS tagging How do we humans understand this language? We know context!! Context is a major factor in understanding language

Context is Everything Word by word translation does not work Walking in Russia with dictionary in hand and no body language is allowed – expressing meaningful phrases would be nearly impossible

Universal Grammar Noam Chomsky “the idea that human languages, as superficially diverse as they are, share some fundamental similarities, and that these are attributable to innate principles unique to language: that deep down, there is only one human language.” Syntactic Structures Noam Chomsky – linguist back in the mid 20th century Frontiers in Psychology journal explains the idea of Universal Grammar Chomsky published Syntactic Structures which explains that every language has a set of rules (grammars) and that through these rules there is an infinite amount of sentences to express. This idea that understanding the grammars (rules) was the key to unlocking the human language led people to focus on the syntax of languages to derive semantics Boise State professor – simple mathematical proof that confirms this idea of grammar allows for infinite expressible sentences

Rule-based NLP What is it? Pros Cons Hand-crafted rules Linguists sit in a room and think of all possible rules associated with a language Example: Context-Free Grammar There is minimal training required for the AI agent to function Not much else Development is dependent on linguists forming rules Not robust – there are many exceptions to rules in languages High maintenance An example of a rule-based approach would be a context-free grammar. In a context-free grammar there is a start symbol, productions, nonterminating symbols, and terminating symbols. The productions are the rules that specify how to parse nonterminating symbols to eventually get to terminating symbols. The popularity of this approach is most likely due to the fact that, since there wasn’t much computing power (well at least not enough required to perform machine learning algorithms) Lack of statistical techniques

Statistical NLP Statistical revolution in 80s/90s Machine Learning systems the ability to automatically learn and improve from experience without being explicitly programmed.  Machine learning focuses on the development of computer programs that can access data and use it learn for themselves.”

Statistical NLP What’s required Main tasks Computer Corpora Software Tokenize Analyze Interpret the data Tokenize: The system will need to be able to extract words, sentences, and punctuations into tokens. Analyze: The system will need to be able to extract the features of these tokenized words and sentences. Features can range from the sequence of words the preceded this word, to the morphology of a word, to even as simple as the number of times a word appears in a document. Interpret the data: Run the features through some probabilistic model that will produce an output that gives direction or meaning to the NLP system. The third task as mentioned above refers to having a probabilistic model with parameters and features that the system can update and tweak. This updating and tweaking of the model is what is how a machine can “learn”. There are many different types of probabilistic models in NLP, but two popular ones are Hidden Markov Models (HMMs) and Neural Networks.

Machine Translation: machine translation is the transformation of text from one language to another performed automatically by a computer. 2 training corpora Target language corpus – to learn what target language sentences look like, learn phrases Source-Target corpus – to find correspondences between languages

Hidden Markov Models “a tool for representing probability distributions over sequences of observations” Markov Property States are not fully observed Quote from International Journal of Pattern Recognition and Artificial Intelligence Markov Property - at a given state, all pertinent data is supplied to predict the future states Through observations (like audio waves, the word(s) preceding a word, the part of speech of a given word) we can infer states

Types of Hidden Markov Models History-based model First-order HMM Mth-order HMM History-based model: Every word depends on all the preceding words --- That’s a ton of parameters!!! Lots of conditional probabilities to compute and does not scale well First-order HMM (sometimes referred to as a bigram HMM): satisfies markov property: probability of next state solely depends on previous state - problem with this? Realistically, it’s insufficient to only consider the preceding word…it’s difficult to truly capture context Mth-order HMM: take the m preceding words into account

State space is the current word we are trying to translate Emission – the corresponding Spanish word Differences – need to be sensitive to what is actually present on the English side (Is this truly an HMM – is state truly hidden??) and needs to remember context of current word How do we keep track of context with HMM? - this type of HMM is more complicated - Probably use an mth-ordered HMM - Probably have a POS-tagger to help with translation for disambiguation

Other application of NLP Document Similarity Optical Character Recognition (OCR) Speech Recognition Natural Language Generation (NLG) Other application of NLP Document Similarity – algorithm to determine how similar documents are OCR - ABBYY FineReader OCR allows businesses to convert documents in just about any format into data that can be easily analyzed and manipulated Speech Recognition – Google Assistant, Siri, Cortana – allows for humans to interact with computers (Google analogy of humans trying to communicate with each other by writing messages to each other) NLG - Natural Language Generation is the process of consuming data, interpreting that data, and producing an unbiased summary of the data in human-understandable language

Conclusion Defined Natural Language Processing Discussed main challenges Described two main approaches of NLP Hidden Markov Models Modern applications of NLP

Questions