NLP. Introduction to NLP Very important for language processing Example in speech recognition: –“recognize speech” vs “wreck a nice beach” Example in.

Slides:



Advertisements
Similar presentations
1 Essential Probability & Statistics (Lecture for CS598CXZ Advanced Topics in Information Retrieval ) ChengXiang Zhai Department of Computer Science University.
Advertisements

Copyright © 2006 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Slide
0 0 Review Probability Axioms –Non-negativity P(A)≥0 –Additivity P(A U B) =P(A)+ P(B), if A and B are disjoint. –Normalization P(Ω)=1 Independence of two.
INTRODUCTION TO ARTIFICIAL INTELLIGENCE Massimo Poesio LECTURE 11 (Lab): Probability reminder.
Chapter 4 Probability and Probability Distributions
Probability Predictions Ch. 1, Act. 5. Probability The study of random events. Random events are things that happen without predictability – e.g. the.
How likely something is to happen.
Conditional Probability and Independence. Learning Targets 1. I can calculate conditional probability using a 2-way table. 2. I can determine whether.
Solve for x. 28 = 4(2x + 1) = 8x = 8x + 8 – 8 – 8 20 = 8x = x Distribute Combine Subtract Divide.
Probability Sample Space Diagrams.
Lecture 10 – Introduction to Probability Topics Events, sample space, random variables Examples Probability distribution function Conditional probabilities.
Random Variables.
ST3236: Stochastic Process Tutorial TA: Mar Choong Hock
June 3, 2008Stat Lecture 6 - Probability1 Probability Introduction to Probability, Conditional Probability and Random Variables Statistics 111 -
8.7 Probability. Ex 1 Find the sample space for each of the following. One coin is tossed. Two coins are tossed. Three coins are tossed.
1 BASIC NOTIONS OF PROBABILITY THEORY. NLE 2 What probability theory is for Suppose that we have a fair dice, with six faces, and that we keep throwing.
Probability Notation Review Prior (unconditional) probability is before evidence is obtained, after is posterior or conditional probability P(A) – Prior.
CSE 221: Probabilistic Analysis of Computer Systems Topics covered: Discrete random variables Probability mass function Distribution function (Secs )
Probability theory Much inspired by the presentation of Kren and Samuelsson.
Chapter 2: Probability.
UNR, MATH/STAT 352, Spring Random variable (rv) – central concept of probability theory Numerical description of experiment.
September SOME BASIC NOTIONS OF PROBABILITY THEORY Universita’ di Venezia 29 Settembre 2003.
Visualizing Events Contingency Tables Tree Diagrams Ace Not Ace Total Red Black Total
C4: DISCRETE RANDOM VARIABLES CIS 2033 based on Dekking et al. A Modern Introduction to Probability and Statistics Longin Jan Latecki.
Review of Probability Theory. © Tallal Elshabrawy 2 Review of Probability Theory Experiments, Sample Spaces and Events Axioms of Probability Conditional.
Chapter 9 Introducing Probability - A bridge from Descriptive Statistics to Inferential Statistics.
Lecture 10 – Introduction to Probability Topics Events, sample space, random variables Examples Probability distribution function Conditional probabilities.
2. Mathematical Foundations
Sets, Combinatorics, Probability, and Number Theory Mathematical Structures for Computer Science Chapter 3 Copyright © 2006 W.H. Freeman & Co.MSCS SlidesProbability.
Sets, Combinatorics, Probability, and Number Theory Mathematical Structures for Computer Science Chapter 3 Copyright © 2006 W.H. Freeman & Co.MSCS SlidesProbability.
1 CY1B2 Statistics Aims: To introduce basic statistics. Outcomes: To understand some fundamental concepts in statistics, and be able to apply some probability.
Discrete Random Variables: PMFs and Moments Lemon Chapter 2 Intro to Probability
1 Chapters 6-8. UNIT 2 VOCABULARY – Chap 6 2 ( 2) THE NOTATION “P” REPRESENTS THE TRUE PROBABILITY OF AN EVENT HAPPENING, ACCORDING TO AN IDEAL DISTRIBUTION.
Random Variables. A random variable X is a real valued function defined on the sample space, X : S  R. The set { s  S : X ( s )  [ a, b ] is an event}.
Dr. Omar Al Jadaan Probability. Simple Probability Possibilities and Outcomes Expressed in the form of a fraction A/B Where A is the occurrence B is possible.
November 2004CSA4050: Crash Concepts in Probability1 CSA4050: Advanced Topics in NLP Probability I Experiments/Outcomes/Events Independence/Dependence.
The Cartoon Guide to Statistics
CS433 Modeling and Simulation Lecture 03 – Part 01 Probability Review 1 Dr. Anis Koubâa Al-Imam Mohammad Ibn Saud University
Copyright © Cengage Learning. All rights reserved. Elementary Probability Theory 5.
Mathematical Foundations Elementary Probability Theory Essential Information Theory Updated 11/11/2005.
PROBABILITY IN OUR DAILY LIVES
12/7/20151 Probability Introduction to Probability, Conditional Probability and Random Variables.
Conditional Probability Mass Function. Introduction P[A|B] is the probability of an event A, giving that we know that some other event B has occurred.
Probability (outcome k) = Relative Frequency of k
Copyright © 2006 The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Review of Statistics I: Probability and Probability Distributions.
Probability and Distributions. Deterministic vs. Random Processes In deterministic processes, the outcome can be predicted exactly in advance Eg. Force.
Measuring chance Probabilities FETP India. Competency to be gained from this lecture Apply probabilities to field epidemiology.
AP STATISTICS Section 7.1 Random Variables. Objective: To be able to recognize discrete and continuous random variables and calculate probabilities using.
Probability. Probability Probability is fundamental to scientific inference Probability is fundamental to scientific inference Deterministic vs. Probabilistic.
Unit 4 Section 3.1.
Probability VOCAB!. What is probability? The probability of an event is a measure of the likelihood that the event will occur. When all outcomes are equally.
9/14/1999 JHU CS /Jan Hajic 1 Introduction to Natural Language Processing Probability AI-Lab
Chapter 2: Probability. Section 2.1: Basic Ideas Definition: An experiment is a process that results in an outcome that cannot be predicted in advance.
Statistical NLP: Lecture 4 Mathematical Foundations I: Probability Theory (Ch2)
Essential Probability & Statistics (Lecture for CS397-CXZ Algorithms in Bioinformatics) Jan. 23, 2004 ChengXiang Zhai Department of Computer Science University.
Random Variables Lecture Lecturer : FATEN AL-HUSSAIN.
PROBABILITY 1. Basic Terminology 2 Probability 3  Probability is the numerical measure of the likelihood that an event will occur  The probability.
Statistical NLP Course for Master in Computational Linguistics 2nd Year Diana Trandabat.
Probability. Today we will look at… 1.Quick Recap from last week 2.Terminology relating to events and outcomes 3.Use of sample spaces when dealing with.
Conditional Probability 423/what-is-your-favorite-data-analysis-cartoon 1.
Umm Al-Qura University
Introduction to Probability
Business Statistics Topic 4
Probability Review Definitions/Identities Random Variables
Advanced Artificial Intelligence
Statistical NLP: Lecture 4
I flip a coin two times. What is the sample space?
ASV Chapters 1 - Sample Spaces and Probabilities
POPULATION (of “units”)
M248: Analyzing data Block A UNIT A3 Modeling Variation.
Presentation transcript:

NLP

Introduction to NLP

Very important for language processing Example in speech recognition: –“recognize speech” vs “wreck a nice beach” Example in machine translation: –“l’avocat general”: “the attorney general” vs. “the general avocado” Probabilities make it possible to combine evidence from multiple sources in a systematic way.

Probability theory –predicting how likely it is that something will happen Experiment (trial) –e.g., throwing a coin Possible outcomes –heads or tails Sample spaces –discrete or continuous Events –  is the certain event –  is the impossible event –event space - all possible events

Probabilities –numbers between 0 and 1 Probability distribution –distributes a probability mass of 1 throughout the sample space . Example: –A fair coin is tossed three times. –What is the probability of 3 heads? –What is the probability of 2 heads?

Frequentist –I threw the coin 10 times and it turned up heads 5 times Subjective –I am willing to bet 50 cents on heads

p(  ) = 0 P(certain event)=1 p(X)  p(Y), if X  Y p(X  Y) = p(X) + p(Y), if X  Y= 

Prior and posterior probability Conditional probability P(A|B) = P(A  B) P(B)  A B ABAB

Six-sided fair die –P(D even)=? –P(D>=4)=? –P(D even|D>=4)=? –P(D odd|D>=4)=? Multiple conditions –P(D odd|D>=4, D<=5)=?

Six-sided fair die –P(D even)=3/6=1/2 –P(D>=4)=3/6=1/2 –P(D even|D>=4)=2/3 –P(D odd|D>=4)=1/3 Multiple conditions –P(D odd|D>=4, D<=5)=1/2

P(w 1,w 2,w 3 …w n ) = ? Using the chain rule: –P(w 1,w 2,w 3 …w n ) =P(w 1 ) P(w 2 |w 1 ) P(w 3 |w 1,w 2 )… P(w n |w 1,w 2 …w n-1 ) This rule is used in many ways in statistical NLP, more specifically in Markov Models

Two events are independent when –P(A  B) = P(A)P(B) Unless P(B)=0 this is equivalent to saying that P(A) = P(A|B) If two events are not independent, they are considered dependent

Adding constraints –P(walk=yes|weather=nice) –P(walk=yes|weather=nice,freetime=yes,crowded=yes) –More accurate –But more difficult to estimate Removing constraints (Backoff) –P(walk=yes|weather=nice,freetime=yes,crowded=yes) –P(walk=yes|weather=nice,freetime=yes) –P(walk=yes|weather=nice) –Note that it is not possible to do backoff on the left hand side of the conditional

X:   R n Simply a function: The numbers are generated by a stochastic process with a certain probability distribution Example –the discrete random variable X that is the sum of the faces of two randomly thrown fair dice Probability mass function (pmf) which gives the probability that the random variable has different numeric values: P(x) = P(X = x) = P(A x ) where A x = {    : X(  ) = x}

If a random variable X is distributed according to the pmf p(x), then we write X ~ p(x) For a discrete random variable, we have  p(x i ) = P(  ) = 1

p(1) = 1/6 p(2) = 1/6 etc. P(D)=? P(D) = {1/6, 1/6, 1/6, 1/6, 1/6, 1/6} P(D|odd) = {1/3, 0, 1/3, 0, 1/3, 0}

NLP