CPSC 7373: Artificial Intelligence Lecture 12: Hidden Markov Models and Filters Jiang Bian, Fall 2012 University of Arkansas at Little Rock.

Slides:



Advertisements
Similar presentations
Lecture 16 Hidden Markov Models. HMM Until now we only considered IID data. Some data are of sequential nature, i.e. have correlations have time. Example:
Advertisements

Hidden Markov Models. Room Wandering I’m going to wander around my house and tell you objects I see. Your task is to infer what room I’m in at every point.
Lirong Xia Probabilistic reasoning over time Tue, March 25, 2014.
Automatic Speech Recognition II  Hidden Markov Models  Neural Network.
Reasoning Under Uncertainty: Bayesian networks intro Jim Little Uncertainty 4 November 7, 2014 Textbook §6.3, 6.3.1, 6.5, 6.5.1,
Hidden Markov Models Reading: Russell and Norvig, Chapter 15, Sections
Chapter 15 Probabilistic Reasoning over Time. Chapter 15, Sections 1-5 Outline Time and uncertainty Inference: ltering, prediction, smoothing Hidden Markov.
Introduction to Hidden Markov Models
Hidden Markov Models Bonnie Dorr Christof Monz CMSC 723: Introduction to Computational Linguistics Lecture 5 October 6, 2004.
2004/11/161 A Tutorial on Hidden Markov Models and Selected Applications in Speech Recognition LAWRENCE R. RABINER, FELLOW, IEEE Presented by: Chi-Chun.
Hidden Markov Models Adapted from Dr Catherine Sweeney-Reed’s slides.
Ch-9: Markov Models Prepared by Qaiser Abbas ( )
Chapter 6: HIDDEN MARKOV AND MAXIMUM ENTROPY Heshaam Faili University of Tehran.
Hidden Markov Models Theory By Johan Walters (SR 2003)
Hidden Markov Models in NLP
Hidden Markov Model based 2D Shape Classification Ninad Thakoor 1 and Jean Gao 2 1 Electrical Engineering, University of Texas at Arlington, TX-76013,
Hidden Markov Models (HMMs) Steven Salzberg CMSC 828H, Univ. of Maryland Fall 2010.
Advanced Artificial Intelligence
Introduction to Sequence Models. Sequences Many types of information involve sequences: -Financial data: -DNA: -Robot motionRobot motion -Text: “Jack.
Albert Gatt Corpora and Statistical Methods Lecture 8.
Probabilistic reasoning over time So far, we’ve mostly dealt with episodic environments –One exception: games with multiple moves In particular, the Bayesian.
PatReco: Hidden Markov Models Alexandros Potamianos Dept of ECE, Tech. Univ. of Crete Fall
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Hidden Markov Models K 1 … 2. Outline Hidden Markov Models – Formalism The Three Basic Problems of HMMs Solutions Applications of HMMs for Automatic Speech.
. Hidden Markov Models with slides from Lise Getoor, Sebastian Thrun, William Cohen, and Yair Weiss.
Chapter 3 (part 3): Maximum-Likelihood and Bayesian Parameter Estimation Hidden Markov Model: Extension of Markov Chains All materials used in this course.
CPSC 422, Lecture 14Slide 1 Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 14 Feb, 4, 2015 Slide credit: some slides adapted from Stuart.
CS 188: Artificial Intelligence Fall 2009 Lecture 19: Hidden Markov Models 11/3/2009 Dan Klein – UC Berkeley.
Particle Filters++ TexPoint fonts used in EMF.
CHAPTER 15 SECTION 3 – 4 Hidden Markov Models. Terminology.
Ch10 HMM Model 10.1 Discrete-Time Markov Process 10.2 Hidden Markov Models 10.3 The three Basic Problems for HMMS and the solutions 10.4 Types of HMMS.
Particle Filter & Search
BINF6201/8201 Hidden Markov Models for Sequence Analysis
Fundamentals of Hidden Markov Model Mehmet Yunus Dönmez.
CS774. Markov Random Field : Theory and Application Lecture 19 Kyomin Jung KAIST Nov
UIUC CS 498: Section EA Lecture #21 Reasoning in Artificial Intelligence Professor: Eyal Amir Fall Semester 2011 (Some slides from Kevin Murphy (UBC))
Homework 1 Reminder Due date: (till 23:59) Submission: – – Write the names of students in your team.
Kalman Filter Notes Prateek Tandon.
Mobile Robot Localization (ch. 7)
Processing Sequential Sensor Data The “John Krumm perspective” Thomas Plötz November 29 th, 2011.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Reestimation Equations Continuous Distributions.
CS 416 Artificial Intelligence Lecture 17 Reasoning over Time Chapter 15 Lecture 17 Reasoning over Time Chapter 15.
CS Statistical Machine learning Lecture 24
1 CS 552/652 Speech Recognition with Hidden Markov Models Winter 2011 Oregon Health & Science University Center for Spoken Language Understanding John-Paul.
1 CONTEXT DEPENDENT CLASSIFICATION  Remember: Bayes rule  Here: The class to which a feature vector belongs depends on:  Its own value  The values.
CHAPTER 8 DISCRIMINATIVE CLASSIFIERS HIDDEN MARKOV MODELS.
 Present by 陳群元.  Introduction  Previous work  Predicting motion patterns  Spatio-temporal transition distribution  Discerning pedestrians  Experimental.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Elements of a Discrete Model Evaluation.
CPS 170: Artificial Intelligence Markov processes and Hidden Markov Models (HMMs) Instructor: Vincent Conitzer.
Probability and Time. Overview  Modelling Evolving Worlds with Dynamic Baysian Networks  Simplifying Assumptions Stationary Processes, Markov Assumption.
Hidden Markov Models Sean Callen Joel Henningsen.
CS Statistical Machine learning Lecture 25 Yuan (Alan) Qi Purdue CS Nov
EEL 6586: AUTOMATIC SPEECH PROCESSING Hidden Markov Model Lecture Mark D. Skowronski Computational Neuro-Engineering Lab University of Florida March 31,
Hidden Markov Model Parameter Estimation BMI/CS 576 Colin Dewey Fall 2015.
Particle filters for Robot Localization An implementation of Bayes Filtering Markov Localization.
Instructor: Eyal Amir Grad TAs: Wen Pu, Yonatan Bisk Undergrad TAs: Sam Johnson, Nikhil Johri CS 440 / ECE 448 Introduction to Artificial Intelligence.
Visual Recognition Tutorial1 Markov models Hidden Markov models Forward/Backward algorithm Viterbi algorithm Baum-Welch estimation algorithm Hidden.
CS 541: Artificial Intelligence Lecture VIII: Temporal Probability Models.
Hidden Markov Models HMM Hassanin M. Al-Barhamtoshy
Instructor: Vincent Conitzer
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 14
CSCI 5822 Probabilistic Models of Human and Machine Learning
CS 188: Artificial Intelligence Spring 2007
HCI/ComS 575X: Computational Perception
Hassanin M. Al-Barhamtoshy
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 14
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Instructor: Vincent Conitzer
Instructor: Vincent Conitzer
Presentation transcript:

CPSC 7373: Artificial Intelligence Lecture 12: Hidden Markov Models and Filters Jiang Bian, Fall 2012 University of Arkansas at Little Rock

Hidden Markov Models Hidden Markov Models (HMMs) – to analyze, or – to predict – time series Applications: – Robotics – Medical – Finance – Speech and – Language Technologies – etc.

Bayes Network of HMMs HMMs – A sequence of states that evolves over time; – Each state only depends on the previous state; and – Each state emits a measurement Filters – Kalman Filters – Particle Filters S1 S2 S3 Sn... Z1 Z2 Z3 Zn

Markov Chain R R S S The initial state: P(R0) = 1 P(S0) = 0 P(R1) = ??? P(R2) = ??? P(R3) = ???

Markov Chain R R S S The initial state: P(R0) = 1 P(S0) = 0 P(R1) = 0.6 P(R1|R0) * P(R0) = 0.6 P(R2) = 0.44 P(R2) = P(R2|R1) * P(R1) + P(R2|S1)*P(S1) = 0.44 P(R3) = P(R3) = P(R3|R2) * P(R2) + P(R3|S2)*P(S2) = 0.376

Markov Chain P(A0) = 1 – P(A1) = ??? – P(A2) = ??? – P(A3) = ??? A A B B 0.5 1

Markov Chain P(A0) = 1 – P(A1) = 0.5 – P(A2) = 0.75 – P(A3) = A A B B 0.5 1

Stationary Distribution P(A1000) = ??? Stationary Distribution: – P(At) = P(At-1) – P(At|At-1)*P(At-1) +P(At|Bt-1)*P(Bt-1) – P(At) = X X = 0.5 * X + 1 * (1- X) X = 2/3 = P(A) P(B) = 1 – P(A) = 1/3 A A B B 0.5 1

Stationary Distribution R R S S

Stationary Distribution R R S S

Transition Probabilities Finding the transition probabilities from observation. – e.g., R, S, S, S, R, S, R (7 days) Maximum Likelihood – P(R0) = 1 – P(S|R) = 1, P(R|R) = 0 – P(S|S) = 0.5, P(R|S) = 0.5 R R S S ? ? ? ?

Transition Probabilities - Quiz Finding the transition probabilities from observation. – e.g., S, S, S, S, S, R, S, S, S, R, R Maximum Likelihood – P(R0) = ??; P(S0) = ?? – P(S|R) = ??, P(R|R) = ?? – P(S|S) = ??, P(R|S) = ?? R R S S ? ? ? ?

Transition Probabilities - Quiz Finding the transition probabilities from observation. – e.g., S, S, S, S, S, R, S, S, S, R, R Maximum Likelihood – P(R0) = 0; P(S0) = 1 – P(S|R) = 0.5, P(R|R) = 0.5 – P(S|S) = 0.75, P(R|S) = 0.25 R R S S ? ? ? ?

Laplacian Smoothing Laplacian smoothing; k = 1 – P(R0) = ??; P(S0) = ?? – P(S|R) = ??, P(R|R) = ?? – P(S|S) = ??, P(R|S) = ?? R R S S ? ? ? ?

Laplacian Smoothing R, S, S, S, S Laplacian smoothing; k = 1 – P(R0) = 2/3; P(S0) = 1/3 – P(S|R) = 4/5, P(R|R) = 1/5 – P(S|S) = 2/3, P(R|S) = 1/3 R R S S ? ? ? ?

Hidden Markov Models Suppose that I can’t observe the weather (e.g., I am grounded with no windows); But I can make a guess based on whether my wife carries a umbrella; and – P(U|R) = 0.9; P(-U|R) = 0.1 – P(U|S) = 0.2; P(-U|S) = 0.8 P(R1|U1) = P(U1|R1) * P(R1) / P(U1) = 0.9 * 0.4 / (0.4* *0.6) = 0.75 – P(R1) = P(R1|R0)P(R0) + P(R1|S0) P(S0) = 0.4 R R S S U U U U P(R0) = ½ P(S0) = ½

Specification of an HMM N - number of states – Q = {q 1 ; q 2 ; : : : ;q T } - set of states M - the number of symbols (observables) – O = {o 1 ; o 2 ; : : : ;o T } - set of symbols A - the state transition probability matrix – aij = P(q t+1 = j|q t = i) B- observation probability distribution – b j (k) = P(o t = k|q t = j) i ≤ k ≤ M π - the initial state distribution

Specification of an HMM Full HMM is thus specified as a triplet: – λ = (A,B,π)

Central problems in HMM modelling Problem 1 Evaluation: – Probability of occurrence of a particular observation sequence, O = {o 1,…,o k }, given the model – P(O|λ) – Complicated – hidden states – Useful in sequence classification

Central problems in HMM modelling Problem 2 Decoding: – Optimal state sequence to produce given observations, O = {o 1,…,o k }, given model – Optimality criterion – Useful in recognition problems

Central problems in HMM modelling Problem 3 Learning: – Determine optimum model, given a training set of observations – Find λ, such that P(O|λ) is maximal

Particle Filter

Particle Filters

Sensor Information: Importance Sampling

Robot Motion

Sensor Information: Importance Sampling

Robot Motion

1. Algorithm particle_filter( S t-1, u t-1 z t ): For Generate new samples 4. Sample index j(i) from the discrete distribution given by w t-1 5. Sample from using and 6. Compute importance weight 7. Update normalization factor 8. Insert 9. For 10. Normalize weights Particle Filter Algorithm

Particle Filters Pros: – easy to implement – Work well in many applications Cons: – Don’t work in high dimension spaces – Problems with degenerate conditions. Very few particles, and Not much noise in either measurements or controls.