Hidden Markov Models Dave DeBarr

Slides:



Advertisements
Similar presentations
Lecture 16 Hidden Markov Models. HMM Until now we only considered IID data. Some data are of sequential nature, i.e. have correlations have time. Example:
Advertisements

CPSC 422, Lecture 16Slide 1 Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 16 Feb, 11, 2015.
Russell and Norvig, AIMA : Chapter 15 Part B – 15.3,
Hidden Markov Models Reading: Russell and Norvig, Chapter 15, Sections
Chapter 15 Probabilistic Reasoning over Time. Chapter 15, Sections 1-5 Outline Time and uncertainty Inference: ltering, prediction, smoothing Hidden Markov.
Hidden Markov Models Bonnie Dorr Christof Monz CMSC 723: Introduction to Computational Linguistics Lecture 5 October 6, 2004.
Cognitive Computer Vision
Advanced Artificial Intelligence
Sequential Modeling with the Hidden Markov Model Lecture 9 Spoken Language Processing Prof. Andrew Rosenberg.
1 Reasoning Under Uncertainty Over Time CS 486/686: Introduction to Artificial Intelligence Fall 2013.
INTRODUCTION TO Machine Learning 3rd Edition
Part II. Statistical NLP Advanced Artificial Intelligence Hidden Markov Models Wolfram Burgard, Luc De Raedt, Bernhard Nebel, Lars Schmidt-Thieme Most.
Speech Recognition. What makes speech recognition hard?
… Hidden Markov Models Markov assumption: Transition model:
On Recognizing Music Using HMM Following the path craved by Speech Recognition Pioneers.
Hidden Markov Models. Hidden Markov Model In some Markov processes, we may not be able to observe the states directly.
CS 188: Artificial Intelligence Fall 2009 Lecture 21: Speech Recognition 11/10/2009 Dan Klein – UC Berkeley TexPoint fonts used in EMF. Read the TexPoint.
Temporal Processes Eran Segal Weizmann Institute.
CSCI 5582 Fall 2006 CSCI 5582 Artificial Intelligence Lecture 15 Jim Martin.
1 Hidden Markov Model Instructor : Saeed Shiry  CHAPTER 13 ETHEM ALPAYDIN © The MIT Press, 2004.
CPSC 422, Lecture 14Slide 1 Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 14 Feb, 4, 2015 Slide credit: some slides adapted from Stuart.
CHAPTER 15 SECTION 3 – 4 Hidden Markov Models. Terminology.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
1 7-Speech Recognition (Cont’d) HMM Calculating Approaches Neural Components Three Basic HMM Problems Viterbi Algorithm State Duration Modeling Training.
HMM - Basics.
Speech, Perception, & AI Artificial Intelligence CMSC March 5, 2002.
Entropy & Hidden Markov Models Natural Language Processing CMSC April 22, 2003.
LML Speech Recognition Speech Recognition Introduction I E.M. Bakker.
UIUC CS 498: Section EA Lecture #21 Reasoning in Artificial Intelligence Professor: Eyal Amir Fall Semester 2011 (Some slides from Kevin Murphy (UBC))
Speech, Perception, & AI Artificial Intelligence CMSC February 13, 2003.
Homework 1 Reminder Due date: (till 23:59) Submission: – – Write the names of students in your team.
Kalman Filter Notes Prateek Tandon.
Hidden Markov Models: Decoding & Training Natural Language Processing CMSC April 24, 2003.
CS 416 Artificial Intelligence Lecture 17 Reasoning over Time Chapter 15 Lecture 17 Reasoning over Time Chapter 15.
CS Statistical Machine learning Lecture 24
The famous “sprinkler” example (J. Pearl, Probabilistic Reasoning in Intelligent Systems, 1988)
CPSC 422, Lecture 15Slide 1 Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 15 Oct, 14, 2015.
Probabilistic reasoning over time Ch. 15, 17. Probabilistic reasoning over time So far, we’ve mostly dealt with episodic environments –Exceptions: games.
CPS 170: Artificial Intelligence Markov processes and Hidden Markov Models (HMMs) Instructor: Vincent Conitzer.
Probability and Time. Overview  Modelling Evolving Worlds with Dynamic Baysian Networks  Simplifying Assumptions Stationary Processes, Markov Assumption.
Statistical Models for Automatic Speech Recognition Lukáš Burget.
Probability and Time. Overview  Modelling Evolving Worlds with Dynamic Baysian Networks  Simplifying Assumptions Stationary Processes, Markov Assumption.
EEL 6586: AUTOMATIC SPEECH PROCESSING Hidden Markov Model Lecture Mark D. Skowronski Computational Neuro-Engineering Lab University of Florida March 31,
Automated Speach Recognotion Automated Speach Recognition By: Amichai Painsky.
CPSC 7373: Artificial Intelligence Lecture 12: Hidden Markov Models and Filters Jiang Bian, Fall 2012 University of Arkansas at Little Rock.
Hidden Markov Models: Probabilistic Reasoning Over Time Natural Language Processing CMSC February 22, 2005.
Instructor: Eyal Amir Grad TAs: Wen Pu, Yonatan Bisk Undergrad TAs: Sam Johnson, Nikhil Johri CS 440 / ECE 448 Introduction to Artificial Intelligence.
CS 541: Artificial Intelligence Lecture VIII: Temporal Probability Models.
Hidden Markov Models Wassnaa AL-mawee Western Michigan University Department of Computer Science CS6800 Adv. Theory of Computation Prof. Elise De Doncker.
EEL 6586: AUTOMATIC SPEECH PROCESSING Hidden Markov Model Lecture
Hidden Markov Models (HMM)
Statistical Models for Automatic Speech Recognition
Instructor: Vincent Conitzer
An INTRODUCTION TO HIDDEN MARKOV MODEL
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 15
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 14
Probabilistic Reasoning over Time
Probabilistic Reasoning over Time
CS 188: Artificial Intelligence Spring 2007
CS 188: Artificial Intelligence Fall 2008
Statistical Models for Automatic Speech Recognition
Instructors: Fei Fang (This Lecture) and Dave Touretzky
Hidden Markov Model LR Rabiner
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 14
Speech recognition, machine learning
Instructor: Vincent Conitzer
CSCI 5582 Artificial Intelligence
Instructor: Vincent Conitzer
Speech recognition, machine learning
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 14
Presentation transcript:

Hidden Markov Models Dave DeBarr

Overview General Characteristics Simple Example Speech Recognition

Andrei Markov Russian statistician (1856 – 1922) Studied temporal probability models Markov assumption –State t depends only on a bounded subset of State 0:t-1 First-order Markov process –P(State t | State 0:t-1 ) = P(State t | State t-1 ) Second-order Markov process –P(State t | State 0:t-1 ) = P(State t | State t-2:t-1 )

Hidden Markov Model (HMM) Evidence can be observed, but the state is hidden Three components –Priors (initial state probabilities) –State transition model –Evidence observation model Changes are assumed to be caused by a stationary process –The transition and observation models do not change

Simple HMM Security guard resides in underground facility (with no way to see if it is raining) Wants to determine the probability of rain given whether the director brings an umbrella P(Rain 0 = t) = 0.50

What can you do with an HMM? Filtering –P(State t | Evidence 1:t ) Prediction –P(State t+k | Evidence 1:t ) Smoothing –P(State k | Evidence 1:t ) Most likely explanation –argmax State 1:t P(State 1:t | Evidence 1:t )

Filtering (the forward algorithm) P(Rain 1 = t) = Σ Rain 0 P(Rain 1 = t | Rain 0 ) P(Rain 0 ) =0.70 * * 0.50 = 0.50 P(Rain 1 = t | Umbrella 1 = t) = α P(Umbrella 1 = t | Rain 1 = t) P(Rain 1 = t) = α * 0.90 * 0.50 = α *0.45 ≈ P(Rain 2 = t | Umbrella 1 = t) = Σ Rain 1 P(Rain 2 = t | Rain 1 ) P(Rain 1 | Umbrella 1 = t) = 0.70 * * ≈ P(Rain 2 = t | Umbrella 1 = t, Umbrella 2 = t) = α P(Umbrella 2 = t | Rain 2 = t) P(Rain 2 = t | Umbrella 1 = t) = α * 0.90 * ≈ α * ≈ 0.883

Smoothing (the forward-backward algorithm) P(Umbrella 2 = t | Rain 1 = t) = Σ Rain 2 P(Umbrella 2 = t | Rain 2 ) P(* | Rain 2 ) P(Rain 2 | Rain 1 = t) = 0.9 * 1.0 * * 1.0 * 0.3 = 0.69 P(Rain 1 = t | Umbrella 1 = t, Umbrella 2 = t) = α * * 0.69 ≈ α * 0.56 ≈ 0.883

Most Likely Explanation (the Viterbi algorithm) P(Rain 1 = t, Rain 2 = t | Umbrella 1 = t, Umbrella 2 = t) = P(Umbrella 1 = t | Rain 1 = t) * P(Rain 2 = t | Rain 1 = t) * P (Umbrella 2 = t | Rain 2 = t) = * 0.70 * 0.90 ≈ 0.515

Speech Recognition (signal preprocessing)

Speech Recognition (models) P(Words | Signal) = α P(Signal | Words) P(Words) Decomposes into an acoustic model and a language model –Ceiling or Sealing –High ceiling or High sealing A state in a continuous speech HMM may be labeled with a phone, a phone state, and a word

Speech Recognition (phones) Human languages use a limited repertoire of sounds

Speech Recognition (phone model) Acoustic signal for [t] –Silent beginning –Small explosion in the middle –(Usually) Hissing at the end

Speech Recognition (pronounciation model) Coarticulation and dialect variations

Speech Recognition (language model) Can be as simple as bigrams P(Word i | Word 1:i-1 ) = P(Word i | Word i-1 )

References Artificial Intelligence: A Modern Approach –Second Edition (2003) –Stuart Russell & Peter Norvig Hidden Markov Model Toolkit (HTK) – –Nice tutorial (from data prep to evaluation)