Presentation is loading. Please wait.

Presentation is loading. Please wait.

Hidden Markov Models (HMMs) Chapter 3 (Duda et al.) – Section 3.10 (Warning: this section has lots of typos) CS479/679 Pattern Recognition Spring 2013.

Similar presentations


Presentation on theme: "Hidden Markov Models (HMMs) Chapter 3 (Duda et al.) – Section 3.10 (Warning: this section has lots of typos) CS479/679 Pattern Recognition Spring 2013."— Presentation transcript:

1 Hidden Markov Models (HMMs) Chapter 3 (Duda et al.) – Section 3.10 (Warning: this section has lots of typos) CS479/679 Pattern Recognition Spring 2013 – Dr. George Bebis

2 Sequential vs Temporal Patterns Sequential patterns: – The order of data points is irrelevant. Temporal patterns: – The order of data points is important (i.e., time series). – Data can be represented by a number of states. – States at time t are influenced directly by states in previous time steps (i.e., correlated).

3 Hidden Markov Models (HMMs) HMMs are appropriate for problems that have an inherent temporality. – Speech recognition – Gesture recognition – Human activity recognition

4 First-Order Markov Models Represented by a graph where every node corresponds to a state ω i. The graph can be fully-connected with self-loops.

5 First-Order Markov Models (cont’d) Links between nodes ω i and ω j are associated with a transition probability: P(ω(t+1)=ω j / ω(t)=ω i )=α ij which is the probability of going to state ω j at time t+1 given that the state at time t was ω i (first-order model).

6 First-Order Markov Models (cont’d) Markov models are fully described by their transition probabilities α ij The following constraints should be satisfied:

7 Example: Weather Prediction Model Assume three weather states: – ω 1 : Precipitation (rain, snow, hail, etc.) – ω 2 : Cloudy – ω 3 : Sunny Transition Matrix ω 1 ω 2 ω 3 ω1ω1ω2ω2ω3ω3ω1ω1ω2ω2ω3ω3 ω1ω1ω1ω1 ω2ω2ω2ω2 ω3ω3ω3ω3

8 Computing the probability P(ω T ) of a sequence of states ω T Given a sequence of states ω T =(ω(1), ω(2),..., ω(T)), the probability that the model generated ω T is equal to the product of the corresponding transition probabilities: where P(ω(1)/ ω(0))=P(ω(1)) is the prior probability of the first state.

9 Example: Weather Prediction Model (cont’d) What is the probability that the weather for eight consecutive days is: “sunny-sunny-sunny-rainy-rainy-sunny-cloudy-sunny” ? ω 8 =ω 3 ω 3 ω 3 ω 1 ω 1 ω 3 ω 2 ω 3 P(ω 8 )=P(ω 3 )P(ω 3 /ω 3 )P(ω 3 /ω 3 ) P(ω 1 /ω 3 ) P(ω 1 /ω 1 ) P( ω 3 /ω 1 ) P(ω 2 /ω 3 )P(ω 3 /ω 2 )=1.536 x 10 -4

10 Limitations of Markov models In Markov models, each state is uniquely associated with an observable event. Once an observation is made, the state of the system is trivially retrieved. Such systems are not of practical use for most applications.

11 Hidden States and Observations Assume that each state can generate a number of outputs (i.e., observations) according to some probability distribution. Each observation can potentially be generated at any state. State sequence is not directly observable (i.e., hidden) but can be approximated from observation sequence.

12 First-order HMMs Augment Markov model such that when it is in state ω(t) it also emits some symbol v(t) (visible state) among a set of possible symbols. We have access to the visible states v(t) only, while ω(t) are unobservable.

13 Example: Weather Prediction Model (cont’d) v 1 : temperature v 2 : humidity etc. Observations:

14 Observation Probabilities When the model is in state ω j at time t, the probability of emitting a visible state v k at that time is denoted as: P(v(t)=v k / ω(t)= ω j )=b jk where (observation probabilities) For every sequence of hidden states, there is an associated sequence of visible states: ω T =(ω(1), ω(2),..., ω(T))  V T =(v(1), v(2),..., v(T))

15 Absorbing State ω 0 Given a state sequence and its corresponding observation sequence: ω T =(ω(1), ω(2),..., ω(T))  V T =(v(1), v(2),..., v(T)) we assume that ω(T)=ω 0 is some absorbing state, which uniquely emits symbol v(T)=v 0 Once entering the absorbing state, the system can not escape from it.

16 HMM Formalism An HMM is defined by {Ω, V,  – Ω : {ω 1 … ω n } are the possible states – V : {v 1 …v m } are the possible observations –  i  are the prior state probabilities – A = {a ij } are the state transition probabilities – B = {b ik } are the observation state probabilities

17 Some Terminology Causal: the probabilities depend only upon previous states. Ergodic: Given some starting state, every one of the states has a non-zero probability of occurring. “left-right” HMM

18 Coin toss example You are in a room with a barrier (e.g., a curtain) through which you cannot see what is happening on the other side. On the other side of the barrier is another person who is performing a coin (or multiple coin) toss experiment. The other person will tell you only the result of the experiment, not how he obtained that result. e.g., V T =HHTHTTHH...T=v(1),v(2),..., v(T)

19 Coin toss example (cont’d) Problem: derive an HMM model to explain the observed sequence of heads and tails. – The coins represent the hidden states since we do not know which coin was tossed each time. – The outcome of each toss represents an observation. – A “likely” sequence of coins (state sequence) may be inferred from the observations. – The state sequence might not be unique in general.

20 Coin toss example: 1-fair coin model There are 2 states, each associated with either heads (state1) or tails (state2). Observation sequence uniquely defines the states (i.e., states are not hidden). observation probabilities

21 Coin toss example: 2-fair coins model There are 2 states, each associated with a coin; a third coin is used to decide which of the fair coins to flip. Neither state is uniquely associated with either heads or tails. observation probabilities

22 Coin toss example: 2-biased coins model There are 2 states, each associated with a biased coin; a third coin is used to decide which of the biased coins to flip. Neither state is uniquely associated with either heads or tails. observation probabilities

23 Coin toss example:3-biased coins model There are 3 states, each state associated with a biased coin; we decide which coin to flip using some way (e.g., other coins). Neither state is uniquely associated with either heads or tails. observation probabilities

24 Which model is best? Since the states are not observable, the best we can do is to select the model θ that best explains the observations: max θ P(V T / θ) Long observation sequences are typically better in selecting the best model.

25 Classification Using HMMs Given an observation sequence V T and set of possible models θ, choose the model with the highest probability P(θ / V T ). Bayes rule:

26 Three basic HMM problems Evaluation – Determine the probability P(V T ) that a particular sequence of visible states V T was generated by a given model (i.e., Forward/Backward algorithm). Decoding – Given a sequence of visible states V T, determine the most likely sequence of hidden states ω T that led to those observations (i.e., using Viterbi algorithm). Learning – Given a set of visible observations, determine a ij and b jk (i.e., using EM algorithm - Baum-Welch algorithm).

27 Evaluation The probability that a model produces V T can be computed using the theorem of total probability: where ω r T =(ω(1), ω(2),..., ω(T)) is a possible state sequence and r max is the max number of state sequences. For a model with c states ω 1, ω 2,..., ω c, r max =c T

28 Evaluation (cont’d) We can rewrite each term as follows: Combining the two equations we have:

29 Evaluation (cont’d) Given a ij and b jk, it is straightforward to compute P(V T ). What is the computational complexity? O(T r max )= O(T c T )

30 Recursive computation of P(V T ) (HMM Forward) v(T) v(1)v(t) v(t+1) ω(1)ω(t)ω(t+1)ω(T) ωiωiωiωi ωjωjωjωj...

31 Recursive computation of P(V T ) (HMM Forward) (cont’d) or using marginalization:

32 Recursive computation of P(V T ) (HMM Forward) (cont’d) ω0ω0ω0ω0

33 (i.e., corresponds to state ω(T)=ω 0 ) for j=1 to c do What is the computational complexity in this case? O(T c 2 ) (if t=T, j=0)

34 Example ω 0 ω 1 ω 2 ω 3 ω 0 ω 1 ω 2 ω 3 ω0ω0ω1ω1ω2ω2ω3ω3ω0ω0ω1ω1ω2ω2ω3ω3 ω0ω0ω1ω1ω2ω2ω3ω3ω0ω0ω1ω1ω2ω2ω3ω3

35 Example (cont’d) Similarly for t=2,3,4 Finally: V T =v 1 v 3 v 2 v 0 0.2 0.2 0.8 initial state

36 Recursive computation of P(V T ) (HMM backward) v(1) ω(1) ω(t)ω(t+1)ω(T) v(t)v(t+1)v(T)... ωiωiωiωi ωjωjωjωj β j (t+1) /ω (t+1)=ω j ) β i (t) i ωiωiωiωi

37 Recursive computation of P(V T ) (HMM backward) (cont’d) =ω j )) or i v(1) ω(1) ω(t)ω(t+1)ω(T) v(t)v(t+1)v(T) ωiωiωiωi ωjωjωjωj

38 Recursive computation of P(V T ) (HMM backward) (cont’d)

39 Decoding Find the most probable sequence of hidden states. Use an optimality criterion - different optimality criteria lead to different solutions. Algorithm 1: choose the states ω(t) which are individually most likely.

40 Decoding – Algorithm 1

41 Decoding (cont’d) Algorithm 2: at each time step t, find the state that has the highest probability α i (t) (i.e., use forward algorithm with minor changes).

42 Decoding – Algorithm 2

43 Decoding – Algorithm 2 (cont’d)

44 There is no guarantee that the path is a valid one. The path might imply a transition that is not allowed by the model. not allowed since ω 32 =0 0 1 2 3 4 0 1 2 3 4 Example:

45 Decoding (cont’d) Algorithm 3: find the single best sequence ω T by maximizing P(ω T /V T ) This is the most widely used algorithm known as Viterbi algorithm.

46 Decoding – Algorithm 3 maximize: P(ω T /V T )

47 Decoding – Algorithm 3 (cont’d) recursion (similar to Forward Algorithm, except that it uses maximization over previous states instead of summation\)

48 Learning Determine the transition and emission probabilities a ij and b jv from a set of training examples (i.e., observation sequences V 1 T, V 2 T,..., V n T ). There is no known way to find the ML solution analytically. – It would be easy if we knew the hidden states – Hidden variable problem  use EM algorithm!

49 Learning (cont’d) EM algorithm – Update a ij and b jk iteratively to better explain the observed training sequences. V: V 1 T, V 2 T,..., V n T Expectation step: p(ω T /V, θ) Maximization step: θ t+1 =argmax θ E[log p(ω T,V T / θ)/ V T, θ t ]

50 Learning (cont’d) Updating transition/emission probabilities:

51 Learning (cont’d) Define the probability of transitioning from ω i to ω j at step t given V T : Expectation step

52 Learning (cont’d) t-1 α i (t-1) β j (t) t a ij b jv(t)

53 Learning (cont’d) Maximization step

54 Learning (cont’d) Maximization step

55 Practical Problems How do we decide on the number of states and the structure of the model? – Use domain knowledge otherwise very hard problem! What about the size of observation sequence ? – Should be sufficiently long to guarantee that all state transitions will appear a sufficient number of times. – A large number of training data is necessary to learn the HMM parameters.


Download ppt "Hidden Markov Models (HMMs) Chapter 3 (Duda et al.) – Section 3.10 (Warning: this section has lots of typos) CS479/679 Pattern Recognition Spring 2013."

Similar presentations


Ads by Google