Download presentation
Presentation is loading. Please wait.
1
Hidden Markov chain models (state space model)
General problem: have a set of noisy measurements of a process. I.e. two components: A real-world process. Measurements that give some information about the state of the process at a given time. Assumptions: Process is Markovian (previous state summarizes all relevant history). Observational noise is independent. This is the data weβve got. Z1 Z2 Z3 Zk Zk+1 Zn This is what weβre interested in. (The process) X1 . . . . . . X2 X3 Xk Xk+1 Xn
2
Hidden Markov model specification
We have two components in a hidden Markov chain model: A hidden process which is Markovian: π π 1 , π 2 , π 3 ,β―, π πβ1 , π π ,β―, π π =π( π 1 ) π=2 π π( π π | π πβ1 ) For each time point, what is needed is the dependency of the process state on the previous state: π( π π | π πβ1 ). This distribution will contain all process-related parameters and is called the system equation (Dependency on process parameters have been suppressed.) The dependency of each measurement on the process state at the given time: π( π π | π π ). This is called the observational equation and contains all the observation-related parameters. PS: Both process and measurements can be vectors. The number of components in the measurements can vary and do not need to match the number of components in the process! This is the data weβve got. Z1 Z2 Z3 Zk-1 Zk Zn X1 . . . . . . X2 X3 Xk-1 Xk Xn This is the process weβre interested in.
3
Hidden Markov model specification
What we want from the process: Inference about the state of the process at a given time The likelihood! If we have the likelihood, we can do inference on the parameters of the process (and the observations), as well as do model selection! πΏ π β‘π( π 1 , π 2 , π 3 ,β―, π πβ1 , π π ,β―, π π )= π( π 1 )π( π 2 |π 1 )π( π 3 |π 1 ,π 2 )β― π( π π |π 1 ,π 2 , π 3 ,β―, π πβ1 ) β― π( π π |π 1 ,π 2 , π 3 ,β―, π πβ1 ) PS: This is completely general! This is the data weβve got. Z1 Z2 Z3 Zk-1 Zk Zn X1 . . . . . . X2 X3 Xk-1 Xk Xn This is the process weβre interested in.
4
Doing inference on hidden times series models - filtering
Data Z1 Z2 Z3 Zk-1 Zk Zn X1 . . . . . . X2 X3 Xk-1 Xk Xn Process Procedure: Go stepwise through the observations from 1 to n. For observation k-1, assume you have the state inference conditioned on all data up to that observation: f(Xk-1|Zk-1,β¦,Z1). Then: Predict next system state: Use the state equation and the law of total probability calculate f(Xk|Zk-1,β¦,Z1). Predict next observation: Use the observational equation and the law of total probability to get f(Zk|Zk-1,β¦,Z1). Update process state with current observation: Use Bayesβ formula to calculate f(Xk|Zk, Zk-1, β¦,Z1). (You need f(Xk|Zk-1,β¦,Z1). , f(Zk|Zk-1,β¦,Z1) and the observational equation π( π π | π π ) to do this). Go back to step 1, now for Xk+1. Likelihood: LοΊf(Z1,β¦, Zn|ο±)=f(Z1) f(Z2|Z1)β¦f(Zk|Zk-1)... f(Zn|Zn-1).
5
The Kalman filter Steps i and ii can usually not be performed analytically for continuous states and measurements! However, if the system and observational equations are both linear and normal, this can be done! System equation: π π = πΉ π π πβ1 + π’ π + π π , where π π ~π 0 , π π Observational equation: π π = π» π π π + πΏ π , where πΏ π ~π 0 , π
π We only need to keep track of the mean and the variance, since everything is normal. Let π₯ π|πβ1 and π₯ π|π the mean of the distribution of π π given all measurements up to k-1 and k, respectively. Similarly π π|πβ1 and π π|π denotes the variance of π π given all measurements up to k-1 and k, respectively. Let π§ π|πβ1 and π π|πβ1 denoted the mean and variance of π π given all measurement up to k-1. Predict system: π₯ π|πβ1 = πΉ π π₯ πβ1|πβ1 + π’ π , π π|πβ1 = πΉ π π πβ1|πβ1 πΉβ² π + π π Predict observation: π§ π|πβ1 = π» π π₯ π|πβ1 , π π|πβ1 = π» π π π|πβ1 π»β² π + π
π Update with current observation: π₯ π|π = π₯ π|πβ1 + π π|πβ1 π»β² π π π|πβ1 β1 ( π π β π§ π|πβ1 ), π π|π =(πΌβ π π|πβ1 π»β² π π π|πβ1 β1 π» π ) π π|πβ1 Go back to step 1, now for Xk+1. πΏ= π=1 π π π ( π π |ππππ= π» π π₯ π|πβ1 , π£ππ= π π|πβ1 )
6
Kalman smoother We might also be interested in the inferring the state given *all* the measurements. This can also be done analytically for a linear normal system, using the Kalman smoother. Start with π₯ π|π and π π|π , i.e. inference for the last state given all measurements. Run backwards, using Bayes equation to incorporate the extra measurements step by step: Let πΆ π β‘ π π|π πΉ β² π π π+1|π β1 π₯ π|π = π₯ π|π + πΆ π π₯ π+1|π β π’ π+1 β πΉ π+1 π₯ π|π π π|π = π π|π + πΆ π ( π π+1|π β π π+1|π ) πΆβ² π Go from k to k-1. Typically the uncertainty of this state inference will Β«bubble upΒ» where you have bad/no observations and narrow down where you have good/many observations.
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.