Lecture II-4: Filtering, Sequential Estimation, and the Ensemble Kalman Filter Lecture Outline: A Typical Filtering Problem Sequential Estimation Formal.

Slides:



Advertisements
Similar presentations
Introduction to Data Assimilation NCEO Data-assimilation training days 5-7 July 2010 Peter Jan van Leeuwen Data Assimilation Research Center (DARC) University.
Advertisements

1 12. Principles of Parameter Estimation The purpose of this lecture is to illustrate the usefulness of the various concepts introduced and studied in.
Chapter 7 Title and Outline 1 7 Sampling Distributions and Point Estimation of Parameters 7-1 Point Estimation 7-2 Sampling Distributions and the Central.
Lecture 3 Probability and Measurement Error, Part 2.
The structure and evolution of stars
STAT 497 APPLIED TIME SERIES ANALYSIS
Observers and Kalman Filters
Sérgio Pequito Phd Student
Advanced data assimilation methods- EKF and EnKF Hong Li and Eugenia Kalnay University of Maryland July 2006.
Estimation and the Kalman Filter David Johnson. The Mean of a Discrete Distribution “I have more legs than average”
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
Lecture II-2: Probability Review
Principles of the Global Positioning System Lecture 10 Prof. Thomas Herring Room A;
Slam is a State Estimation Problem. Predicted belief corrected belief.
Principles of the Global Positioning System Lecture 11 Prof. Thomas Herring Room A;
Binary Variables (1) Coin flipping: heads=1, tails=0 Bernoulli Distribution.
Regression and Correlation Methods Judy Zhong Ph.D.
Colorado Center for Astrodynamics Research The University of Colorado STATISTICAL ORBIT DETERMINATION Project Report Unscented kalman Filter Information.
ECE 8443 – Pattern Recognition LECTURE 06: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Bias in ML Estimates Bayesian Estimation Example Resources:
Lecture II-1: Data Assimilation Overview Lecture Outline: Objectives and methods of data assimilation Definitions and terminology Examples State-space.
880.P20 Winter 2006 Richard Kass 1 Confidence Intervals and Upper Limits Confidence intervals (CI) are related to confidence limits (CL). To calculate.
Applications of Bayesian sensitivity and uncertainty analysis to the statistical analysis of computer simulators for carbon dynamics Marc Kennedy Clive.
Computer vision: models, learning and inference Chapter 19 Temporal models.
1 Institute of Engineering Mechanics Leopold-Franzens University Innsbruck, Austria, EU H.J. Pradlwarter and G.I. Schuëller Confidence.
Computer Vision Group Prof. Daniel Cremers Autonomous Navigation for Flying Robots Lecture 6.2: Kalman Filter Jürgen Sturm Technische Universität München.
CSDA Conference, Limassol, 2005 University of Medicine and Pharmacy “Gr. T. Popa” Iasi Department of Mathematics and Informatics Gabriel Dimitriu University.
Kalman Filter (Thu) Joon Shik Kim Computational Models of Intelligence.
Assimilation of HF Radar Data into Coastal Wave Models NERC-funded PhD work also supervised by Clive W Anderson (University of Sheffield) Judith Wolf (Proudman.
Module 1: Statistical Issues in Micro simulation Paul Sousa.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Deterministic vs. Random Maximum A Posteriori Maximum Likelihood Minimum.
Modern Navigation Thomas Herring
Overview Particle filtering is a sequential Monte Carlo methodology in which the relevant probability distributions are iteratively estimated using the.
ECE 8443 – Pattern Recognition LECTURE 07: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Class-Conditional Density The Multivariate Case General.
A unifying framework for hybrid data-assimilation schemes Peter Jan van Leeuwen Data Assimilation Research Center (DARC) National Centre for Earth Observation.
Research Vignette: The TransCom3 Time-Dependent Global CO 2 Flux Inversion … and More David F. Baker NCAR 12 July 2007 David F. Baker NCAR 12 July 2007.
Data assimilation and forecasting the weather (!) Eugenia Kalnay and many friends University of Maryland.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 07: BAYESIAN ESTIMATION (Cont.) Objectives:
What is Data Assimilation ? Data Assimilation: Data assimilation seeks to characterize the true state of an environmental system by combining information.
ECE-7000: Nonlinear Dynamical Systems Overfitting and model costs Overfitting  The more free parameters a model has, the better it can be adapted.
An Introduction to Kalman Filtering by Arthur Pece
An Introduction To The Kalman Filter By, Santhosh Kumar.
STOCHASTIC HYDROLOGY Stochastic Simulation of Bivariate Distributions Professor Ke-Sheng Cheng Department of Bioenvironmental Systems Engineering National.
BASIC STATISTICAL CONCEPTS Statistical Moments & Probability Density Functions Ocean is not “stationary” “Stationary” - statistical properties remain constant.
SLAM Tutorial (Part I) Marios Xanthidis.
Hydrologic Data Assimilation with a Representer-Based Variational Algorithm Dennis McLaughlin, Parsons Lab., Civil & Environmental Engineering, MIT Dara.
Lecture II-3: Interpolation and Variational Methods Lecture Outline: The Interpolation Problem, Estimation Options Regression Methods –Linear –Nonlinear.
The Unscented Particle Filter 2000/09/29 이 시은. Introduction Filtering –estimate the states(parameters or hidden variable) as a set of observations becomes.
Stochastic Hydrology Random Field Simulation Professor Ke-Sheng Cheng Department of Bioenvironmental Systems Engineering National Taiwan University.
A Random Subgrouping Scheme for Ensemble Kalman Filters Yun Liu Dept. of Atmospheric and Oceanic Science, University of Maryland Atmospheric and oceanic.
Colorado Center for Astrodynamics Research The University of Colorado 1 STATISTICAL ORBIT DETERMINATION Kalman Filter with Process Noise Gauss- Markov.
Regularization of energy-based representations Minimize total energy E p (u) + (1- )E d (u,d) E p (u) : Stabilizing function - a smoothness constraint.
Copyright © Cengage Learning. All rights reserved. 5 Joint Probability Distributions and Random Samples.
G. Cowan Lectures on Statistical Data Analysis Lecture 9 page 1 Statistical Data Analysis: Lecture 9 1Probability, Bayes’ theorem 2Random variables and.
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
Computacion Inteligente Least-Square Methods for System Identification.
Probabilistic Robotics Bayes Filter Implementations Gaussian filters.
Multiple Random Variables and Joint Distributions
Chapter 7. Classification and Prediction
Data Assimilation Research Testbed Tutorial
STATISTICAL ORBIT DETERMINATION Kalman (sequential) filter
12. Principles of Parameter Estimation
Sampling Distributions and Estimation
Data Assimilation Theory CTCD Data Assimilation Workshop Nov 2005
PSG College of Technology
Course: Autonomous Machine Learning
Stochastic Hydrology Random Field Simulation
Filtering and State Estimation: Basic Concepts
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
12. Principles of Parameter Estimation
Kalman Filter: Bayes Interpretation
Presentation transcript:

Lecture II-4: Filtering, Sequential Estimation, and the Ensemble Kalman Filter Lecture Outline: A Typical Filtering Problem Sequential Estimation Formal Bayesian Solution to the Nonlinear Filtering Problem The Linear Case -- Classical Kalman Filtering Practical Solution of the Nonlinear problem – Ensemble Kalman Filtering SGP97 Case Study

A Filtering Problem -- Soil Moisture Problem is to characterize time-varying soil saturation at nodes on a discrete grid. Estimates rely on scattered meteorological measurements and on microwave remote sensing observations (brightness temperature). Estimate must be compatible with relevant mass and energy conservation equations. State and meas. equations are nonlinear. Time t i-1 Time t i Time t i+1 Meteorological station Satellite footprint Surface soil saturation contour State Eq. (Land surface model) Meas Eq. (Radiative transfer model)

Limitations of the Variational Approach for Filtering Applications Variational methods provide an effective way to estimate the mode of f y|z (y|z) for interpolation and smoothing problems. However, they have the following limitations: The adjoint equations for a complex model may be difficult to derive, especially if the model includes discontinuous functions. Most variational methods rely on gradient-based search algorithms which may have trouble converging. Variational data assimilation algorithms tend to be complex, fragile, and difficult to implement, particularly for large time-dependent problems. Variational algorithms generally provide no information on the likely range of system states around the conditional mode. Most variational methods are unable to account for time-varying model input errors when problem size is large. Considering these limitations, it seems reasonable examine other alternatives for the filtering problem. One option is sequential estimation.

Sequential Estimation For filtering problems, where the emphasis is on characterizing current rather than historical conditions, it is natural to divide the data assimilation process into discrete stages which span measurement times. With these defintions we can define two distinct types of estimation steps: To illustrate, suppose that the measurements available at all times through the current time t i are assembled in a large array Z i : Z i = [ z 1, z 2, …, z i ] = Set of all measurements through time t i Then the conditional PDF of y(t i ), given all measurements through t i, is f yi|Zi [ y(t i )|Z i ] and the conditional PDF of the state at some later time (e.g. t i+1 ) given the same measurement information is f y,i+1|Zi [ y(t i+1 )|Z i ]. A propagation step -- Describes how the conditional PDF changes between two adjacent measurement times (e.g. between t i and t i+1 ). An update step -- Describes how the conditional PDF changes at a given time (e.g. t i+1 ) when new measurements are incorporated. These steps are carried out sequentially, starting from the initial time between t 0 through the final time t I.

Propagating and Updating Conditional PDFs The propagation and update steps of the sequential estimation process are related as follows: zizi Z i = [Z i-1, z i ] t0t0 z1z1 Z 1 = [z 1 ] t1t1 z2z2 Z 2 = [Z 1, z 2 ] t2t2 titi t i+1 Algorithm initialized with unconditional (prior) PDF at t 0 f yi| zi-1 [ y i |Z i-1 ]f y,i+1| zi [ y i+1 |Z i ] f yi| zi [ y i |Z i ] Update i Propagation i to i +1 f y0 [ y 0 ] f y1 [ y 1 ] f y1| z1 [ y 1 |Z 1 ] Propagation 1 to 2 f y2| z1 [ y 2 |Z 1 ] Update 1 Propagation 0 to 1 Meas. i Meas. 1Meas. 2

Formal Solution to the Nonlinear Bayesian Filtering Problem Bayes theorem provides the basis for deriving the PDF transformations required during the propgation and update steps of the sequential filtering procedure. For example, when the random input and the measurement error are additive and independent from time to time the transformation from f yi| Zi [ y i |Z i ] to f yi+1| Zi [ y i+1 |Z i ] is: Propagation from t i to t i+1 : The transformation from f yi+1| Zi [ y i+1 |Z i ] to f yi+1| Zi+1 [ y i+1 |Z i+1 ] is: Update at t i+1 : These expressions constitute a formal sequential solution to the nonlinear filtering problem since everything on the right-hand side of each equation is known from the previous step. Although it is generally not possible to evaluate the required PDFs these Bayesian equations provide the basis for a number of practical approximate solutions to the filtering problem.

The Linear Bayesian Filtering Problem If u i, y 0, and  i are independent multivariate normal random variables then the conditional PDFs appearing in the Bayesian propagation and update equations are all multivariate normal. These PDFs are completely defined by their means and covariances, which can be derived from the general expression for the moments of a conditional multivariate normal PDF. When the state and measurement equations are linear and the input and measurement error PDFs are multivariate normal the Bayesian propagation and update expressions yield a convenient sequential algorithm known as the Kalman filter. Suppose that the state and measurement equations have the following forms: random For simplicity, assume a measurement is taken at every discretized time

The Classical Kalman Filter The conditional mean and covariance given by the Bayesian filter for the linear multivariate case are: Propagation from t i to t i+1 : Update at t i+1 : Note that the expression for the updated conditional mean (the estimate ) consists of two parts: A forecast E ( y i+1 |Z i ) based on data collected through t i A correction term which depends on the difference between the forecast and the new measurement z i taken at t i The Kalman gain K i+1 determines the weight given to the correction term. forecastcorrection Kalman gain

Kalman Filter Example - 1 A simple example illustrates the forecast - correction tradeoff at the heart of the Kalman filter. Consider the following scalar state and measurement equations: In this special case the only input is a random initial condition with a zero mean and a specified covariance (actually a scalar variance) C yy,0. The state remains fixed at this initial value but is observed with a set of noisy measurements. The filter equations are: Note that the filter tends to ignore the measurements ( K i+1  0 ) when the measurement error variance is large and it tends to ignore the forecast ( K i+1  1 ) when this variance is small. random Propagation: Update: Kalman Gain

Kalman Filter Example - 2 In this example the covariance and Kalman gain both decrease over time and the updated estimate converges to the true value of the unknown initial condition. The Kalman gain is smaller and the convergence is slower when the measurement error variance is larger. Updated estimate True value Kalman gain 1/t (Sample mean) Time Note that the Kalman gain approaches 1/t for large time. This implies that the estimate of the unknown initial condition is nearly equal to the sample mean of the measurements for the conditions specified here (variance of both initial condition and measurement error = 1.0).

Practical Solution of the Nonlinear Bayesian Filtering Problem One option is to use ensemble (or Monte Carlo) methods to approximate the probabilistic information conveyed by the conditional PDFs f yi| Zi [ y i |Z i ] and f yi+1| Zi [ y i+1 |Z i ]. The basic concept is to generate populations of random inputs, measurement errors, and states which have the desired PDFs. This is easiest to illustrate within the two step sequential structure adopted earlier: The general Bayesian nonlinear filtering algorithm is not practical for large problems. The Kalman filter is intended for linear problems but can be applied to weakly nonlinear problems if the state and measurement equations are linearized about the most recent updated estimate. Unfortunately, the resulting extended Kalman filter tends to be unstable. In order to solve the large nonlinear problems encountered in data assimilation we need to take a different approach. Propagation from t i to t i+1 : Obtain L random replicates of the state y i from the update at t i. Generate L random replicates of the input vector u i during the interval [ t i, t i+1 ). Solve the state equation for each replicate to construct a population of L random replicates for the state y i+1 at the end of the interval. Update at t i+1 : Generate L random replicates of the measurement error  i+1 at the new measurement time t i+1. Then use Bayes theorem (or a suitable approximation) to update each replicate of the state y i+ 1. These replicates serve as the initial conditions for the next propagation step.

Ensemble filtering Propagation of conditional probability density (formal Bayesian filtering): Evolution of random replicates in ensemble (Ensemble filtering): y l (t i+1 | Z i+1 ) y l (t i | Z i ) Time Propagate forward in time titi t i+1 Update with new measurement y l (t i+1 | Z i ) Propagate forward in time p(yi|Zi)p(yi|Zi) p(y i+1 |Z i ) Update with new measurement (z i+1 ) t i+1 Time titi p(y i+1 |Z i+1 ) Ensemble filtering propagates only replicates (no PDFs). But how should update be performed? It is not practical to construct complete multivariate PDF and update with Bayes theorem.

The Ensemble Kalman Filter The updating problem simplifies greatly if we assume p[y( t i+1 )| Z i+1 ] is Gaussian. Then update for replicate l is: K i+1 = Kalman gain derived from propagated ensemble sample covariance Cov[y(t i+1 )| Z i ] and  l i is a random measurement error generated in the filter. After each replicate is updated it is propagated to next measurement time. There is no need to update the sample covariance. This is the ensemble Kalman filter (EnKF). Potential Pitfalls Appropriateness of the Kalman update for non-Gaussian density functions? Need to construct, store, and manipulate large covariance matrices (as in classical Kalman filter)

Ensemble Kalman Filtering Example - 1 Plot below shows ensemble Kalman filter trajectories for a simple two state time series model with perfect initial conditions, additive input error, and noisy measurements at every fifth time step: 20 replicates are shown in green Measurements are shown as magenta dots The true state (used to generate the measurements) is shown in black The mean of the replicates is shown in red Note how ensemble spreads over time until update, when replicates converge toward measurement Time State 1

Ensemble Kalman Filtering Example - 2 It is difficult and not very useful to try to estimate the complete multivariate conditional PDF of y from the ensemble. However, it is possible to estimate means, median, modes, covariances, marginal PDFs and other useful properties of the state. Plot below shows histogram (approximate marginal PDF) for state 1 of the time series example. This histogram is derived from a 200 replicate ensemble at t = State 1 Number of replicates

Case Study Area Aircraft microwave measurements SGP97 Experiment - Soil Moisture Campaign

Ensemble Kalman Filter Test – SGP97 Soil Moisture Problem “Measured” radiobrightness “True” radiobrightness “True” soil, canopy moisture and temperature Soil properties and land use Land surface model Mean initial conditions Mean land- atmosphere boundary fluxes Radiative transfer model Random input error Random initial condition error Random meas. error Ensemble Kalman Filter Estimated radiobrightness and soil moisture Soil properties and land use, mean fluxes and initial conditions, error covariances Estimation error Observing System Simulation Experiment (OSSE)

Synthetic Experiment (OSSE) based on SGP97 Field Campaign Synthetic experiment uses real soil, landcover, and precipitation data from SGP97 (Oklahoma). Radiobrightness measurements are generated from our land surface and radiative transfer models, with space/time correlated model error (process noise) and measurement error added. SGP97 study area, showing principal inputs to data assimilation algorithm:

Area-average top node saturation estimation error Normalized error for open-loop prediction (no microwave meas.) = 1.0 Compare jumps in EnKF estimates at measurement time to variational benchmark (smoothing solution). EnKF error generally increases between measurements. Increasing ensemble size

Error vs. Ensemble Size Error decreases faster up to ~500 replicates but then levels off. Does this reflect impact of non-Gaussian density (good covariance estimate is not sufficient)?

Actual and Expected Error top node saturation rms error, Ne = 500 [-] day of year actual error (rms) expected forecast error (ensemble-derived) expected analysis error (ensemble-derived) Ensemble Kalman filter consistently underestimates rms error, even when input statistics are specified perfectly. Non-Gaussian behavior?

Ensemble Error Distribution Errors appear to be more Gaussian at intermediate moisture values and more skewed at high or low values. Uncertainty is small just after a storm, grows with drydown, and decreases again when soil is very dry.

Model Error Estimates Ensemble Kalman filter provides discontinuous but generally reasonable estimates of model error but sample problem. Compare to smoothed error estimate from variational benchmark. Specified error statistics are perfect.

Summary Variational methods can work well for interpolation and smoothing problems but have conceptual and computational deficiencies that limit their applicability to filtering problems Sequential Bayesian filtering provides a formal solution to the nonlinear filtering problem but is not feasible to use for large problems. Classical Kalman filtering provides a good solution to linear multivariate normal problems of moderate size but it is not suitable for nonlinear problems. Ensemble Kalman filtering provides an efficient option for the solving large nonlinear filtering problems encountered in data assimilation applications. Ensemble propagation characterizes distribution of system states (e.g. soil moisture) while making relatively few assumptions. Approach accommodates very general descriptions of model error. Most ensemble filter updates are based on Gaussian assumption. Validity of this assumption is problem-dependent.