Introduction to Data Assimilation NCEO Data-assimilation training days 5-7 July 2010 Peter Jan van Leeuwen Data Assimilation Research Center (DARC) University.

Slides:



Advertisements
Similar presentations
Pattern Recognition and Machine Learning
Advertisements

1 -Classification: Internal Uncertainty in petroleum reservoirs.
Data-Assimilation Research Centre
Variational data assimilation and forecast error statistics
Mobile Robot Localization and Mapping using the Kalman Filter
Introduction to Data Assimilation Peter Jan van Leeuwen IMAU.
The Inverse Regional Ocean Modeling System:
Initialization Issues of Coupled Ocean-atmosphere Prediction System Climate and Environment System Research Center Seoul National University, Korea In-Sik.
The General Linear Model. The Simple Linear Model Linear Regression.
Aspects of Conditional Simulation and estimation of hydraulic conductivity in coastal aquifers" Luit Jan Slooten.
Visual Recognition Tutorial
Ibrahim Hoteit KAUST, CSIM, May 2010 Should we be using Data Assimilation to Combine Seismic Imaging and Reservoir Modeling? Earth Sciences and Engineering.
Sérgio Pequito Phd Student
Lecture 10: Support Vector Machines
Tracking with Linear Dynamic Models. Introduction Tracking is the problem of generating an inference about the motion of an object given a sequence of.
Lecture II-2: Probability Review
CSC2535: 2013 Advanced Machine Learning Lecture 3a: The Origin of Variational Bayes Geoffrey Hinton.
Principles of the Global Positioning System Lecture 10 Prof. Thomas Herring Room A;
Slam is a State Estimation Problem. Predicted belief corrected belief.
Principles of the Global Positioning System Lecture 11 Prof. Thomas Herring Room A;
Colorado Center for Astrodynamics Research The University of Colorado STATISTICAL ORBIT DETERMINATION Project Report Unscented kalman Filter Information.
Particle Filters in high dimensions Peter Jan van Leeuwen and Mel Ades Data-Assimilation Research Centre DARC University of Reading Lorentz Center 2011.
Nonlinear Data Assimilation and Particle Filters
Computing a posteriori covariance in variational DA I.Gejadze, F.-X. Le Dimet, V.Shutyaev.
CSDA Conference, Limassol, 2005 University of Medicine and Pharmacy “Gr. T. Popa” Iasi Department of Mathematics and Informatics Gabriel Dimitriu University.
Computer Vision - A Modern Approach Set: Tracking Slides by D.A. Forsyth The three main issues in tracking.
Linear Regression Andy Jacobson July 2006 Statistical Anecdotes: Do hospitals make you sick? Student’s story Etymology of “regression”
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 3: LINEAR MODELS FOR REGRESSION.
ECE 8443 – Pattern Recognition LECTURE 07: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Class-Conditional Density The Multivariate Case General.
Mean Field Variational Bayesian Data Assimilation EGU 2012, Vienna Michail Vrettas 1, Dan Cornford 1, Manfred Opper 2 1 NCRG, Computer Science, Aston University,
Dusanka Zupanski And Scott Denning Colorado State University Fort Collins, CO CMDL Workshop on Modeling and Data Analysis of Atmospheric CO.
SUPA Advanced Data Analysis Course, Jan 6th – 7th 2009 Advanced Data Analysis for the Physical Sciences Dr Martin Hendry Dept of Physics and Astronomy.
University of Colorado Boulder ASEN 5070: Statistical Orbit Determination I Fall 2014 Professor Brandon A. Jones Lecture 18: Minimum Variance Estimator.
2004 SIAM Annual Meeting Minisymposium on Data Assimilation and Predictability for Atmospheric and Oceanographic Modeling July 15, 2004, Portland, Oregon.
A unifying framework for hybrid data-assimilation schemes Peter Jan van Leeuwen Data Assimilation Research Center (DARC) National Centre for Earth Observation.
Applications of optimal control and EnKF to Flow Simulation and Modeling Florida State University, February, 2005, Tallahassee, Florida The Maximum.
University of Colorado Boulder ASEN 5070: Statistical Orbit Determination I Fall 2014 Professor Brandon A. Jones Lecture 26: Singular Value Decomposition.
MODEL ERROR ESTIMATION EMPLOYING DATA ASSIMILATION METHODOLOGIES Dusanka Zupanski Cooperative Institute for Research in the Atmosphere Colorado State University.
Research Vignette: The TransCom3 Time-Dependent Global CO 2 Flux Inversion … and More David F. Baker NCAR 12 July 2007 David F. Baker NCAR 12 July 2007.
Data assimilation and forecasting the weather (!) Eugenia Kalnay and many friends University of Maryland.
Geology 5670/6670 Inverse Theory 21 Jan 2015 © A.R. Lowry 2015 Read for Fri 23 Jan: Menke Ch 3 (39-68) Last time: Ordinary Least Squares Inversion Ordinary.
An Introduction to Kalman Filtering by Arthur Pece
Data Modeling Patrice Koehl Department of Biological Sciences National University of Singapore
Bayes Theorem The most likely value of x derived from this posterior pdf therefore represents our inverse solution. Our knowledge contained in is explicitly.
Short Introduction to Particle Filtering by Arthur Pece [ follows my Introduction to Kalman filtering ]
Prepared by Dusanka Zupanski and …… Maximum Likelihood Ensemble Filter: application to carbon problems.
Lecture II-3: Interpolation and Variational Methods Lecture Outline: The Interpolation Problem, Estimation Options Regression Methods –Linear –Nonlinear.
Geogg124: Data assimilation P. Lewis. What is Data Assimilation? Optimal merging of models and data Models Expression of current understanding about process.
Colorado Center for Astrodynamics Research The University of Colorado 1 STATISTICAL ORBIT DETERMINATION Kalman Filter with Process Noise Gauss- Markov.
G. Cowan Lectures on Statistical Data Analysis Lecture 9 page 1 Statistical Data Analysis: Lecture 9 1Probability, Bayes’ theorem 2Random variables and.
Future Directions in Ensemble DA for Hurricane Prediction Applications Jeff Anderson: NCAR Ryan Torn: SUNY Albany Thanks to Chris Snyder, Pavel Sakov The.
École Doctorale des Sciences de l'Environnement d’Île-de-France Année Universitaire Modélisation Numérique de l’Écoulement Atmosphérique et Assimilation.
École Doctorale des Sciences de l'Environnement d’ Î le-de-France Année Modélisation Numérique de l’Écoulement Atmosphérique et Assimilation.
The Ensemble Kalman filter
June 20, 2005Workshop on Chemical data assimilation and data needs Data Assimilation Methods Experience from operational meteorological assimilation John.
3. Linear Models for Regression 後半 東京大学大学院 学際情報学府 中川研究室 星野 綾子.
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 1: INTRODUCTION.
Hybrid Data Assimilation
Probability Theory and Parameter Estimation I
Data Assimilation Theory CTCD Data Assimilation Workshop Nov 2005
Ch3: Model Building through Regression
CH 5: Multivariate Methods
Introduction to particle filter
Filtering and State Estimation: Basic Concepts
Introduction to particle filter
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
Principles of the Global Positioning System Lecture 11
Kalman Filter: Bayes Interpretation
Sarah Dance DARC/University of Reading
Presentation transcript:

Introduction to Data Assimilation NCEO Data-assimilation training days 5-7 July 2010 Peter Jan van Leeuwen Data Assimilation Research Center (DARC) University of Reading

Basic estimation theory Observation: T 0 = T + e 0 First guess: T m = T + e m E{e 0 } = 0 E{e m } = 0 Assume a linear best estimate: T n = a T 0 + b T m with T n = T + e n. 1) Gives: E{e n } = E{T n -T} = E{aT 0 +bT m -T} = = E{ae 0 +be m + (a+b-1) T} = (a+b-1) T = 0 Hence b=1-a. Find a and b such that 1) E{e n } = 0 2) E{e n 2 } minimal E{e 0 2 } = s 0 2 E{e m 2 } = s m 2 E{e 0 e m } = 0

Basic estimation theory 2) E{e n 2 } minimal gives: E{e n 2 } = E{(T n –T) 2 } = E{(aT 0 + bT m – T) 2 } = = E{(ae 0 +be m ) 2 } = a 2 E{e 0 2 } + b 2 E{e m 2 } = = a 2 s (1-a) 2 s m 2 a = __________ and sm2sm2 s s m 2 This has to be minimal, so the derivative wrt a has to be zero: 2 a s (1-a) s m 2 = 0, so (s s m 2 )a – s m 2 = 0, hence: b = 1-a = __________ s02s02 s s m 2 s n 2 = E{e n 2 } = ______________ = ________ (s s m 2 ) 2 s m 4 s s 0 4 s m 2 s 0 2 s m 2 s s m 2

Solution: T n = _______ T 0 + _______ T m sm2sm2 s02s02 s s m 2 ___ = ___ + ___ 111 sm2sm2 s02s02 sn2sn2 and Note: s n smaller than s 0 and s m ! Basic estimation theory Best Linear Unbiased Estimate BLUE Just least squares!!!

Can we generalize this? More dimensions Nonlinear estimates (why linear?) Observations that are not directly modeled Biases (Is this reference to the truth necessary?)

P(u) u (m/s) The basics: probability density functions

The model pdf p (u(x1), u(x2), T(x3), … ) u(x1) u(x2) T(x3)

Observations In situ observations: irregular in space and time e.g. sparse hydrographic observations, Satellite observations: indirect e.g. of the sea-surface

The solution is a pdf! Bayes theorem: Data assimilation: general formulation

Bayes Theorem Conditional pdf: Similarly: Combine: Even better:

Filters and smoothers Filter: solve 3D problem sequentially Smoother: solve 4D problem once Note: I assumed a dynamical (=time-varying) model here.

Most present-day data-assimilation methods assume that the model and observation pdf are Gaussian distributed. Advantage: only first two moments, mean and covariance, needed. Disadvantage: This is not always a good approximation. Gaussian distribution: The Gaussian assumption I Multivariate Gaussian for n-dimensional model vector (state or parameters or a combination):

The Gaussian assumption II The pdf of the observations is actually (and luckily) the pdf of the observations given the model state. Hence we need the model equivalent of the observations. We can express each observation as: in which d is an m-dimensional observation vector. The observation pdf is now given by:

(Ensemble) Kalman Filter I Use Gaussianity in Bayes at a specific time: Hence Complete the squares to find again a Gaussian (only for linear H !!!):

(Ensemble) Kalman Filter III Both lead to the Kalman filter equations, which again are just the least squares solutions: Two possibilities to find the expressions for the mean and covariance: 1)Completing the squares 2)Writing the new solution as a linear combination of model and observations, assume unbiased and minimal trace of error covariance.

Spatial correlation of SSH and SST in the Indian Ocean x x Haugen and Evensen, 2002 The error covariance: tells us how model variables co-vary Likewise you can find the covariance of a model parameter with any other model variable.

(Ensemble) Kalman Filter IV Interestingly Old solution: T n = _______ T 0 + _______ T m sm2sm2 s02s02 s m 2 + s 0 2 Hence we find the same solution, but now generalized to more dimensions and partial or indirect observations (H operator)!

(Ensemble) Kalman Filter V Operational steps: 1)Determine mean and covariance from the prior pdf 2)Use Kalman filter equations to update the model variables 3)Propagate the new forward in time using the model equations 4)Propagate the new covariance in time using the linearized model equations 5)Back to 2) In the ensemble Kalman filter the mean and covariance are found from an ensemble of model states/runs.

Propagation of pdf in time: ensemble or particle methods

And parameter estimation? Several ways: 1)Augment the state vector with the parameters (i.e. add parameters to the state vector). This makes the parameter estimates time variable. How do we find the covariances? See Friday. 2) Prescribe the covariances between parameters and observed variables. 3)Find best fits using Monte-Carlo runs (lots of tries, can be expensive).

Nonlinear measurement operators H (but observations still Gaussian distributed ) 1) Use the nonlinear H in the Kalman filter equations (no rigorous derivation possible). 2) Augment the state vector x with H(x) (still assuming Gausian pdf for new model vector). 3)Solve the nonlinear problem. For large-scale problems variational methods tend to be efficient: 3DVar. (Sometimes also used when H is linear.)

Variational methods Recall that for Gaussian distributed model and observations the posterior pdf can be written as: A variational method looks for the most probable state, which is the maximum of this posterior pdf also called the mode. Instead of looking for the maximum one solves for the minimum of a so-called costfunction.

The pdf can be rewritten as in which Find min J from variational derivative: J is costfunction or penalty function Variational methods

Gradient descent methods J model variable

3DVar 3DVar solves for the minimum of the costfucntion, which is the maximum of the posterior pdf. Advantages: 1)When preconditioned appropriately it can handle very large dimensional systems (e.g. 10^8 variables) 2)It uses an adjoint, which tells us the linear sensitivity of each model variable to all other model variables. This is useful for e.g. observation impact. Disadvantages: 1)The adjoint has to be coded 2)It doesnt automatically provide an error estimate. For weakly nonlinear systems the inverse of the 2 nd variational derivative, the Hessian, is equal to the covariance, but it comes at great cost.

4DVar There is an interesting extension to this formulation to a smoother. Filter: Solve a 3D problem at each observation time. 4DVar: Solve a 3D problem at the beginning of the time window using all observations. Note we get a new full 4D solution!

4DVar: The observation pdf The part of the costfunction related to the observations can again be written as: But now we have sets of observations at different times t obs, and H is a complex operator that includes the forward model. If observation errors at different times are uncorrelated we find:

4DVar: the dynamical model The dynamical model is denoted by M: Using the model operator twice brings us to the next time step: And some short-hand notation:

4DVar The total costfunction that we have to minimize now becomes: in which the measurement operator H i contains the forward model: defined as

Weak-constraint 4DVar So far, we have considered errors in observations and the model initial condition. What about the forward model? If the forward model has errors, i.e. if there are errors present in the model equations the model pdf has to be extended. This leads to extra terms in the costfunction:

Parameter estimation Bayes: Looks simple, but we dont observe model parameters…. We observe model fields, so: in which H has to be found from model integrations.

And what if the distributions are non-Gaussian? Fully non-Gaussian data-assimilation methods exist, but none has proven to work for large-scale systems. A few do have potential among which the Particle Filter. It has been shown that the curse of dimensionality has been cured….

Summary and outlook 1We know how to formulate the data assimilation problem using Bayses Theorem. 2We have derived the Kalman Filter and shown that it is the best linear unbiased estimator (BLUE). 3We derived 3D and 4DVar and discussed some of their properties. 4This forms the basis for what is to come in the next days! 5Notation: when we talk about the state vector and its observations notation will be changed: becomes x, and d becomes y. Beware !!!