École Doctorale des Sciences de l'Environnement d’Île-de-France Année Universitaire 2014-2015 Modélisation Numérique de l’Écoulement Atmosphérique et Assimilation.

Slides:



Advertisements
Similar presentations
Variational data assimilation and forecast error statistics
Advertisements

General Linear Model With correlated error terms  =  2 V ≠  2 I.
Pattern Classification. Chapter 2 (Part 1): Bayesian Decision Theory (Sections ) Introduction Bayesian Decision Theory–Continuous Features.
Hidden Markov Models. A Hidden Markov Model consists of 1.A sequence of states {X t |t  T } = {X 1, X 2,..., X T }, and 2.A sequence of observations.
Observers and Kalman Filters
Parameter Estimation: Maximum Likelihood Estimation Chapter 3 (Duda et al.) – Sections CS479/679 Pattern Recognition Dr. George Bebis.
Data assimilation schemes in numerical weather forecasting and their link with ensemble forecasting Gérald Desroziers Météo-France, Toulouse, France.
0 Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Validation Olivier Talagrand WMO Workshop on 4D-Var and Ensemble Kalman Filter Intercomparisons Buenos Aires, Argentina 13 November 2008.
Probability theory 2011 The multivariate normal distribution  Characterizing properties of the univariate normal distribution  Different definitions.
Data Basics. Data Matrix Many datasets can be represented as a data matrix. Rows corresponding to entities Columns represents attributes. N: size of the.
Tch-prob1 Chapter 4. Multiple Random Variables Ex Select a student’s name from an urn. S In some random experiments, a number of different quantities.
Visual Recognition Tutorial1 Random variables, distributions, and probability density functions Discrete Random Variables Continuous Random Variables.
Basic Mathematics for Portfolio Management. Statistics Variables x, y, z Constants a, b Observations {x n, y n |n=1,…N} Mean.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Gaussians Pieter Abbeel UC Berkeley EECS Many slides adapted from Thrun, Burgard and Fox, Probabilistic Robotics TexPoint fonts used in EMF. Read the TexPoint.
Probability theory 2008 Outline of lecture 5 The multivariate normal distribution  Characterizing properties of the univariate normal distribution  Different.
The joint probability distribution function of X and Y is denoted by f XY (x,y). The marginal probability distribution function of X, f X (x) is obtained.
Random Variable and Probability Distribution
Lecture II-2: Probability Review
Joint Probability distribution
Joint Probability Distributions
Modern Navigation Thomas Herring
Separate multivariate observations
Today Wrap up of probability Vectors, Matrices. Calculus
Lecture 11: Kalman Filters CS 344R: Robotics Benjamin Kuipers.
Chapter 15 Modeling of Data. Statistics of Data Mean (or average): Variance: Median: a value x j such that half of the data are bigger than it, and half.
III. Multi-Dimensional Random Variables and Application in Vector Quantization.
Federal Department of Home Affairs FDHA Federal Office of Meteorology and Climatology MeteoSwiss High-resolution data assimilation in COSMO: Status and.
Linear Regression Andy Jacobson July 2006 Statistical Anecdotes: Do hospitals make you sick? Student’s story Etymology of “regression”
0 K. Salah 2. Review of Probability and Statistics Refs: Law & Kelton, Chapter 4.
Maximum Likelihood Estimation and Simplified Kalman Filter tecniques for real time Data Assimilation.
Data assimilation and forecasting the weather (!) Eugenia Kalnay and many friends University of Maryland.
Statistics for Business & Economics
CHAPTER 5 SIGNAL SPACE ANALYSIS
Chapter 3: Maximum-Likelihood Parameter Estimation l Introduction l Maximum-Likelihood Estimation l Multivariate Case: unknown , known  l Univariate.
Geology 5670/6670 Inverse Theory 21 Jan 2015 © A.R. Lowry 2015 Read for Fri 23 Jan: Menke Ch 3 (39-68) Last time: Ordinary Least Squares Inversion Ordinary.
Conditional Probability Distributions Eran Segal Weizmann Institute.
Additional Topics in Prediction Methodology. Introduction Predictive distribution for random variable Y 0 is meant to capture all the information about.
III. Multi-Dimensional Random Variables and Application in Vector Quantization.
Geology 6600/7600 Signal Analysis 02 Sep 2015 © A.R. Lowry 2015 Last time: Signal Analysis is a set of tools used to extract information from sequences.
Math 4030 – 6a Joint Distributions (Discrete)
Chapter 20 Classification and Estimation Classification – Feature selection Good feature have four characteristics: –Discrimination. Features.
Brief Review Probability and Statistics. Probability distributions Continuous distributions.
1 Probability and Statistical Inference (9th Edition) Chapter 4 Bivariate Distributions November 4, 2015.
École Doctorale des Sciences de l'Environnement d’Île-de-France Année Universitaire Modélisation Numérique de l’Écoulement Atmosphérique et Assimilation.
Geology 5670/6670 Inverse Theory 28 Jan 2015 © A.R. Lowry 2015 Read for Fri 30 Jan: Menke Ch 4 (69-88) Last time: Ordinary Least Squares: Uncertainty The.
- 1 - Preliminaries Multivariate normal model (section 3.6, Gelman) –For a multi-parameter vector y, multivariate normal distribution is where  is covariance.
École Doctorale des Sciences de l'Environnement d’ Î le-de-France Année Modélisation Numérique de l’Écoulement Atmosphérique et Assimilation.
École Doctorale des Sciences de l'Environnement d’Île-de-France Année Universitaire Modélisation Numérique de l’Écoulement Atmosphérique et Assimilation.
MathematicalMarketing Slide 5.1 OLS Chapter 5: Ordinary Least Square Regression We will be discussing  The Linear Regression Model  Estimation of the.
École Doctorale des Sciences de l'Environnement d’ Î le-de-France Année Modélisation Numérique de l’Écoulement Atmosphérique et Assimilation.
1 Ka-fu Wong University of Hong Kong A Brief Review of Probability, Statistics, and Regression for Forecasting.
Probabilistic Robotics Bayes Filter Implementations Gaussian filters.
Part 3: Estimation of Parameters. Estimation of Parameters Most of the time, we have random samples but not the densities given. If the parametric form.
Colorado Center for Astrodynamics Research The University of Colorado 1 STATISTICAL ORBIT DETERMINATION Statistical Interpretation of Least Squares ASEN.
CWR 6536 Stochastic Subsurface Hydrology Optimal Estimation of Hydrologic Parameters.
Introduction to Vectors and Matrices
Biointelligence Laboratory, Seoul National University
CS479/679 Pattern Recognition Dr. George Bebis
Probability Theory and Parameter Estimation I
CH 5: Multivariate Methods
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
Error statistics in data assimilation
OVERVIEW OF LINEAR MODELS
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Introduction to Vectors and Matrices
Multivariate Methods Berlin Chen
Multivariate Methods Berlin Chen, 2005 References:
Topic 11: Matrix Approach to Linear Regression
Presentation transcript:

École Doctorale des Sciences de l'Environnement d’Île-de-France Année Universitaire Modélisation Numérique de l’Écoulement Atmosphérique et Assimilation de Données Olivier Talagrand Cours 4 23 Avril 2015

After N. Gustafsson

Optimal Interpolation Random field  (  ) Observation network  1,  2, …,  p For one particular realization of the field, observations y j =  (  j ) +  j, j = 1, …, p, making up vector y = (y j ) Estimate x =  (  ) at given point , in the form x a =  +  j  j y j =  +  T y, where  = (  j )  and the  j ’s being determined so as to minimize the expected quadratic estimation error E[(x-x a ) 2 ]

Optimal Interpolation (continued 1) Solution x a = E(x)  + E(x’y’ T ) [E(y’y’ T )] -1  y - E(y)  = E(x)  + C xy [C yy ] -1  y - E(y)  i. e.,  T = C xy [C yy ] -1  = E(x)  -  T E(y) Estimate is unbiased E(x-x a ) = 0 Minimized quadratic estimation error E[(x-x a ) 2 ] = E(x’ 2 ) - E[(x’ a ) 2 ]) = C xx - C xy [C yy ] -1 C yx Estimation made in terms of deviations x’ and y’ from expectations E(x) and E(y).

Optimal Interpolation (continued 2) x a = E(x)  + E(x’y’ T ) [E(y’y’ T )] -1  y - E(y)   y j =  (  j ) +  j E(y j ’y k ’)= E[  ’(  j ) +  j ’][  ’(  k ) +  k ’] If observation errors  j are mutually uncorrelated, have common variance r, and are uncorrelated with field , then E(y j ’y k ’)= C  (  j,  k ) + r  jk and E(x’y j ’)= C  ( ,  j )

Optimal Interpolation (continued 3) x a = E(x)  + C xy [C yy ] -1  y - E(y)  Vector  = (  j )  [C yy ] -1  y - E(y)  is independent of variable to be estimated x a = E(x)  +  j  j E(x’y j ’)  a (  ) = E[  (  )]  +  j  j E[  ’(  ) y j ’] = E[  (  )]  +  j  j C  ( ,  j ) Correction made on background expectation is a linear combination of the p functions C  ( ,  j ) C  ( ,  j ), considered as a function of estimation position , is the representer associated with observation y j.

Optimal Interpolation (continued 4) Univariate interpolation. Each physical field (e. g. temperature) determined from observations of that field only. Multivariate interpolation. Observations of different physical fields are used simultaneously. Requires specification of cross-covariances between various fields. Cross-covariances between mass and velocity fields can simply be modelled on the basis of geostrophic balance. Cross-covariances between humidity and temperature (and other) fields still a problem.

After N. Gustafsson

After A. Lorenc, MWR, 1981

Optimal Interpolation (continued 5) Observation vector y Estimation of a scalar x x a = E(x)  + C xy [C yy ] -1  y - E(y)  p a  E[(x-x a ) 2 ] = E(x’ 2 ) - E[(x’ a ) 2 ]) = C xx - C xy [C yy ] -1 C yx Estimation of a vector x x a = E(x)  + C xy [C yy ] -1  y - E(y)  P a  E[(x-x a ) (x-x a ) T ] = E(x’x’ T ) - E(x’ a x’ aT ) = C xx - C xy [C yy ] -1 C yx

Optimal Interpolation (continued 6) x a = E(x)  + C xy [C yy ] -1  y - E(y)  P a = C xx - C xy [C yy ] -1 C yx If probability distribution for couple (x, y) is Gaussian (with, in particular, covariance matrix then Optimal Interpolation achieves Bayesian estimation, in the sense that P(x | y)  N [x a, P a ]

Best Linear Unbiased Estimate State vector x, belonging to state space S (dim S = n), to be estimated. Available data in the form of  A ‘background’ estimate (e. g. forecast from the past), belonging to state space, with dimension n x b = x +  b  An additional set of data (e. g. observations), belonging to observation space, with dimension p y = Hx +  H is known linear observation operator. Assume probability distribution is known for the couple (  b,  ). Assume E(  b ) = 0, E(  ) = 0, E(  b  T ) = 0 (not restrictive) Set E(  b  b T ) = P b (also often denoted B), E(  T ) = R

Best Linear Unbiased Estimate (continuation 1) x b = x +  b (1) y = Hx +  (2) A probability distribution being known for the couple (  b,  ), eqs (1-2) define probability distribution for the couple (x, y), with E(x) = x b, x’ = x - E(x) = -  b E(y) = Hx b, y’ = y - E(y) = y - Hx b =  - H  b d  y - Hx b is called the innovation vector.

Best Linear Unbiased Estimate (continuation 2) Apply formulæ for Optimal Interpolation x a = x b + P b H T [HP b H T + R] -1 (y - Hx b ) P a = P b - P b H T [HP b H T + R] -1 HP b x a is the Best Linear Unbiased Estimate (BLUE) of x from x b and y. Equivalent set of formulæ x a = x b + P a H T R -1 (y - Hx b ) [P a ] -1 = [P b ] -1 + H T R -1 H Matrix K = P b H T [HP b H T + R] -1 = P a H T R -1 is gain matrix. If probability distributions are globally gaussian, BLUE achieves bayesian estimation, in the sense that P(x | x b, y) = N [x a, P a ].

Best Linear Unbiased Estimate (continuation 3) H can be any linear operator Example : (scalar) satellite observation x = (x 1, …, x n ) T temperature profile Observation y =  i  h i x i +  = Hx + ,  H = (h 1, …, h n ), E(  2 ) = r Backgroundx b = (x 1 b, …, x n b ) T, error covariance matrix P b = (p ij b ) x a = x b + P b H T [HP b H T + R] -1 (y - Hx b ) [HP b H T + R] -1 (y - Hx b ) = (y -   h  x  b ) / (  ij h i h j p ij b + r)   scalar !  P b = p b I n x i a = x i b + p b h i   P b = diag(p ii b ) x i a = x i b + p ii b h i   General case x i a = x i b +  j p ij b h j  Each level i is corrected, not only because of its own contribution to the observation, but because of the contribution of the other levels to which its background error is correlated.