EnKF Overview and Theory

Slides:



Advertisements
Similar presentations
Data Assimilation Andrew Collard. Overview Introduction to Atmospheric Data Assimilation Control Variables Observations Background Error Covariance Summary.
Advertisements

Polly Smith, Alison Fowler, Amos Lawless School of Mathematical and Physical Sciences, University of Reading Exploring coupled data assimilation using.
Use of Kalman filters in time and frequency analysis John Davis 1st May 2011.
Initialization Issues of Coupled Ocean-atmosphere Prediction System Climate and Environment System Research Center Seoul National University, Korea In-Sik.
Effects of model error on ensemble forecast using the EnKF Hiroshi Koyama 1 and Masahiro Watanabe 2 1 : Center for Climate System Research, University.
1 What constrains spread growth in forecasts initialized from ensemble Kalman filters? Tom Hamill (& Jeff Whitaker) NOAA Earth System Research Lab Boulder,
Jidong Gao and David Stensrud Some OSSEs on Assimilation of Radar Data with a Hybrid 3DVAR/EnKF Method.
Ibrahim Hoteit KAUST, CSIM, May 2010 Should we be using Data Assimilation to Combine Seismic Imaging and Reservoir Modeling? Earth Sciences and Engineering.
Assimilating Sounding, Surface and Profiler Observations with a WRF-based EnKF for An MCV Case during BAMEX Zhiyong Meng & Fuqing Zhang Texas A&M University.
1/20 Accelerating minimizations in ensemble variational assimilation G. Desroziers, L. Berre Météo-France/CNRS (CNRM/GAME)
Review of “Covariance Localization” in Ensemble Filters Tom Hamill NOAA Earth System Research Lab, Boulder, CO NOAA Earth System Research.
ASSIMILATION of RADAR DATA at CONVECTIVE SCALES with the EnKF: PERFECT-MODEL EXPERIMENTS USING WRF / DART Altuğ Aksoy National Center for Atmospheric Research.
WWRP/THORPEX WORKSHOP on 4D-VAR and ENSEMBLE KALMAN FILTER INTER-COMPARISONS BUENOS AIRES - ARGENTINA, NOVEMBER 2008 Topics for Discussion in Session.
Advanced data assimilation methods- EKF and EnKF Hong Li and Eugenia Kalnay University of Maryland July 2006.
Accounting for correlated AMSU-A observation errors in an ensemble Kalman filter 18 August 2014 The World Weather Open Science Conference, Montreal, Canada.
UNBIASED ESTIAMTION OF ANALYSIS AND FORECAST ERROR VARIANCES
Introduction to Numerical Weather Prediction and Ensemble Weather Forecasting Tom Hamill NOAA-CIRES Climate Diagnostics Center Boulder, Colorado USA.
Background In deriving basic understanding of atmospheric phenomena, the analysis often revolves around discovering and exploiting relationships between.
Models for model error –Additive noise. What is Q(x 1, x 2, t 1, t 2 )? –Covariance inflation –Multiplicative noise? –Parameter uncertainty –“Structural”
A comparison of hybrid ensemble transform Kalman filter(ETKF)-3DVAR and ensemble square root filter (EnSRF) analysis schemes Xuguang Wang NOAA/ESRL/PSD,
Francesca Marcucci, Lucio Torrisi with the contribution of Valeria Montesarchio, ISMAR-CNR CNMCA, National Meteorological Center,Italy First experiments.
Ensemble-Based Atmospheric Data Assimilation: A Tutorial Thomas M. Hamill NOAA-CIRES Climate Diagnostics Center Boulder, Colorado USA 80301
Kalman filtering techniques for parameter estimation Jared Barber Department of Mathematics, University of Pittsburgh Work with Ivan Yotov and Mark Tronzo.
Ensemble Data Assimilation and Uncertainty Quantification Jeffrey Anderson, Alicia Karspeck, Tim Hoar, Nancy Collins, Kevin Raeder, Steve Yeager National.
EnKF Overview and Theory
Observing Strategy and Observation Targeting for Tropical Cyclones Using Ensemble-Based Sensitivity Analysis and Data Assimilation Chen, Deng-Shun 3 Dec,
Eidgenössisches Departement des Innern EDI Bundesamt für Meteorologie und Klimatologie MeteoSchweiz Statistical Characteristics of High- Resolution COSMO.
WWOSC 2014, Aug 16 – 21, Montreal 1 Impact of initial ensemble perturbations provided by convective-scale ensemble data assimilation in the COSMO-DE model.
ISDA 2014, Feb 24 – 28, Munich 1 Impact of ensemble perturbations provided by convective-scale ensemble data assimilation in the COSMO-DE model Florian.
Computational Issues: An EnKF Perspective Jeff Whitaker NOAA Earth System Research Lab ENIAC 1948“Roadrunner” 2008.
1 ESTIMATING THE STATE OF LARGE SPATIOTEMPORALLY CHAOTIC SYSTEMS: WEATHER FORECASTING, ETC. Edward Ott University of Maryland Main Reference: E. OTT, B.
CSDA Conference, Limassol, 2005 University of Medicine and Pharmacy “Gr. T. Popa” Iasi Department of Mathematics and Informatics Gabriel Dimitriu University.
Federal Department of Home Affairs FDHA Federal Office of Meteorology and Climatology MeteoSwiss High-resolution data assimilation in COSMO: Status and.
MPO 674 Lecture 20 3/26/15. 3d-Var vs 4d-Var.
Data assimilation and observing systems strategies Pierre Gauthier Data Assimilation and Satellite Meteorology Division Meteorological Service of Canada.
Dusanka Zupanski And Scott Denning Colorado State University Fort Collins, CO CMDL Workshop on Modeling and Data Analysis of Atmospheric CO.
Ensemble Data Assimilation for the Mesoscale: A Progress Report Massimo Bonavita, Lucio Torrisi, Francesca Marcucci CNMCA National Meteorological Service.
Sophie RICCI CALTECH/JPL Post-doc Advisor : Ichiro Fukumori The diabatic errors in the formulation of the data assimilation Kalman Filter/Smoother system.
Applications of optimal control and EnKF to Flow Simulation and Modeling Florida State University, February, 2005, Tallahassee, Florida The Maximum.
Research Vignette: The TransCom3 Time-Dependent Global CO 2 Flux Inversion … and More David F. Baker NCAR 12 July 2007 David F. Baker NCAR 12 July 2007.
Data assimilation, short-term forecast, and forecasting error
Data assimilation and forecasting the weather (!) Eugenia Kalnay and many friends University of Maryland.
Ensemble Kalman Filter in a boundary layer 1D numerical model Samuel Rémy and Thierry Bergot (Météo-France) Workshop on ensemble methods in meteorology.
A kernel-density based ensemble filter applicable (?) to high-dimensional systems Tom Hamill NOAA Earth System Research Lab Physical Sciences Division.
Local Predictability of the Performance of an Ensemble Forecast System Liz Satterfield and Istvan Szunyogh Texas A&M University, College Station, TX Third.
Implementation and Testing of 3DEnVAR and 4DEnVAR Algorithms within the ARPS Data Assimilation Framework Chengsi Liu, Ming Xue, and Rong Kong Center for.
Nonlinear State Estimation
A Random Subgrouping Scheme for Ensemble Kalman Filters Yun Liu Dept. of Atmospheric and Oceanic Science, University of Maryland Atmospheric and oceanic.
The application of ensemble Kalman filter in adaptive observation and information content estimation studies Junjie Liu and Eugenia Kalnay July 13th, 2007.
École Doctorale des Sciences de l'Environnement d’Île-de-France Année Universitaire Modélisation Numérique de l’Écoulement Atmosphérique et Assimilation.
École Doctorale des Sciences de l'Environnement d’ Î le-de-France Année Modélisation Numérique de l’Écoulement Atmosphérique et Assimilation.
The Ensemble Kalman filter
June 20, 2005Workshop on Chemical data assimilation and data needs Data Assimilation Methods Experience from operational meteorological assimilation John.
© Crown copyright Met Office Mismatching Perturbations at the Lateral Boundaries in Limited-Area Ensemble Forecasting Jean-François Caron … or why limited-area.
ECMWF/EUMETSAT NWP-SAF Satellite data assimilation Training Course Mar 2016.
Hybrid Data Assimilation
Data Assimilation.
Methods for dealing with spurious covariances arising from small samples in ensemble data assimilation Jeff Whitaker NOAA Earth.
Data Assimilation Theory CTCD Data Assimilation Workshop Nov 2005
BACY = Basic Cycling A COSMO Data Assimilation Testbed for Research and Development Roland Potthast, Hendrik Reich, Christoph Schraff, Klaus.
background error covariance matrices Rescaled EnKF Optimization
Vertical localization issues in LETKF
New Approaches to Data Assimilation
FSOI adapted for used with 4D-EnVar
Mohamed Jardak (presenter), N. E. Bowler, A. M. Clayton, G. W
Comparison of different combinations of ensemble-based and variational data assimilation approaches for deterministic NWP Mark Buehner Data Assimilation.
Principles of the Global Positioning System Lecture 13
Project Team: Mark Buehner Cecilien Charette Bin He Peter Houtekamer
Kalman Filter: Bayes Interpretation
Sarah Dance DARC/University of Reading
Presentation transcript:

EnKF Overview and Theory Jeff Whitaker jeffrey.s.whitaker@noaa.gov Will give a brief overview of the role of DA in NWP, then get into the specifics of the algorithms we use in NWP and how ensembles fit in. In the latter half of the talk I’ll describe where operational NWP centers, particularly our own, are heading and what the main outstanding issues are.

The Numerical Weather Prediction Process Forecast Model 06 UTC obs Analysis Data Assimilation 00 UTC Forecast To understand what data assimilation is, let’s look at the big picture of numerical weather prediction. Every 6-h we collect all the obs (including satellites, aircraft, radiosondes) These observations are blended with a forecast from a weather prediction model in the process we call DA, to produce new initial conditions for the model (what we call the analysis). That analysis is used to start a new forecast, which is blended with obs 6-h later, and the process continues. The analyses, and hence the forecasts, get more accurate when we get better observations, or the forecast model or data assimilation algorithm improves. The model is very important in this process, since it carries the information from past observvations forward in time. As Sandy showed, most of the improvement we have seen in 5-d forecasts has come from improvements to the model and the assimilation process over the last decade. Analyses and forecasts become more accurate when: Observations, forecast model and/or data assimilation components improve. Forecast model carries information from past observations.

The Data Assimilation Process Bayes Thm + Guassian assumption (Kalman Filter) xa = xb + K(yo-Hxb) where K =PbHT(HPbHT + R)-1 Pa = (I-KH) Pb Background forecast xb with uncertainty Pb Obs yo with uncertainty R Forecast model xa  xb, Pb Pb Cycling xa and xb easy, but Pb and Pa are huge matrices!

The EnKF approach Instead of evolving the full error covariance matrix (Pa), evolve a random sample (via an ensemble forecast). In the DA step, Pb is a sample estimate from the forecast ensemble, and each ensemble member is updated individually.

The EnKF DA Process EnKF – estimate Pb from ensemble update every member with xa = xb + K(yo-Hxb) where K =PbHT(HPbHT + R)-1 Ensemble of background forecasts xb Obs yo with uncertainty R Forecast model xa  xb for each member Cycle every ensemble member instead of propagating Pb

Two categories of EnKF ‘stochastic’ EnKF (original formulation by Houtekamer and Mitchell, 1998 MWR) treats obs as ensemble by adding N(0,R) noise. Every member updated with the same familiar KF equation (simple!) Used by Environment Canada. ‘deterministic’ EnKF (ETKF, Bishop et al 2001 MWR; EnSRF, Whitaker and Hamill 2002 MWR) avoids this by constructing analysis perturbations so that Pa consistent with KF is obtained. Mean updated with Kalman update equation. Perturbations updated differently. More accurate than stochastic approach for small ensembles.

Computational shortcuts in EnKF: (1) Simplifying Kalman gain calculation The Kalman gain is computed without explicitly forming Pb – only matrices of size (M x N) and (N x N) are needed where N is the number of obs and M is the number of state vars (N << M usually). The key here is that the huge matrix Pb is never explicitly formed

Computational shortcuts in EnKF: (2) serial processing of observations (requires observation error covariance R to be diagonal) Method 1 Observations 1 and 2 Background forecasts EnKF Analyses Method 2 We can make this matrices even smaller by using a trick, which takes advantage of the fact that R is often diagonal, i.e. observation errors are uncorrelated. In that case, you can assimilate obs in batches, or even one by one and get the same answer as when they are all assimilated at once. For exampe, if we have two observations, we can calculate the full Kalman gain and update the model state in one go. Or, we can assimilate just the first ob, update the model state, use that model state as the backgound for assimilating the second ob, then assimilate that one. If R is diagonal, the result is the same. Observation 1 Observation 2 Background forecasts EnKF Analyses after obs 1 EnKF Analyses

NOAA EnKF implements two algorithms Serial EnKF (observations processed one at a time) Both stochastic and deterministic options available. Local Ensemble Transform Filter (LETKF) Update computed with all obs at once, but matrices are kept small by updating each grid point independently using only nearby observations.

LETKF Algorithm

Consequences of sampling error Ensemble sizes we can afford are O(100-1000). Rank of full Pb is at least O(106) for current global ensemble resolution. Pb is estimated from XbXbT, so most of eigenvalues are zero. Errors in individual elements of Pb are O(Ne-1/2), but correlated across elements. EnKF fails miserably if raw sample covariance used. Pa grossly underestimated, spread collapses, data assimilation ignores all observations.

Covariance localization – the secret sauce that makes it all work. Basic idea: small values in Pb cannot be estimated accurately, just setting them to zero will increase rank of sample estimate. Hypothesis: Covariances decrease in significance with distance. Method: Taper covariance estimate to zero away from (block) diagonal using a pre-specific function. Also makes algorithms faster, since observations cannot impact the entire state.

Localization: a simple example Estimates of covariances from a small ensemble will be noisy, with signal-to-noise small especially when covariance is small

Localization: a real-world example AMSUA_n15 channel 6 radiance at 150E,-50S. Increment to level 30 (~310mb) temperature for a 1K O-F for 40,80,160,320 and 640 ens members with no localization.

Localization: a real-world example

What about uncertainty in the model itself What about uncertainty in the model itself? Not included in Pb if every member run with the same model Must account for the background error any difference between simulated and true environment. Methods used so far: multiplicative inflation (mult. ens perts by a factor > 1). additive inflation (random perts added to each member – e.g. differences between 24 and 48-h forecasts valid at the same time). model-based schemes (e.g. stochastic kinetic energy backscatter for representing unresolved processes, multi-model/multi-parameterization). Opnl NCEP system used use a combination of 1) and 2), now using a combination of 1) and 3). Only 1) is taken care of within the EnKF itself. 2) Is a separate step (runs after EnKF update and before forecast step) 3) happens inside forecast model.

Relaxation To Prior Spread (RTPS) Inflation Described in DOI: 10 Relaxation To Prior Spread (RTPS) Inflation Described in DOI: 10.1175/MWR-D-11-00276.1 Inflate posterior spread (std. dev) sa back toward prior spread sb Equivalent to Fuqing Zhang and his collaborators first proposed this type of inflation in a 2005 paper. Their version, which I call ‘relaxation to prior perturbation’ inflation, relaxes the posterior perturbations back to the prior with a coefficient alpha, so that if alpha=1 the posterior perts are completely replace by the prior perts, and if alpha=0 nothing is done. Since the posterior perturbations are much smaller than the prior perturbations in regions of dense obs, for a given value of alpha, more inflation is applied where the obs have had a larger impact on the analysis. Accounts for assimilation related errors, which will occur where observations have a larger impact. Also, model error is a bigger fraction of total background error where observations are dense. Here we propose an alternative approach, which we call ‘relaxation to prior stdev, which relaxes the ensemble spread back to the prior spread. This can be expressed a multiplicative inflation where the inflation coefficient depends on the fractional reduction of the ensemble spread that results from assimlating observations.

Why does Pb matter? Flow dependence and unobserved variables ps ob First-Guess Precipitable Water First-Guess SLP contours A simple example of the benefits… This case involves a northward flow of warm and moist air ahead of a maritime cyclone (aka atmospheric river). White contours are background MSLP, colored field is background total column water vapor, and red contours are TCWV increment for a single surface pressure observation at the location denoted by the yellow dot. Covariance between surface pressure at the observation location and the TCWV field estimated from an EnKF ensemble allows the pressure ob to adjust the position ‘atmopsheric river consistent with the surface cyclone. DA system ‘knows’ there is a cyclone/atm river in the background, makes adjustments when given observations taking the dynamics into account. In the old 3DVar system, the climatogical weights were set such that surface pressure obs could not change the humidity field at all. This is because, on average, there is very little correlation between errors in the surface pressure field and errors in the humidity field. However, in certain dynamical situations these correlations can be large, and the ensemble-based DA methods allow much more information to extracted from the observations in those cases. Surface pressure observation can improve analysis of integrated water vapor (through flow-dependent cross-variable relationships). If climo Pb were used (3DVar) there would be no vapor increment. 18

Why combine EnKF and Var? Features from EnKF Features from Var Can propagate Pb from across assimilation windows Treatment of sampling error in ensemble Pb estimate does not depend on H. More flexible treatment of model error (can be treated in ensemble) Dual-resolution capability – can produce a high-res “control” analysis. Automatic initialization of ensemble forecasts. Ease of adding extra constraints to cost function, including a static Pb component. Why not just stand-alone EnKF? Operational centers have used Var systems for many years – existing infrastructure. Var solver makes some aspects easier (localization, extra constraints, dual-res, blending of specified static covaraince).

What is limiting performance now? Sampling error Run larger ensembles. Better (flow and scale dependent) localization methods. Better treatment of non-local observations (model-space localization). Model error Run higher resolution ensembles. Better parameterizations of model uncertainty. Non-gaussian error statistics. Some phenomena not observed often/well enough, dynamical error growth will be nonlinear. Can arise from displacement errors in coherent features. Variables that are physically bounded (humidity, wind speed).