Ensemble-Based Atmospheric Data Assimilation: A Tutorial Thomas M. Hamill NOAA-CIRES Climate Diagnostics Center Boulder, Colorado USA 80301

Slides:



Advertisements
Similar presentations
Introduction to Data Assimilation NCEO Data-assimilation training days 5-7 July 2010 Peter Jan van Leeuwen Data Assimilation Research Center (DARC) University.
Advertisements

ECMWF Data Assimilation Training Course - Kalman Filter Techniques
Initialization Issues of Coupled Ocean-atmosphere Prediction System Climate and Environment System Research Center Seoul National University, Korea In-Sik.
Effects of model error on ensemble forecast using the EnKF Hiroshi Koyama 1 and Masahiro Watanabe 2 1 : Center for Climate System Research, University.
1 What constrains spread growth in forecasts initialized from ensemble Kalman filters? Tom Hamill (& Jeff Whitaker) NOAA Earth System Research Lab Boulder,
Ensemble-4DVAR for NCEP hybrid GSI- EnKF data assimilation system 1 5 th EnKF workshop, New York, May 22, 2012 Xuguang Wang, Ting Lei University of Oklahoma,
Jidong Gao and David Stensrud Some OSSEs on Assimilation of Radar Data with a Hybrid 3DVAR/EnKF Method.
Balance in the EnKF Jeffrey D. Kepert Centre for Australian Weather and Climate Research A partnership between the Australian Bureau of Meteorology and.
Ibrahim Hoteit KAUST, CSIM, May 2010 Should we be using Data Assimilation to Combine Seismic Imaging and Reservoir Modeling? Earth Sciences and Engineering.
Data assimilation schemes in numerical weather forecasting and their link with ensemble forecasting Gérald Desroziers Météo-France, Toulouse, France.
Toward a Real Time Mesoscale Ensemble Kalman Filter Gregory J. Hakim Dept. of Atmospheric Sciences, University of Washington Collaborators: Ryan Torn (UW)
Review of “Covariance Localization” in Ensemble Filters Tom Hamill NOAA Earth System Research Lab, Boulder, CO NOAA Earth System Research.
ASSIMILATION of RADAR DATA at CONVECTIVE SCALES with the EnKF: PERFECT-MODEL EXPERIMENTS USING WRF / DART Altuğ Aksoy National Center for Atmospheric Research.
Advanced data assimilation methods with evolving forecast error covariance Four-dimensional variational analysis (4D-Var) Shu-Chih Yang (with EK)
Advanced data assimilation methods- EKF and EnKF Hong Li and Eugenia Kalnay University of Maryland July 2006.
Ensemble Kalman Filter Methods
Accounting for correlated AMSU-A observation errors in an ensemble Kalman filter 18 August 2014 The World Weather Open Science Conference, Montreal, Canada.
Current Status of the Development of the Local Ensemble Transform Kalman Filter at UMD Istvan Szunyogh representing the UMD “Chaos-Weather” Group Ensemble.
Two Methods of Localization  Model Covariance Matrix Localization (B Localization)  Accomplished by taking a Schur product between the model covariance.
Introduction to Numerical Weather Prediction and Ensemble Weather Forecasting Tom Hamill NOAA-CIRES Climate Diagnostics Center Boulder, Colorado USA.
Background In deriving basic understanding of atmospheric phenomena, the analysis often revolves around discovering and exploiting relationships between.
A comparison of hybrid ensemble transform Kalman filter(ETKF)-3DVAR and ensemble square root filter (EnSRF) analysis schemes Xuguang Wang NOAA/ESRL/PSD,
Numerical Weather Prediction Division The usage of the ATOVS data in the Korea Meteorological Administration (KMA) Sang-Won Joo Korea Meteorological Administration.
Colorado Center for Astrodynamics Research The University of Colorado STATISTICAL ORBIT DETERMINATION Project Report Unscented kalman Filter Information.
Ensemble-variational sea ice data assimilation Anna Shlyaeva, Mark Buehner, Alain Caya, Data Assimilation and Satellite Meteorology Research Jean-Francois.
Kalman filtering techniques for parameter estimation Jared Barber Department of Mathematics, University of Pittsburgh Work with Ivan Yotov and Mark Tronzo.
Ensemble Data Assimilation and Uncertainty Quantification Jeffrey Anderson, Alicia Karspeck, Tim Hoar, Nancy Collins, Kevin Raeder, Steve Yeager National.
EnKF Overview and Theory
Parameter estimation: To what extent can data assimilation techniques correctly uncover stochasticity? Jim Hansen MIT, EAPS (with lots.
Computational Issues: An EnKF Perspective Jeff Whitaker NOAA Earth System Research Lab ENIAC 1948“Roadrunner” 2008.
1 ESTIMATING THE STATE OF LARGE SPATIOTEMPORALLY CHAOTIC SYSTEMS: WEATHER FORECASTING, ETC. Edward Ott University of Maryland Main Reference: E. OTT, B.
Exploring sample size issues for 6-10 day forecasts using ECMWF’s reforecast data set Model: 2005 version of ECMWF model; T255 resolution. Initial Conditions:
CSDA Conference, Limassol, 2005 University of Medicine and Pharmacy “Gr. T. Popa” Iasi Department of Mathematics and Informatics Gabriel Dimitriu University.
Federal Department of Home Affairs FDHA Federal Office of Meteorology and Climatology MeteoSwiss High-resolution data assimilation in COSMO: Status and.
MPO 674 Lecture 20 3/26/15. 3d-Var vs 4d-Var.
Hybrid Variational/Ensemble Data Assimilation
Dusanka Zupanski And Scott Denning Colorado State University Fort Collins, CO CMDL Workshop on Modeling and Data Analysis of Atmospheric CO.
1 GSI/ETKF Regional Hybrid Data Assimilation with MMM Hybrid Testbed Arthur P. Mizzi NCAR/MMM 2011 GSI Workshop June 29 – July 1, 2011.
Data Assimilation Using Modulated Ensembles Craig H. Bishop, Daniel Hodyss Naval Research Laboratory Monterey, CA, USA September 14, 2009 Data Assimilation.
Ensemble Data Assimilation for the Mesoscale: A Progress Report Massimo Bonavita, Lucio Torrisi, Francesca Marcucci CNMCA National Meteorological Service.
2004 SIAM Annual Meeting Minisymposium on Data Assimilation and Predictability for Atmospheric and Oceanographic Modeling July 15, 2004, Portland, Oregon.
Applications of optimal control and EnKF to Flow Simulation and Modeling Florida State University, February, 2005, Tallahassee, Florida The Maximum.
MODEL ERROR ESTIMATION EMPLOYING DATA ASSIMILATION METHODOLOGIES Dusanka Zupanski Cooperative Institute for Research in the Atmosphere Colorado State University.
Research Vignette: The TransCom3 Time-Dependent Global CO 2 Flux Inversion … and More David F. Baker NCAR 12 July 2007 David F. Baker NCAR 12 July 2007.
Data assimilation and forecasting the weather (!) Eugenia Kalnay and many friends University of Maryland.
Ensemble data assimilation in an operational context: the experience at the Italian Weather Service Massimo Bonavita and Lucio Torrisi CNMCA-UGM, Rome.
Adaptive Hybrid EnKF-OI for State- Parameters Estimation in Contaminant Transport Models Mohamad E. Gharamti, Johan Valstar, Ibrahim Hoteit European Geoscience.
A kernel-density based ensemble filter applicable (?) to high-dimensional systems Tom Hamill NOAA Earth System Research Lab Physical Sciences Division.
Local Predictability of the Performance of an Ensemble Forecast System Liz Satterfield and Istvan Szunyogh Texas A&M University, College Station, TX Third.
Implementation and Testing of 3DEnVAR and 4DEnVAR Algorithms within the ARPS Data Assimilation Framework Chengsi Liu, Ming Xue, and Rong Kong Center for.
University of Oklahoma, Norman, OK
A Random Subgrouping Scheme for Ensemble Kalman Filters Yun Liu Dept. of Atmospheric and Oceanic Science, University of Maryland Atmospheric and oceanic.
École Doctorale des Sciences de l'Environnement d’ Î le-de-France Année Modélisation Numérique de l’Écoulement Atmosphérique et Assimilation.
École Doctorale des Sciences de l'Environnement d’Île-de-France Année Universitaire Modélisation Numérique de l’Écoulement Atmosphérique et Assimilation.
École Doctorale des Sciences de l'Environnement d’ Î le-de-France Année Modélisation Numérique de l’Écoulement Atmosphérique et Assimilation.
The Ensemble Kalman filter
ECMWF/EUMETSAT NWP-SAF Satellite data assimilation Training Course Mar 2016.
Hybrid Data Assimilation
Methods for dealing with spurious covariances arising from small samples in ensemble data assimilation Jeff Whitaker NOAA Earth.
Data Assimilation Theory CTCD Data Assimilation Workshop Nov 2005
EnKF Overview and Theory
background error covariance matrices Rescaled EnKF Optimization
Ensemble variance loss in transport models:
New Approaches to Data Assimilation
FSOI adapted for used with 4D-EnVar
Initial trials of 4DEnVAR
Project Team: Mark Buehner Cecilien Charette Bin He Peter Houtekamer
Hartmut Bösch and Sarah Dance
Kalman Filter: Bayes Interpretation
Sarah Dance DARC/University of Reading
Presentation transcript:

Ensemble-Based Atmospheric Data Assimilation: A Tutorial Thomas M. Hamill NOAA-CIRES Climate Diagnostics Center Boulder, Colorado USA

Motivation Ensemble forecasting : Provides flow-dependent estimate of uncertainty of the forecast. Data assimilation : requires information about uncertainty in prior forecast and observations. More accurate estimate of uncertainty, less error in the analysis. Improved initial conditions, improved ensemble forecasts. Ensemble forecasting  data assimilation.

Example: where flow-dependent first-guess errors help

New observation is inconsistent with first guess

3D-Var produces an unappealing analysis, bowing the front out just in the region of the observation. The way an observation influences surrounding points is always the same in 3D-Var --- typically concentric rings of decreasing influence the greater the distance from the observation.

Here’s the front you’d re- draw based on synoptic intuition, understanding that the errors are flow dependent. How can we teach data assimilation algorithms to do this? In this special case, your synoptic intuition would tell you that temperature errors ought to be correlated along the front, from northeast to southwest. However, you wouldn’t expect that same NE-SW correlation if the observation were near the warm front. Analyses might be improved dramatically if the errors could have this flow- dependency.

Ensemble-based data assimilation algorithms Can use ensemble to model the statistics of the first guess (“background”) errors. Tests in simple models show dramatically improved sets of objective analyses These sets of objective analyses could also be useful for initializing ensemble forecasts.

4D-Var: As far as data assimilation can go? Finds model trajectory that best fits observations over a time window. BUT: –Requires linear tangent and adjoint (difficult to code, and are linearity of error growth assumptions met?) –What if forecast model trajectory unlike real atmosphere’s trajectory? (model error) Ensemble-based approaches avoid these problems (though they have their own, as we’ll see).

Example: assimilation of a sparse network of surface-pressure observations

From first principles: Bayesian data assimilation A manipulation of Baye’s Rule, assuming observation errors are independent in time. = observations = [Today’s, all previous] = (unknown) true model state “posterior”“prior”

Bayesian data assimilation: 2-D example Computationally expensive! Here, probabilities explicitly updated on 100x100 grid; costs multiply geometrically with the number of dimensions of model state.

Data assimilation terminology y : Observation vector (raobs, satellite, etc.) x b : Background state vector (1 st guess) x a : Analysis state vector H : (Linear) operator to convert model state  obs R : Observation - error covariance matrix P b : Background - error covariance matrix P a : Analysis - error covariance matrix

Simplifying Bayesian D.A.: toward the Kalman Filter (manipulation of Baye’s rule ) Maximizing equivalent to minimizing i.e., minimizing the functional so

Kalman filter derivation continued After much math, plus other assumptions about linearity, we end up with the “Kalman filter” equations (see Lorenc, QJRMS, 1986). How the background errors at the next data assimilation time are estimated. M is the “tangent linear” of the forecast model M Assumed. This “update” equation tells how to estimate the analysis state. A weighted correction of the difference between the observation and the background is added to the background. K is the “Kalman Gain Matrix.” It indicates how much to weight the observations relative to the background and how to spread their influence to other grid points P a is the “analysis-error covariance. The Kalman filter indicates not only the most likely state but also quantifies the uncertainty in the analysis state. How the forecast is propagated. In “extended Kalman filter” M is replaced by fully nonlinear M

Kalman filter update equation: example in 1-D

Kalman filter problems Covariance propagation step P t a  P b t+1 very expensive Still need tangent linear/adjoint models for evolving covariances, linear error growth assumption questionable.

From the Kalman filter to the ensemble Kalman Filter What if we estimate P b from a random ensemble of forecasts? (Evensen, JGR 1994) Let’s design a procedure so if error growth is linear and ensemble size infinite, gives same result as Kalman filter.

Canonical ensemble Kalman filter update equations H = (possibly nonlinear) operator from model to observation space x = state vector ( i for ith member ) (1)An ensemble of parallel data assimilation cycles is conducted, assimilating perturbed observations. (2)Background-error covariances are estimated using the ensemble. Notes:

The ensemble Kalman filter: a schematic (This schematic is a bit of an inappropriate simplification, for EnKF uses every member to estimate background- error covariances)

How the EnKF works: 2-D example Start with a random sample from bimodal distribution used in previous Bayesian data assimilation example. Contours reflect the Gaussian distribution fitted to ensemble data.

Why perturb the observations? histograms denote the ensemble values; heavy black line denotes theoretical expected analysis-error covariance

How the EnKF may achieve its improvement relative to 3D-Var: better background-error covariances Output from a “single-observation” experiment. The EnKF is cycled for a long time in a dry, primitive equation model. The cycle is interrupted and a single observation 1K greater than first guess is assimilated. Maps of the analysis minus first guess are plotted. These “analysis increments” are proportional to the background-error covariances between every other model grid point and the observation location.

Why are the biggest increments not located at the observation?

Analysis increment: vertical cross section through front

More examples of flow-dependent background-error covariances

Computational trickery in EnKF: (1) serial processing of observations EnKF Background forecasts Observations 1 and 2 Analyses EnKF Background forecasts Observation 1 Analyses after obs 1 EnKF Observation 2 Analyses Method 1 Method 2

Computational trickery in EnKF: (2) Simplifying Kalman gain calculation The key here is that the huge matrix P b is never explicitly formed

Different implementations of ensemble filters Double EnKF (Houtekamer and Mitchell, MWR, March 1998) Ensemble adjustment filter (EnAF; Anderson, MWR, Dec 2001) Ensemble square-root filter (EnSRF; Whitaker and Hamill, MWR, July 2002) Ensemble transform Kalman filter (ETKF; Bishop et al, MWR, March 2001) Local EnKF (Ott et al, U. Maryland; Tellus, in press) Others as well (Lermusiaux, Pham, Keppenne, Heemink, etc.)

Why different implementations? Save computational expense Problems with perturbed obs. Suppose P b ~ N(0,1), R ~ N(0,1), E[corr(x b,y’)] = 0 Random sample x b : [0.19, 0.06, 1.29, 0.36, -0.61] Random sample y’ : [-1.04, 0.46, -0.12, -0.65, 2.10] Sample corr (x b,y’) = ! Noise added through perturbed observations can introduce spurious correlations between background, observation errors. However, there may be some advantages to perturbed observations in situations where the prior is highly nonlinear. See Lawson and Hansen, MWR, Aug 2004.

Adjusting noisy background- error covariances: “covariance localization”

Covariance localization in practice from Hamill review paper, to appear in upcoming Cambridge Press book. See also Houtekamer and Mitchell, MWR, March 1998

How does covariance localization make up for larger ensemble? (a) eigenvalue spectrum from small ensemble too steep, not enough variance in trailing directions. (b) Covariance localization adds directions, flattens eigenvalue spectrum) source: Hamill et al, MWR, Nov 2001

Covariance localization and size of the ensemble Smaller ensembles achieve lowest error and comparable spread/error with a tighter localization function (from Houtekamer and Mitchell, MWR, Jan 2001)

Problem: “filter divergence” Ensemble of solutions drifts away from true solution. During data assimilation, small variance in background forecasts causes data assimilation to ignore influence of new observations.

Filter divergence: some causes Observation error statistics incorrect. Too small an ensemble. –Poor covariance estimates just due to random sampling. –Not enough ensemble members. If M members and G > M growing directions, no variance in some directions. Model error. –Not enough resolution. Interaction of small scales with larger scales impossible. –Deterministic sub-gridscale parameterizations. –Other model aspects unperturbed (e.g., land surface condition) –Others

Effects of misspecifying observation error characteristics

Possible filter divergence remedies Higher-resolution model, more members (but costly!) Covariance localization (discussed before) Possible ways to parameterize model error –Apply bias correction –Covariance inflation –Integrate stochastic noise –Add simulated model error noise at data assimilation time. –Multi-model ensembles

Remedy: covariance inflation r is inflation factor Disadvantage: what if model error in altogether different subspace? source: Anderson and Anderson, MWR, Dec 1999

Remedy: integrating stochastic noise

Remedy : integrating stochastic noise, continued Questions –What is structure of S(x,t)? –Integration methodology for noise? –Will this produce ~ balanced covariances? –Will noise project upon growing structures and increase overall variance? Early experiments in ECMWF ensemble to simulate stochastic effects of sub-gridscale in parameterizations (Buizza et al., QJRMS, 1999).

Remedy : adding noise at data assimilation time Idea follows Dee (Apr 1995 MWR) and Mitchell and Houtekamer (Feb 2000 MWR)

Remedy : adding noise at data assimilation time Integrate deterministic model forward to next analysis time. Then add noise to deterministic forecasts consistent with Q.

Remedy : adding noise at data assimilation time (cont’d) Forming proper covariance model Q important Mitchell and Houtekamer: estimate parameters of Q from data assimilation innovation statistics. (MWR, Feb 2000) Hamill and Whitaker: estimate from differences between lo-res, hi-res simulations (

Remedy : multi-model ensemble? Integrate different members with different models/different parameterizations. Initial testing shows covariances from such an ensemble are highly unbalanced (M. Buehner, RPN Canada, personal communication).

Applications of ensemble filter: adaptive observations Can predict the expected reduction in analysis or forecast error from inserting a rawinsonde at a particular location. Source: Hamill and Snyder, MWR, June See also Bishop et al., MWR, March 2001

Applications of ensemble filter: analysis-error covariance singular vectors Source: Hamill et al., MWR, Aug 2003

Applications of ensemble filters: setting initial perturbations for ensembles Idea: Use ETKF to set structure of initial condition perturbations around 3D-Var analysis Source: Wang and Bishop, JAS, May 2003 Matrix of ensemble perturbations formed as in EnKF Goal: form transformation matrix T so as to get analysis perturbations.

Benefits of ETKF vs. Bred They grow faster. Ensemble mean forecast error is lower Stronger spread-skill relationship Source: Wang and Bishop, JAS, May 2003

Conclusions Ensemble data assimilation literature growing rapidly because of –Great results in simple models –Coding ease (no tangent linear/adjoint) –Conceptual appeal (more gracefully handles some nonlinearity) However: –Computationally expensive –Requires careful modeling of observation, forecast-error covariances to exploit benefits. Exploration of these issues relatively new. Acknowledgments: Chris Snyder, Jeff Whitaker, Jeff Anderson, Peter Houtekamer For more information – –