A kernel-density based ensemble filter applicable (?) to high-dimensional systems Tom Hamill NOAA Earth System Research Lab Physical Sciences Division.

Slides:



Advertisements
Similar presentations
1 What constrains spread growth in forecasts initialized from ensemble filters? Tom Hamill (& Jeff Whitaker) NOAA Earth System Research Lab Boulder, Colorado,
Advertisements

Data-Assimilation Research Centre
Biointelligence Laboratory, Seoul National University
Kriging.
Probability Distributions CSLU 2850.Lo1 Spring 2008 Cameron McInally Fordham University May contain work from the Creative Commons.
Initialization Issues of Coupled Ocean-atmosphere Prediction System Climate and Environment System Research Center Seoul National University, Korea In-Sik.
Effects of model error on ensemble forecast using the EnKF Hiroshi Koyama 1 and Masahiro Watanabe 2 1 : Center for Climate System Research, University.
1 What constrains spread growth in forecasts initialized from ensemble Kalman filters? Tom Hamill (& Jeff Whitaker) NOAA Earth System Research Lab Boulder,
June 17th 2003 Spruce VI1 On the use of statistics in complex weather and climate models Andreas Hense Meteorological Institute University Bonn.
Representing Model Error in Ensemble DA Chris Snyder (NCAR) NCAR is supported by the National Science Foundation.
Empirical Localization of Observation Impact in Ensemble Filters Jeff Anderson IMAGe/DAReS Thanks to Lili Lei, Tim Hoar, Kevin Raeder, Nancy Collins, Glen.
Balance in the EnKF Jeffrey D. Kepert Centre for Australian Weather and Climate Research A partnership between the Australian Bureau of Meteorology and.
Review of “Covariance Localization” in Ensemble Filters Tom Hamill NOAA Earth System Research Lab, Boulder, CO NOAA Earth System Research.
ASSIMILATION of RADAR DATA at CONVECTIVE SCALES with the EnKF: PERFECT-MODEL EXPERIMENTS USING WRF / DART Altuğ Aksoy National Center for Atmospheric Research.
Advanced data assimilation methods- EKF and EnKF Hong Li and Eugenia Kalnay University of Maryland July 2006.
Introduction to Numerical Weather Prediction and Ensemble Weather Forecasting Tom Hamill NOAA-CIRES Climate Diagnostics Center Boulder, Colorado USA.
A comparison of hybrid ensemble transform Kalman filter(ETKF)-3DVAR and ensemble square root filter (EnSRF) analysis schemes Xuguang Wang NOAA/ESRL/PSD,
Binary Variables (1) Coin flipping: heads=1, tails=0 Bernoulli Distribution.
1 CE 530 Molecular Simulation Lecture 7 David A. Kofke Department of Chemical Engineering SUNY Buffalo
1 Statistical Mechanics and Multi- Scale Simulation Methods ChBE Prof. C. Heath Turner Lecture 11 Some materials adapted from Prof. Keith E. Gubbins:
PATTERN RECOGNITION AND MACHINE LEARNING
The impact of localization and observation averaging for convective-scale data assimilation in a simple stochastic model Michael Würsch, George C. Craig.
Ensemble Data Assimilation and Uncertainty Quantification Jeffrey Anderson, Alicia Karspeck, Tim Hoar, Nancy Collins, Kevin Raeder, Steve Yeager National.
EnKF Overview and Theory
Particle Filters in high dimensions Peter Jan van Leeuwen and Mel Ades Data-Assimilation Research Centre DARC University of Reading Lorentz Center 2011.
Nonlinear Data Assimilation and Particle Filters
Eidgenössisches Departement des Innern EDI Bundesamt für Meteorologie und Klimatologie MeteoSchweiz Statistical Characteristics of High- Resolution COSMO.
ISDA 2014, Feb 24 – 28, Munich 1 Impact of ensemble perturbations provided by convective-scale ensemble data assimilation in the COSMO-DE model Florian.
Computational Issues: An EnKF Perspective Jeff Whitaker NOAA Earth System Research Lab ENIAC 1948“Roadrunner” 2008.
1 ESTIMATING THE STATE OF LARGE SPATIOTEMPORALLY CHAOTIC SYSTEMS: WEATHER FORECASTING, ETC. Edward Ott University of Maryland Main Reference: E. OTT, B.
Exploring sample size issues for 6-10 day forecasts using ECMWF’s reforecast data set Model: 2005 version of ECMWF model; T255 resolution. Initial Conditions:
MPO 674 Lecture 20 3/26/15. 3d-Var vs 4d-Var.
1 Statistical Distribution Fitting Dr. Jason Merrick.
Probabilistic Robotics Bayes Filter Implementations.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Deterministic vs. Random Maximum A Posteriori Maximum Likelihood Minimum.
ECE 8443 – Pattern Recognition LECTURE 07: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Class-Conditional Density The Multivariate Case General.
Mixture Models, Monte Carlo, Bayesian Updating and Dynamic Models Mike West Computing Science and Statistics, Vol. 24, pp , 1993.
Dusanka Zupanski And Scott Denning Colorado State University Fort Collins, CO CMDL Workshop on Modeling and Data Analysis of Atmospheric CO.
SUPA Advanced Data Analysis Course, Jan 6th – 7th 2009 Advanced Data Analysis for the Physical Sciences Dr Martin Hendry Dept of Physics and Astronomy.
10/18/2011 Youngsun Jung and Ming Xue CAPS/OU with help from Tim Supinie.
Data Assimilation Using Modulated Ensembles Craig H. Bishop, Daniel Hodyss Naval Research Laboratory Monterey, CA, USA September 14, 2009 Data Assimilation.
A unifying framework for hybrid data-assimilation schemes Peter Jan van Leeuwen Data Assimilation Research Center (DARC) National Centre for Earth Observation.
ECE 8443 – Pattern Recognition LECTURE 10: HETEROSCEDASTIC LINEAR DISCRIMINANT ANALYSIS AND INDEPENDENT COMPONENT ANALYSIS Objectives: Generalization of.
Applications of optimal control and EnKF to Flow Simulation and Modeling Florida State University, February, 2005, Tallahassee, Florida The Maximum.
MODEL ERROR ESTIMATION EMPLOYING DATA ASSIMILATION METHODOLOGIES Dusanka Zupanski Cooperative Institute for Research in the Atmosphere Colorado State University.
Data assimilation and forecasting the weather (!) Eugenia Kalnay and many friends University of Maryland.
Example: Bioassay experiment Problem statement –Observations: At each level of dose, 5 animals are tested, and number of death are observed.
ECE-7000: Nonlinear Dynamical Systems Overfitting and model costs Overfitting  The more free parameters a model has, the better it can be adapted.
Boosted Particle Filter: Multitarget Detection and Tracking Fayin Li.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 12: Advanced Discriminant Analysis Objectives:
A Random Subgrouping Scheme for Ensemble Kalman Filters Yun Liu Dept. of Atmospheric and Oceanic Science, University of Maryland Atmospheric and oceanic.
Pg 1 SAMSI UQ Workshop; Sep Ensemble Data Assimilation and Uncertainty Quantification Jeff Anderson National Center for Atmospheric Research.
1 Tom Edgar’s Contribution to Model Reduction as an introduction to Global Sensitivity Analysis Procedure Accounting for Effect of Available Experimental.
Future Directions in Ensemble DA for Hurricane Prediction Applications Jeff Anderson: NCAR Ryan Torn: SUNY Albany Thanks to Chris Snyder, Pavel Sakov The.
École Doctorale des Sciences de l'Environnement d’ Î le-de-France Année Modélisation Numérique de l’Écoulement Atmosphérique et Assimilation.
Ch 1. Introduction Pattern Recognition and Machine Learning, C. M. Bishop, Updated by J.-H. Eom (2 nd round revision) Summarized by K.-I.
Chapter 15: Exploratory data analysis: graphical summaries CIS 3033.
Jeffrey Anderson, NCAR Data Assimilation Research Section
Data Assimilation Research Testbed Tutorial
Methods for dealing with spurious covariances arising from small samples in ensemble data assimilation Jeff Whitaker NOAA Earth.
Data Assimilation Theory CTCD Data Assimilation Workshop Nov 2005
Verifying and interpreting ensemble products
Precipitation Products Statistical Techniques
Jeff Anderson, NCAR Data Assimilation Research Section
Nathalie Voisin, Andy W. Wood and Dennis P. Lettenmaier
2. University of Northern British Columbia, Prince George, Canada
Parametric Methods Berlin Chen, 2005 References:
Ch 3. Linear Models for Regression (2/2) Pattern Recognition and Machine Learning, C. M. Bishop, Previously summarized by Yung-Kyun Noh Updated.
Sarah Dance DARC/University of Reading
Presentation transcript:

A kernel-density based ensemble filter applicable (?) to high-dimensional systems Tom Hamill NOAA Earth System Research Lab Physical Sciences Division NOAA Earth System Research Laboratory Presentation at 91st AMS Annual, Seattle, WA, 15th IOAS-AOLS 1

“Research is the process of going up alleys to see if they are blind.” Marston Bates 2

What to do when in ensemble filters when prior is obviously non-Gaussian? Cycled EnKF can create non-Gaussian states, especially when (a) ensemble size is large, and (b) when the forecast model contains significant nonlinearity. See Lawson and Hansen, MWR, 2004; Mitchell and Houtekamer, MWR, 2009 and Anderson, MWR, Example: data assimilation with EAKF in simple dynamical system where x t+1 =x t (x t + αx t |x t |) where α=0.2, n=20, obs~ N(0,1) from Jeff Anderson, MWR,

Gaussian distributions easy. Most common data assimilation updates are well behaved when prior and observation are normally distributed, as shown here, with an analytical solution for this 1-D problem. 4

Estimating prior as fully non-Gaussian may work when state dimension is small. Here ensemble is used to estimate probability density in 2-D, which is then Bayesian updated to an observation in one component. Approaches such as these are computationally and scientifically impractical for large dimensional systems, e.g., “curse of dimensionality.” from Hamill, chapter 6 in “Predictability of Weather and Climate” 5

Flavors of non-Gaussianity (1)Ensemble non-Gaussian, but very high correlation between state components, i.e, effectively non-Gaussian only in 1 direction. (potentially solveable, & intended focus of this talk) (2) Ensemble non-Gaussian, curved attractor (well beyond my current capabilities & available computational resources) 6

Two insights of Jeff Anderson relevant to non-Gaussian data assimilation (of the first type) (1)Serial ensemble filters can split the update into two steps (Anderson, MWR, 2003) : (a)Update prior at observation location to the observation (b) Regress increments to the rest of the state to the updated observation prior. 7

Insight 2: only in updating the observation prior, relax the Gaussianity assumption. “Rank Histogram Filter” -- a probability mass of 1 / (n+1) is assigned between each ensemble member observation prior. Given the observation likelihood (dashed), piecewise construct a product of the prior and likelihood. For i th sorted observation prior, determine the value associated with the posterior CDF at i / (n+1). This replaces step (1) of Anderson (2003). From Anderson, MWR, (2010). 8

Anderson’s tests with Lorenz ‘96 model F=8.0; fixed localization “half width” of 12 grid points. Observations Adaptive spatially varying inflation following Anderson, Tellus, Note poor performance of both RHF and EnKF when ensemble size is 10 (see Whitaker and Hamill MWR 2002), poor performance of EAKF when ensemble size is large (Lawson and Hansen, MWR 2003). 9 Make a non-Gaussian filter as good as EAKF at small ens sizes? Still as good as EnKF at large ens sizes?

Possible errors introduced with rank histogram filter (RHF) Ensembles may, due to sampling variability, have large or small deviations between members. Are there consequences of having 1/(n+1) probability between each? same 1/(n+1) probability mass between both? 10

Surprisingly, the gap between the 1st and 2nd ensemble member in the prior is obliterated with the rank histogram filter. 11

Kernel density approach? Model observation prior with kernel density estimation techniques. source: Wikipedia, “kernel density estimation.” 12

Example of observation prior update with “kernel density filter” (KDF) If prior ensemble is significantly non-Gaussian (as determined through Kolmogorov-Smirnov test) then replace standard ensemble filter’s obs prior update with KDF update. Model the observation prior (black lines, panel a) with kernel density estimation. Construct posterior as product of prior and likelihood (panel a). Generate posterior members by finding the ensemble members’ quantiles in the prior cdf. Posterior members are the values associated with those quantiles in posterior cdf (panel b). 13

When to apply KDF It’s much more expensive than EnKF, EnSRF, EnAF for updating obs prior. My choice: use only when observation prior is statistically significantly non-Gaussian (Kolmogorov-Smirnov test, α=0.05) Departures from Gaussianity more commonly determined to be significant with large ensemble sizes than with small. 14

Testing in Lorenz ’96 model Classic L96 model, 40 variables, perfect model, F=8.0, dt = 0.05, cycle for 10 years after spinup. Test EnKF, EnSRF, EnAF/KDF, EnAF/RHF over range of localization length scales, ensemble sizes. Covariance inflation : /nanals

Lorenz ‘96, perfect model (obs σ=5.0, 6 h between obs, obs every grid point) 16 (1)RHF/EAKF and KDF/EAKF not employed at smallest ensemble sizes since departures from Gaussianity not detected. Same error as EnSRF. (2) Surprisingly, benefit of RHF/EAKF & KDF/EAKF over EnSRF is at moderate ensemble sizes, with degradation at large ensemble sizes.

Lorenz ‘96, perfect model (obs σ=5.0, 6 h between obs, obs every grid point) 17 EnKF detects non-Gaussianity less frequently than deterministic filters. For typical members, non-Gaussian conditions virtually never diagnosed.

Lawson & Hansen’s insight using Ikeda Map 18 black dots = prior ensemble grey dots = posterior ensemble EnKF & effect of perturbed observations creates more normally distributed posterior, while EnSRF compresses but keeps curved prior shape. (EnAF/KDF or EnAF/RHF would do something similar). Lawson & Hansen, MWR, 2004.

Tests of KDF in global primitive equation model Tested in 2-level dry global PE model recently used in Hamill and Whitaker (2011 MWR). Same uniform observation network. Imperfect model assumption. Bottom line: effectively no change from using EnSRF data assimilation to using EnSRF/KDF data assimilation. – Reason is that KDF is virtually never invoked, since only invoked when prior is significantly non-Gaussian. EnSRF consistently used. Admittedly, it’s still a toy model w/o moisture, other complicating aspects. 19

Why is non-Gaussianity actually more rare than one might think? (1) In practical NWP applications, we often account for model error or sampling variability through additive noise with Gaussian properties. P b = MP a M T + Q x b = Mx a + e, e ~ N(0,Q) 20

Why is non-Gaussianity actually more rare than one might think? (2) Ensemble forecast members in more complex models may project onto several modes, some growing, some decaying. This may randomize the perturbation structures. 21 old perturbation, slowly decaying overall perturbation, superposition of old decaying structures and new growing ones new, rapidly growing perturbation

Conclusions Many previous researchers have dashed themselves to pieces trying to surf the reef that is non-Gaussian data assimilation (me too). Some oldies but goodies, like perturbed-obs EnKF & deterministic filters with noise to account for model and sampling error, continue to be hard to beat for NWP applications. Possible increased relevance of non-Gaussian techniques as ensemble filters begin to deal with moisture-related variables. 22

KDF technical detail. Q: How does one determine the kernel type and width used? A: Gaussian kernel used, but may not be a critical detail. Optimal width is a function of ensemble size, smaller for large sizes. The optimal width estimated by repeatedly: – creating an ensemble drawn from standard normal. – choosing kernel width, creating pdf estimate. – evaluating the integrated square error (ISE) of the estimated to the analytical cdf. – finally, for a given ensemble size, choose the width “w” that has the lowest average integrated square error. – width that’s actually used for an ensemble with a spread of s is s*w. 23