May 30, 2003 Tony Eckel, Eric Grimit, and Cliff Mass UW Atmospheric Sciences This research was supported by the DoD Multidisciplinary University Research.

Slides:



Advertisements
Similar presentations
Chapter 13 – Weather Analysis and Forecasting
Advertisements

ECMWF long range forecast systems
A Brief Guide to MDL's SREF Winter Guidance (SWinG) Version 1.0 January 2013.
Mesoscale Probabilistic Prediction over the Northwest: An Overview Cliff Mass Adrian Raftery, Susan Joslyn, Tilmann Gneiting and others University of Washington.
Gridded OCF Probabilistic Forecasting For Australia For more information please contact © Commonwealth of Australia 2011 Shaun Cooper.
Statistical Postprocessing of Weather Parameters for a High-Resolution Limited-Area Model Ulrich Damrath Volker Renner Susanne Theis Andreas Hense.
The University of Washington Mesoscale Short-Range Ensemble System Eric P. Grimit, F. Anthony Eckel, Richard Steed, Clifford F. Mass University of Washington.
Mesoscale Probabilistic Prediction over the Northwest: An Overview Cliff Mass University of Washington.
Mesoscale Probabilistic Prediction over the Northwest: An Overview Cliff Mass University of Washington.
Description and Preliminary Evaluation of the Expanded UW S hort R ange E nsemble F orecast System Maj. Tony Eckel, USAF University of Washington Atmospheric.
Performance Characteristics of a Pseudo-operational Ensemble Kalman Filter April 2006, EnKF Wildflower Meeting Greg Hakim & Ryan Torn University of Washington.
MOS Developed by and Run at the NWS Meteorological Development Lab (MDL) Full range of products available at:
Overview of the Pacific Northwest Environmental Prediction System.
Statistics, data, and deterministic models NRCSE.
Brian Ancell, Cliff Mass, Gregory J. Hakim University of Washington
EG1204: Earth Systems: an introduction Meteorology and Climate Lecture 7 Climate: prediction & change.
Transitioning unique NASA data and research technologies to the NWS 1 Evaluation of WRF Using High-Resolution Soil Initial Conditions from the NASA Land.
Evaluation of a Mesoscale Short-Range Ensemble Forecasting System over the Northeast United States Matt Jones & Brian A. Colle NROW, 2004 Institute for.
22 May :30 PM General Examination Presentation Toward Short-Range Ensemble Prediction of Mesoscale Forecast Skill Eric P. Grimit University of Washington.
Ensembles and Probabilistic Forecasting. Probabilistic Prediction Because of forecast uncertainties, predictions must be provided in a probabilistic framework,
MOS Performance MOS significantly improves on the skill of model output. National Weather Service verification statistics have shown a narrowing gap between.
19 September :15 PM Ensemble Weather Forecasting Workshop; Val-Morin, QC Canada Toward Short-Range Ensemble Prediction of Mesoscale Forecast Error.
Introduction to Numerical Weather Prediction and Ensemble Weather Forecasting Tom Hamill NOAA-CIRES Climate Diagnostics Center Boulder, Colorado USA.
The Expanded UW SREF System and Statistical Inference STAT 592 Presentation Eric Grimit 1. Description of the Expanded UW SREF System (How is this thing.
Implementation and Evaluation of a Mesoscale Short-Range Ensemble Forecasting System Over the Pacific Northwest Eric P. Grimit and Clifford F. Mass Department.
Ensembles and The Future of Mesoscale Prediction Cliff Mass, University of Washington Tony Eckel, USAF and Univ. of WA.
30 January :00 NWSFO-Seattle An Update on Local MM5 Products at NWSFO-Seattle Eric P. Grimit SCEP NOAA/NWS-Seattle Ph.D. Candidate, University of.
Mesoscale Deterministic and Probabilistic Prediction over the Northwest: An Overview Cliff Mass University of Washington.
Ensemble Post-Processing and it’s Potential Benefits for the Operational Forecaster Michael Erickson and Brian A. Colle School of Marine and Atmospheric.
Effective Mesoscale, Short-Range Ensemble Forecasting Tony Eckel** and Clifford F. Mass Presented By: Eric Grimit University of Washington Atmospheric.
Chapter 13 – Weather Analysis and Forecasting. The National Weather Service The National Weather Service (NWS) is responsible for forecasts several times.
“1995 Sunrise Fire – Long Island” Using an Ensemble Kalman Filter to Explore Model Performance on Northeast U.S. Fire Weather Days Michael Erickson and.
Performance of the MOGREPS Regional Ensemble
SRNWP workshop - Bologne Short range ensemble forecasting at Météo-France status and plans J. Nicolau, Météo-France.
Forecasting and Numerical Weather Prediction (NWP) NOWcasting Description of atmospheric models Specific Models Types of variables and how to determine.
Probabilistic Prediction. Uncertainty in Forecasting All of the model forecasts I have talked about reflect a deterministic approach. This means that.
Estimation & Value of Ambiguity in Ensemble Forecasts Tony Eckel National Weather Service Office of Science and Technology, Silver Spring, MD Mark Allen.
How can LAMEPS * help you to make a better forecast for extreme weather Henrik Feddersen, DMI * LAMEPS =Limited-Area Model Ensemble Prediction.
Preliminary Results of Global Climate Simulations With a High- Resolution Atmospheric Model P. B. Duffy, B. Govindasamy, J. Milovich, K. Taylor, S. Thompson,
23 June :30 PM Session 2: Mesoscale Predictability I; 10th Mesoscale Conference; Portland, OR Toward Short-Range Ensemble Prediction of Mesoscale.
It Never Rains But It Pours: Modeling Mixed Discrete- Continuous Weather Phenomena J. McLean Sloughter This work was supported by the DoD Multidisciplinary.
1 An overview of the use of reforecasts for improving probabilistic weather forecasts Tom Hamill NOAA / ESRL, Physical Sciences Div.
Probabilistic Forecasting. pdfs and Histograms Probability density functions (pdfs) are unobservable. They can only be estimated. They tell us the density,
Short-Range Ensemble Prediction System at INM José A. García-Moya SMNT – INM 27th EWGLAM & 12th SRNWP Meetings Ljubljana, October 2005.
Model Post Processing. Model Output Can Usually Be Improved with Post Processing Can remove systematic bias Can produce probabilistic information from.
Plans for Short-Range Ensemble Forecast at INM José A. García-Moya SMNT – INM Workshop on Short Range Ensemble Forecast Madrid, October,
. Outline  Evaluation of different model-error schemes in the WRF mesoscale ensemble: stochastic, multi-physics and combinations thereof  Where is.
Probabilistic Prediction Cliff Mass University of Washington.
Use of Mesoscale Ensemble Weather Predictions to Improve Short-Term Precipitation and Hydrological Forecasts Michael Erickson 1, Brian A. Colle 1, Jeffrey.
Insights from CMC BAMS, June Short Range The SPC Short-Range Ensemble Forecast (SREF) is constructed by post-processing all 21 members of the NCEP.
18 September 2009: On the value of reforecasts for the TIGGE database 1/27 On the value of reforecasts for the TIGGE database Renate Hagedorn European.
Ensembles and Probabilistic Prediction. Uncertainty in Forecasting All of the model forecasts I have talked about reflect a deterministic approach. This.
Nathalie Voisin 1, Florian Pappenberger 2, Dennis Lettenmaier 1, Roberto Buizza 2, and John Schaake 3 1 University of Washington 2 ECMWF 3 National Weather.
Verification of ensemble precipitation forecasts using the TIGGE dataset Laurence J. Wilson Environment Canada Anna Ghelli ECMWF GIFS-TIGGE Meeting, Feb.
DOWNSCALING GLOBAL MEDIUM RANGE METEOROLOGICAL PREDICTIONS FOR FLOOD PREDICTION Nathalie Voisin, Andy W. Wood, Dennis P. Lettenmaier University of Washington,
VERIFICATION OF A DOWNSCALING SEQUENCE APPLIED TO MEDIUM RANGE METEOROLOGICAL PREDICTIONS FOR GLOBAL FLOOD PREDICTION Nathalie Voisin, Andy W. Wood and.
Toward an Effective Short-Range Ensemble Forecast System October 17, 2003 Maj Tony Eckel, USAF University of Washington Atmospheric Sciences Department.
Improving Numerical Weather Prediction Using Analog Ensemble Presentation by: Mehdi Shahriari Advisor: Guido Cervone.
Update on the Northwest Regional Modeling System 2013
University of Washington Ensemble Systems for Probabilistic Analysis and Forecasting Cliff Mass, Atmospheric Sciences University of Washington.
Overview of Deterministic Computer Models
Ensembles and Probabilistic Prediction
Verifying and interpreting ensemble products
Mesoscale Probabilistic Prediction over the Northwest: An Overview
The Stone Age Prior to approximately 1960, forecasting was basically a subjective art, and not very skillful. Observations were sparse, with only a few.
Nathalie Voisin, Andy W. Wood and Dennis P. Lettenmaier
Post Processing.
The Importance of Reforecasts at CPC
New Developments in Aviation Forecast Guidance from the RUC
Verification of Tropical Cyclone Forecasts
Presentation transcript:

May 30, 2003 Tony Eckel, Eric Grimit, and Cliff Mass UW Atmospheric Sciences This research was supported by the DoD Multidisciplinary University Research Initiative (MURI) program administered by the Office of Naval Research under Grant N

Overview  Review Ensemble Forecasting Theory and Introduce UW’s SREFs  Discuss Results of Model Deficiencies on SREF - Need for Bias Correction - Impact on ensemble spread - Impact on probabilistic forecasts skill  Conclusions

T The true state of the atmosphere exists as a single point in phase space that we never know exactly. A point in phase space completely describes an instantaneous state of the atmosphere. For a model, a point is the vector of values for all parameters (pres, temp, etc.) at all grid points at one time. An analysis produced to run a model like the eta is in the neighborhood of truth. The complete error vector is unknown, but we have some idea of its structure and magnitude. e Chaos drives apart the forecast and true trajectories…predictability error growth. EF can predicted the error magnitude and give a “probabilistic cloud” of forecasts. 12h forecast 36h forecast 24h forecast 48h forecast T 48h verification phase space

e a u c j t g n M T T Analysis Region 48h forecast Region 12h forecast 36h forecast 24h forecast Diagram for: PME ACME core or ACME core+ phase space Plug each IC into the MM5 to create an ensemble of mesoscale forecasts (cloud of future states encompassing truth). 1) Reveal uncertainty in forecast 2) Reduce error by averaging M 3) Yield probabilistic information

e a u c j t g n c T M T Analysis Region 48h forecast Region phase space ACME’s Centroid

e n a c u t g T j M T Analysis Region 48h Forecast Region e a u c j t g n c phase space ACME’s Mirrored Members

FP = 93% Parameter Threshold (EX: precip > 0.5”) FP = ORF = 72% Frequency Initial State Forecast Probability from an Ensemble EF provides an estimate (histogram) of truth’s Probability Density Function (red curve). In a large, well-tuned EF, Forecast Probability (FP) = Observed Relative Frequency (ORF) 24hr Forecast State48hr Forecast State Frequency In practice, things get wacky from Under-sampling of the PDF (too few ensemble members) Poor representation of initial uncertainty Model deficiencies -- Model bias causes a shift in the estimated mean -- Sharing of model errors between EF members leads to reduced variance EF’s estimated PDF does not match truth’s PDF, and Fcst Prob  Obs Rel Freq

UW’s Ensemble of Ensembles # of EF Initial Forecast Forecast Name Members Type Conditions Model(s) Cycle Domain ACME 17SMMA 8 Ind. Analyses, “Standard” 00Z 36km, 12km 1 Centroid, MM5 8 Mirrors ACME core 8SMMA Independent “Standard” 00Z 36km, 12km Analyses MM5 ACME core+ 8PMMA “ “ 8 MM5 00Z 36km, 12km variations PME 8 MMMA “ “ 8 “native” 00Z, 12Z 36km large-scale Homegrown Imported ACME: Analysis-Centroid Mirroring Ensemble PME: Poor Man’s Ensemble MM5: PSU/NCAR Mesoscale Modeling System Version 5 SMMA: Single Model Multi-Analysis PMMA: Perturbed-model Multi-Analysis MMMA: Multi-model Multi-Analysis

Resolution ( 45  N ) Objective Abbreviation/Model/Source Type Computational Distributed Analysis gfs, Global Forecast System, SpectralT254 / L641.0  / L14 SSI National Centers for Environmental Prediction~55km~80km3D Var cmcg, Global Environmental Multi-scale (GEM),SpectralT199 / L  / L113D Var Canadian Meteorological Centre ~70km ~100km eta, Eta limited-area mesoscale model, Finite12km / L60 90km / L37SSI National Centers for Environmental Prediction Diff.3D Var gasp, Global AnalysiS and Prediction model,SpectralT239 / L291.0  / L11 3D Var Australian Bureau of Meteorology~60km~80km jma, Global Spectral Model (GSM),SpectralT106 / L  / L13OI Japan Meteorological Agency~135km~100km ngps, Navy Operational Global Atmos. Pred. System,SpectralT239 / L301.0  / L14OI Fleet Numerical Meteorological & Oceanographic Cntr. ~60km~80km tcwb, Global Forecast System,SpectralT79 / L181.0  / L11 OI Taiwan Central Weather Bureau~180km~80km ukmo, Unified Model, Finite5/6  5/9  /L30same / L123D Var United Kingdom Meteorological Office Diff.~60km “Native” Models/Analyses of the PME

Design of ACME core+ 8  5  3  2  5  3  2  2  8  8 = 921,600 Total possible combinations:

 Total of 129, 48-h forecasts (Oct 31, 2002 – Mar 28, 2003) all initialized at 00z - Missing forecast case days are shaded  Parameters: - 36 km Domain: Mean Sea Level Pressure (MSLP), 500mb Geopotential Height (Z500) - 12 km Domain: Wind 10m (WS10), Temperature at 2m (T2) Research Dataset 36 km Domain (151  127) 12 km Domain (101  103)  Verification: - 36 km Domain: centroid analysis (mean of 8 independent analyses, available at 12h increments) - 12 km Domain: ruc20 analysis (NCEP 20 km mesoscale analysis, available at 3h increments) NovemberDecemberJanuary February March

cmcg* The ACME Process STEP 1: Calculate best guess for truth (the centroid) by averaging all analyses. STEP 2: Find error vector in model phase space between one analysis and the centroid by differencing all state variables over all grid points. STEP 3: Make a new IC by mirroring that error about the centroid. cmcg C cmcg* Sea Level Pressure (mb) ~1000 km cent 170°W 165°W 160°W 155°W 150°W 145°W 140°W 135°W eta ngps tcwb gasp avn ukmo cmcg

MSLP analysis south of the Aleutians at 00Z on Jan 16, 2003 tcwb centroid centroid + (centroid  tcwb)

bias correction…

Overview The two flavors of model deficiencies play a big role in SREF: 1) Systematic: Model bias is a significant fraction of forecast error and must be removed. 2) Stochastic: Random model errors significantly increase uncertainty and must be accounted for.  Bias Correction: A simple method gives good results  Model Error*: Impact on ensemble spread  Final Results: Impact of both on probabilistic forecasts skill * bias-corrected

 Often difficult to completely remove bias within a model’s code  Systematic but complex; involving numerics, parameterizations, resolution, etc.  Depend upon weather regime (time of day, surface characteristics, stability, moisture, etc.)  Cheaper and easier to remove bias through post-processing  Sophisticated routines such as MOS require long training periods (years)  The bulk of bias can be removed with the short term mean error Need for Bias Removal NGPS Forecast vs Analysis Data Info  Single model grid point in eastern WA  Verification: centroid analysis  70 forecasts (Nov 25, 2002 – Feb 7, 2003)  Lead time = 24h GASP Forecast vs AnalysisGFS Forecast vs Analysis GFS-MM5 Forecast vs Analysis

Training Period Bias-corrected Forecast Period Training Period Bias-corrected Forecast Period Training Period Bias-corrected Forecast Period Gridded Bias Removal N number of forecast cases (14) f i,j,t forecast at grid point (i, j ) and lead time (t) o i,j verifying observation For the current forecast cycle: 1) Calculate bias at every grid point and lead time using previous 2 weeks’ forecasts 2) Post-process current forecast to correct for bias: f i,j,t bias-corrected forecast at grid point (i, j ) and lead time (t) * NovemberDecemberJanuary February March

Spatial and Temporal Dependence of Bias GFS-MM5 MSLP Bias at f24 Common Bias Forecast Error > 1 too low < 1 too high

Spatial and Temporal Dependence of Bias GFS-MM5 MSLP Bias at f36 Common Bias Forecast Error > 1 too low < 1 too high

Bias Correction Results biased bias-corrected PME

ACME core Bias Correction Results biased bias-corrected

ACME core+ Bias Correction Results biased bias-corrected

Lead Time (hours) Verification Rank Probability Verification Rank Histogram Record of where verification fell (i.e., its rank) among the ordered ensemble members: Flat Well calibrated EF (truth’s PDF matches EF PDF) U’d Under-dispersive EF (truth “gets away” quite often) Humped Over-dispersive EF

Lead Time (hours) Verification Rank Probability Verification Rank Histogram Record of where verification fell (i.e., its rank) among the ordered ensemble members: Flat Well calibrated EF (truth’s PDF matches EF PDF) U’d Under-dispersive EF (truth “gets away” quite often) Humped Over-dispersive EF

Model Error Impact on ensemble spread …

Ensemble Dispersion (MSLP) Analysis Error Error Growth due to Analysis Error Ensemble Variance (mb 2 ) Error Growth due to Model Error EF Mean’s MSE adjusted by n / n+1 to account for small sample size MSE of EF MEAN

Lead Time (hours) Verification Rank Probability Verification Rank Histogram Record of where verification fell (i.e., its rank) among the ordered ensemble members: Flat Well calibrated EF (truth’s PDF matches EF PDF) U’d Under-dispersive EF (truth “gets away” quite often) Humped Over-dispersive EF

Lead Time (hours) Verification Rank Probability Verification Rank Histogram Record of where verification fell (i.e., its rank) among the ordered ensemble members: Flat Well calibrated EF (truth’s PDF matches EF PDF) U’d Under-dispersive EF (truth “gets away” quite often) Humped Over-dispersive EF

Lead Time (hours) Verification Rank Probability Verification Rank Histogram Record of where verification fell (i.e., its rank) among the ordered ensemble members: Flat Well calibrated EF (truth’s PDF matches EF PDF) U’d Under-dispersive EF (truth “gets away” quite often) Humped Over-dispersive EF

Impact of both on probabilistic forecasts skill

Explain probabilistic forecast verification

P(MSLP < 1001mb) by uniform ranks method, 36h lead time, Sub-domain A Reliability Diagram Comparison PME ACME core Sample Climatology

36km Verification Sub-domain A

~6hr improvement by bias correction ~11hr improvement by multi-model diversity and “global” error growth Skill vs. Lead Time (Sub-domain A)

36km Verification Sub-domain B

Skill vs. Lead Time (all bias –corrected) 36km Sub-domain A (2/3 ocean) P(MSLP < 1001mb) Sample Climatology  23% 36km Sub-domain B (mostly land) P(MSLP < 1011mb) Sample Climatology  20% ~11hr improvement by PME ~22hr improvement by PME ~3hr improvement by ACME core+

Conclusions Caveats: - Consider non-optimal ICs and small EF size? ( still fair comparison between PME and ACME core ) - What about higher skill of PME members? ( not so dramatic after bias correction ) - Does higher resolution of MM5 make comparison unfair? ( fitting to lower res. would decrease  2 ) Why bother with ACME core ? PME is certainly more skilled at the synoptic level, but has little to no mesoscale info. Should these conclusions hold true for mesoscale? YES! Model deficiencies for surface variables (precip, winds, temperature) can be even stronger, so the effect on SREF may be even greater. Demonstrating that is now the focus of my research… P(precip > 0.25” in 6hr) An ensemble’s skill is dramatically improved by: 1) Correcting model bias 2) Accounting for model uncertainty

UW’s Ensemble of Ensembles # of EF Initial Forecast Forecast Name Members Type Conditions Model(s) Cycle Domain ACME 17SMMA 8 Ind. Analyses, “Standard” 00Z 36km, 12km 1 Centroid, MM5 8 Mirrors ACME core 8SMMA Independent “Standard” 00Z 36km, 12km Analyses MM5 ACME core+ 8PMMA “ “ 8 MM5 00Z 36km, 12km variations PME 8 MMMA“ “ 8 “native” 00Z, 12Z 36km large-scale ACNE 9hybrid?8 Ind. Analyses, 9 MM5 00Z, 12Z 36km, 12km MMMA 1 Centroid variations PMMA ACNE: Analysis-Centroid Nudged Ensemble SMMA: Single Model Multi-Analysis PMMA: Perturbed-model Multi-Analysis MMMA: Multi-model Multi-Analysis Proposed

?

Skill Score (SS ) Details (reliability) (resolution) (uncertainty) Brier Score Brier Skill Score n: number of data pairs FP i : forecast probability {0.0…1.0} ORF i : observation {0.0 = yes, 1.0 = no} M : number of probability bins (normally 11) N : number of data pairs in the bin FP * i : binned forecast probability {0.0, 0.1,…1.0} ORF * i : observation for the bin {0.0 = yes, 1.0 = no} SC : sample climatology (total occurrences / total forecasts) Decomposed Brier Score (uses binned FP as in rel. diag.) Skill Score

FP = 77.1% For a certain threshold, say Ws  20kt, the FP is then simply the area under the PDF to the right (1  p value) Ws = { } Wind Speed (kt) Frequency Ideal Calculation of Forecast Probability (FP) Given a very large ensemble, a PDF could be found a grid point for any parameter (e.g., wind speed, Ws). Unfortunately, we work with very small ensembles so we can’t make a good estimate of the PDF. Plus, we often do not even know what PDF shape to fit. So we are forced to estimate FP by other means, for a set of Ws forecasts at a point such as: Note: These are random draws from the PDF above

FP = 7/8 = 87.5% FP = 7/9 + [ (21.1 – 20.0) / (21.1 – 16.5) ] * 1/9 = 80.4% /8 7/8 6/8 5/8 4/8 3/8 2/8 1/8 0/8 Democratic Voting FP Uniform Ranks FP 9/9 8/9 7/9 6/9 5/9 4/9 3/9 2/9 1/9 0/9 “pushes” FP towards the extreme values, so high FP is normally over-forecast and low FP is normally under-forecast. a continuous, more appropriate approximation. 20.0

FP = [ (1 – G CDF (50.0)) / (1 – G CDF (47.8)) ] * 1/9 = 8.5% a b fraction = a / b Uniform Ranks FP 9/9 8/9 7/9 6/9 5/9 4/9 3/9 2/9 1/9 0/9 FP When Threshold Falls in an Extreme Rank  0.0, -  50.0 Use the tail of a Gumbel PDF to approximate the fraction for the last rank.

FP = [ (1 – CDF(50.0)) / (1 – CDF(47.8)) ] * 0.17 = 13.0% Weighted Ranks FP Calibration by Weighted Ranks  0.0, -  50.0 Use the verification rank histogram from past cases to define non-uniform, “weighted ranks”. The ranks to sum up and fraction of the rank where the threshold falls are found the same way as with uniform ranks, but now the probability within each rank is the chance that truth will occur there.

Sample Climatology Skill Zone Uniform Ranks vs. Democratic Voting Data Info  P(MSLP < 1002mb)  Verification: centroid analysis  70 forecasts (Nov 25, 2002 – Feb 7, 2003)  Applied 2-week, running bias correction  36km, Outer Domain  Lead time = 48h UR DV

References