Probabilistic Prediction Cliff Mass University of Washington.

Slides:



Advertisements
Similar presentations
ECMWF long range forecast systems
Advertisements

Report of the Q2 Short Range QPF Discussion Group Jon Ahlquist Curtis Marshall John McGinley - lead Dan Petersen D. J. Seo Jean Vieux.
PRESENTS: FORECASTING FOR OPERATIONS AND DESIGN February 16 th 2011 – Aberdeen.
A Brief Guide to MDL's SREF Winter Guidance (SWinG) Version 1.0 January 2013.
Statistical post-processing using reforecasts to improve medium- range renewable energy forecasts Tom Hamill and Jeff Whitaker NOAA Earth System Research.
Mesoscale Probabilistic Prediction over the Northwest: An Overview Cliff Mass Adrian Raftery, Susan Joslyn, Tilmann Gneiting and others University of Washington.
Predictability and Chaos EPS and Probability Forecasting.
2012: Hurricane Sandy 125 dead, 60+ billion dollars damage.
Description and Preliminary Evaluation of the Expanded UW S hort R ange E nsemble F orecast System Maj. Tony Eckel, USAF University of Washington Atmospheric.
May 30, 2003 Tony Eckel, Eric Grimit, and Cliff Mass UW Atmospheric Sciences This research was supported by the DoD Multidisciplinary University Research.
Revolutions in Remote Sensing Greatly Enhanced Weather Prediction from the 1950s Through Today.
Introduction to Weather Forecasting Cliff Mass Department of Atmospheric Sciences University of Washington.
MOS Developed by and Run at the NWS Meteorological Development Lab (MDL) Full range of products available at:
EG1204: Earth Systems: an introduction Meteorology and Climate Lecture 7 Climate: prediction & change.
Transitioning unique NASA data and research technologies to the NWS 1 Evaluation of WRF Using High-Resolution Soil Initial Conditions from the NASA Land.
Ensemble Prediction in the U.S.. NCEP Global Ensemble System Begun in 1993 with the MRF (now GFS) First tried “lagged” ensembles as basis…using runs of.
Ensembles and Probabilistic Forecasting. Probabilistic Prediction Because of forecast uncertainties, predictions must be provided in a probabilistic framework,
MOS Performance MOS significantly improves on the skill of model output. National Weather Service verification statistics have shown a narrowing gap between.
Consortium Meeting June 3, Thanks Mike! Hit Rates.
Introduction to Numerical Weather Prediction and Ensemble Weather Forecasting Tom Hamill NOAA-CIRES Climate Diagnostics Center Boulder, Colorado USA.
The Expanded UW SREF System and Statistical Inference STAT 592 Presentation Eric Grimit 1. Description of the Expanded UW SREF System (How is this thing.
Uncertainties about Uncertainty Cliff Mass. Major Issues.
A Brief History of Weather Forecasting. The Stone Age Prior to approximately 1955, forecasting was basically a subjective art, and not very skillful.
Ensemble Post-Processing and it’s Potential Benefits for the Operational Forecaster Michael Erickson and Brian A. Colle School of Marine and Atmospheric.
Effective Mesoscale, Short-Range Ensemble Forecasting Tony Eckel** and Clifford F. Mass Presented By: Eric Grimit University of Washington Atmospheric.
A Regression Model for Ensemble Forecasts David Unger Climate Prediction Center.
Chapter 13 – Weather Analysis and Forecasting. The National Weather Service The National Weather Service (NWS) is responsible for forecasts several times.
Statistical Methods for long-range forecast By Syunji Takahashi Climate Prediction Division JMA.
Weather Forecasting - II. Review The forecasting of weather by high-speed computers is known as numerical weather prediction. Mathematical models that.
Lecture II-2: Probability Review
“1995 Sunrise Fire – Long Island” Using an Ensemble Kalman Filter to Explore Model Performance on Northeast U.S. Fire Weather Days Michael Erickson and.
Probabilistic forecasts of (severe) thunderstorms for the purpose of issuing a weather alarm Maurice Schmeits, Kees Kok, Daan Vogelezang and Rudolf van.
Performance of the MOGREPS Regional Ensemble
Hypothesis Testing in Linear Regression Analysis
Forecasting and Numerical Weather Prediction (NWP) NOWcasting Description of atmospheric models Specific Models Types of variables and how to determine.
Probabilistic Prediction. Uncertainty in Forecasting All of the model forecasts I have talked about reflect a deterministic approach. This means that.
Exploring sample size issues for 6-10 day forecasts using ECMWF’s reforecast data set Model: 2005 version of ECMWF model; T255 resolution. Initial Conditions:
How can LAMEPS * help you to make a better forecast for extreme weather Henrik Feddersen, DMI * LAMEPS =Limited-Area Model Ensemble Prediction.
It Never Rains But It Pours: Modeling Mixed Discrete- Continuous Weather Phenomena J. McLean Sloughter This work was supported by the DoD Multidisciplinary.
Improving Ensemble QPF in NMC Dr. Dai Kan National Meteorological Center of China (NMC) International Training Course for Weather Forecasters 11/1, 2012,
Celeste Saulo and Juan Ruiz CIMA (CONICET/UBA) – DCAO (FCEN –UBA)
1 An overview of the use of reforecasts for improving probabilistic weather forecasts Tom Hamill NOAA / ESRL, Physical Sciences Div.
Probabilistic Forecasting. pdfs and Histograms Probability density functions (pdfs) are unobservable. They can only be estimated. They tell us the density,
Short-Range Ensemble Prediction System at INM José A. García-Moya SMNT – INM 27th EWGLAM & 12th SRNWP Meetings Ljubljana, October 2005.
Model Post Processing. Model Output Can Usually Be Improved with Post Processing Can remove systematic bias Can produce probabilistic information from.
Plans for Short-Range Ensemble Forecast at INM José A. García-Moya SMNT – INM Workshop on Short Range Ensemble Forecast Madrid, October,
The Challenge of Mesoscale Probabilistic Weather Prediction and Communication Cliff Mass University of Washington.
GIVE ME ODDS FORECASTING ENSEMBLE. KEY POINTS THERE ARE INEVITABLE UNCERTAINTIES IN NWP DUE TO UNCERTAINTIES IN INITIAL CONDITIONS AND MODEL FORMULATION.
18 September 2009: On the value of reforecasts for the TIGGE database 1/27 On the value of reforecasts for the TIGGE database Renate Hagedorn European.
Chapter 20 Classification and Estimation Classification – Feature selection Good feature have four characteristics: –Discrimination. Features.
An Examination Of Interesting Properties Regarding A Physics Ensemble 2012 WRF Users’ Workshop Nick P. Bassill June 28 th, 2012.
Ensembles and Probabilistic Prediction. Uncertainty in Forecasting All of the model forecasts I have talked about reflect a deterministic approach. This.
Nathalie Voisin 1, Florian Pappenberger 2, Dennis Lettenmaier 1, Roberto Buizza 2, and John Schaake 3 1 University of Washington 2 ECMWF 3 National Weather.
Common verification methods for ensemble forecasts
On the Challenges of Identifying the “Best” Ensemble Member in Operational Forecasting David Bright NOAA/Storm Prediction Center Paul Nutter CIMMS/Univ.
Verification of ensemble precipitation forecasts using the TIGGE dataset Laurence J. Wilson Environment Canada Anna Ghelli ECMWF GIFS-TIGGE Meeting, Feb.
2015 Production Suite Review: Report from NHC 2015 Production Suite Review: Report from NHC Eric S. Blake, Richard J. Pasch, Andrew Penny NCEP Production.
MPO 674 Lecture 2 1/20/15. Timeline (continued from Class 1) 1960s: Lorenz papers: finite limit of predictability? 1966: First primitive equations model.
Lecture 8: Measurement Errors 1. Objectives List some sources of measurement errors. Classify measurement errors into systematic and random errors. Study.
Toward an Effective Short-Range Ensemble Forecast System October 17, 2003 Maj Tony Eckel, USAF University of Washington Atmospheric Sciences Department.
Figures from “The ECMWF Ensemble Prediction System”
University of Washington Ensemble Systems for Probabilistic Analysis and Forecasting Cliff Mass, Atmospheric Sciences University of Washington.
Ensembles and Probabilistic Prediction
Verifying and interpreting ensemble products
Probabilistic Prediction
The Stone Age Prior to approximately 1960, forecasting was basically a subjective art, and not very skillful. Observations were sparse, with only a few.
Statistical Methods For Engineers
Post Processing.
14th Cyclone Workshop Brian Ancell The University of Washington
Christoph Gebhardt, Zied Ben Bouallègue, Michael Buchhold
Presentation transcript:

Probabilistic Prediction Cliff Mass University of Washington

Uncertainty in Forecasting Most numerical weather prediction (NWP) today and most forecast products reflect a deterministic approach. This means that we do the best job we can for a single forecast and do not consider uncertainties in the model, initial conditions, or the very nature of the atmosphere. However, the uncertainties are usually very significant and information on such uncertainty can be very useful.

This is really ridiculous!

The work of Lorenz (1963, 1965, 1968) demonstrated that the atmosphere is a chaotic system, in which small differences in the initialization, well within observational error, can have large impacts on the forecasts, particularly for longer forecasts. In a series of experiments he found that small errors in initial conditions can grow so that all deterministic forecast skill is lost at about two weeks. A Fundamental Issue

Butterfly Effect: a small change at one place in a complex system can have large effects elsewhere

Not unlike a pinball game

Uncertainty Extends Beyond Initial Conditions Also uncertainty in our model physics. –such as microphysics and boundary layer parameterizations. And further uncertainty produced by our numerical methods.

Probabilistic NWP To deal with forecast uncertainty, Epstein (1969) suggested stochastic-dynamic forecasting, in which forecast errors are explicitly considered during model integration. Essentially, uncertainty estimates are added to each term in the primitive equations. This stochastic method was not and still is not computationally practical.

Probabilistic-Ensemble Numerical Prediction (NWP) Another approach, ensemble prediction, was proposed by Leith (1974), who suggested that prediction centers run a collection (ensemble) of forecasts, each starting from a different initial state. The variations in the resulting forecasts could be used to estimate the uncertainty of the prediction. But even the ensemble approach was not possible at this time due to limited computer resources. Became practical in the late 1980s as computer power increased.

Ensemble Prediction Can use ensembles to estimate the probabilities that some weather feature will occur. The ensemble mean is more accurate on average than any individual ensemble member. Forecast skill of the ensemble mean is related to the spread of the ensembles When ensemble forecasts are similar, ensemble mean skill tend to be higher. When forecasts differ greatly, ensemble mean forecast skill tends to be less.

11 T The true state of the atmosphere exists as a single point in phase space that we never know exactly. A point in phase space completely describes an instantaneous state of the atmosphere. (pres, temp, etc. at all points at one time.) Nonlinear error growth and model deficiencies drive apart the forecast and true trajectories (i.e., Chaos Theory) P H A S E S P A C E 12h forecast 36h forecast 24h forecast 48h forecast T 48h observation T T T 12h observation 36h observation 24h observation An analysis produced to run an NWP model is somewhere in a cloud of likely states. Any point in the cloud is equally likely to be the truth. Deterministic Forecasting

12 T Ensemble Forecasting:  Encompasses truth  Reveals flow-dependent uncertainty  Yields objective stochastic forecast T 48h Forecast Region (forecast PDF) Analysis Region (analysis PDF) An ensemble of likely analyses leads to an ensemble of likely forecasts Ensemble Forecasting, a Stochastic Approach P H A S E S P A C E

22 May :30 PMGeneral Examination Presentation Probability Density Functions Usually we fit the distribution of ensemble members with a gaussian or other reasonably smooth theoretical distribution as a first step

A critical issue is the development of ensemble systems that create probabilistic guidance that is both reliable and sharp. We Need to Create Probability Density Functions (PDFs) of Each Variable That have These Characteristics

Elements of a Good Probability Forecast: Sharpness (also known as resolution) –The width of the predicted distribution should be as small as possible. Probability Density Function (PDF) for some forecast quantity Sharp Less Sharp

Elements of a Good Probability Forecast Reliability (also known as calibration) –A probability forecast p, ought to verify with relative frequency p. –Forecasts from climatology are reliable (by definition), so calibration alone is not enough. Reliability Diagram

17 Over many trials, record verification ’ s position (the “ rank ” ) among the ordered EF members. Over-Spread EFUnder-Spread EFReliable EF Cumulative Precip. (mm) Frequency EF PDF (curve) & 8 sample members (bars) True PDF (curve) & verification value (bar ) Verification Rank Histogram (a.k.a., Talagrand Diagram)-Another Measure of Reliability

Brier Score M : number of fcst/obs pairs : forecast probability {0.0…1.0} o j : observation {0.0 = no, 1.0 = yes} Continuous BS = 0 for perfect forecasts BS = 1 for perfectly wrong forecasts Brier Skill Score (BSS) directly examines reliability, resolution, and overall skill Brier Skill Score BSS = 1 for perfect forecasts BSS < 0 for forecasts worse than climo Brier Skill Score′ ADVANTAGES: 1) No need for long-term climatology 2) Can compute and visualize in reliability diagram 00 (reliability, rel) (resolution, res) (uncertainty, unc) I : number of probability bins (normally 11) N : number of data pairs in the bin : binned forecast probability (0.0, 0.1,…1.0 for 11 bins) o i : observed relative frequency for bin i o : sample climatology (total occurrences / total forecasts) Decomposed Brier Score by Discrete, Contiguous Bins

Probabilistic Information Can Produce Substantial Economic and Public Protection Benefits

There is a decision theory on using probabilistic information for economic savings C= cost of protection L= loss if a damaging event occurs Decision theory says you should protect if the probability of occurrence is greater than C/L

Critical Event: surface winds > 50kt Cost (of protecting): $150K Loss (if damage ): $1M C/L =.15 (15%) Hit False Alarm Miss Correct Rejection YES NO YES NO Forecast? Observed? Decision Theory Example $150K$1000K $150K$0K Optimal Threshold = 15%

History of Probabilistic Weather Prediction (in the U.S.)

Early Forecasting Started Probabilistically!!! Early forecasters, faced with large gaps in their young science, understood the uncertain nature of the weather prediction process and were comfortable with a probabilistic approach to forecasting. Cleveland Abbe, who organized the first forecast group in the United States as part of the U.S. Signal Corp, did not use the term “forecast” for his first prediction in 1871, but rather used the term “probabilities,” resulting in him being known as “Old Probabilities” or “Old Probs” to the public.

“Ol Probs” Professor Cleveland Abbe, issued the first public “Weather Synopsis and Probabilities” on February 19, 1871 A few years later, the term indications was substituted for probabilities, and by 1889 the term forecasts received official approval(Murphy 1997).

History of Probabilistic Prediction The first modern operational probabilistic forecasts in the United States were produced in These forecasts, for the probability of precipitation, were produced by human weather forecasters and thus were subjective probabilistic predictions. The first objective probabilistic forecasts were produced as part of the Model Output Statistics (MOS) system that began in 1969.

NOTE: Model Output Statistics (MOS) Based on simple linear regression with 12 predictors. Y = a 0 +a 1 X 1 + a 2 X 2 + a 3 X 3 + a 4 X 4 …

Ensemble Prediction Ensemble prediction began an NCEP in the early 1990s. ECMWF rapidly joined the club. During the past decades the size and sophistication of the NCEP and ECMWF ensemble systems have grown considerably, with the medium-range global ensemble system becoming an integral tool for many forecasters. Also during this period, NCEP has constructed a higher resolution, short-range ensemble system (SREF) that uses breeding to create initial condition variations.

Example: NCEP Global Ensemble System Begun in 1993 with the MRF (now GFS) First tried “lagged” ensembles as basis…using runs of various initializations verifying at the same time. Then used the “breeding” method to find perturbations to the initial conditions of each ensemble members. Breeding adds random perturbations to an initial state, let them grow, then reduce amplitude down to a small level, lets them grow again, etc. Give an idea of what type of perturbations are growing rapidly in the period BEFORE the forecast. Does not include physics uncertainty. Now replaced by Ensemble Transform Filter Approach

NCEP Global Ensemble 20 members at 00, 06, 12, and 18 UTC plus two control runs for each cycle 28 levels T190 resolution (roughly 80km resolution) 384 hours Uses stochastic physics to get some physics diversity

ECMWF Global Ensemble 50 members and 1 control 60 levels T399 (roughly 40 km) through 240 hours, T255 afterwards Singular vector approach to creating perturbations Stochastic physics

Several Nations Have Global Ensembles Too! China, Canada, Japan and others! And there are combinations of global ensembles like: –TIGGE: Thorpex Interative Grand Global Ensemble from ten national NWP centers –NAEFS: North American Ensemble Forecasting System combining U.S. and Canadian Global Ensembles

Popular Ensemble-Based Products

Spaghetti Diagram

Ensemble Mean

37 ‘ Ensemble Spread Chart Global Forecast System (GFS) Ensemble  “ best guess ” = high-resolution control forecast or ensemble mean  ensemble spread = standard deviation of the members at each grid point  Shows where “ best guess ” can be trusted (i.e., areas of low or high predictability)  Details unpredictable aspects of waves: amplitude vs. phase

38 Current Deterministic Meteogram Meteograms Versus “Plume Plots” FNMOC Ensemble Forecast System (EFS)  Data Range = meteogram-type trace of each ensemble member ’ s raw output  Excellent tool for point forecasting, if calibrated  Can easily (and should) calibrate for model bias  Calibrating for ensemble spread problems is difficult  Must use box & whisker, or confidence interval plot for large ensembles

39 Box and Whisker Plots

40

41 11/18 12/ / /00 06 Valid Time (UTC) Misawa AB, Japan Wind Direction AFWA Forecast Multimeteogram JME Cycle: 11Nov06, 18Z RWY: 100/280 15km Resolution Wind Speed (kt) 90% CI Extreme Min Extreme Max Mean Gray shaded area is 90% Confidence Interval (CI)

42 3 Hurricane Track Forecast & Potential

Ensemble-Based Probabilities

Postage Stamp Plots 13: avn* 11: ngps* 12: cmcg* 10: tcwb* 9: ukmo* 8: eta* Verification 1: cent 7: avn 5: ngps 6: cmcg 4: tcwb 3: ukmo 2: eta - Reveals high uncertainty in storm track and intensity - Indicates low probability of Puget Sound wind event SLP and winds

A Number of Nations Are Experimenting with Higher- Resolution Ensembles

European MOGREPS – 24 km resolution – Uses ETKF for diversity breeding) – Stochastic physics

NCEP Short-Range Ensembles (SREF) Resolution of 32 km Out to 87 h twice a day (09 and 21 UTC initialization) Uses both initial condition uncertainty (breeding) and physics uncertainty. Uses the Eta and Regional Spectral Models and recently the WRF model (21 total members)

SREF Current System ModelRes (km) LevelsMembers Cloud PhysicsConvection RSM-SAS 4528Ctl,n,p GFS physics Simple Arak-Schubert RSM-RAS 4528n,pGFS physics Relaxed Arak-Schubert Eta-BMJ 3260Ctl,n,pOp FerrierBetts-Miller-Janjic Eta-SAT 3260n,pOp FerrierBMJ-moist prof Eta-KF 3260Ctl,n,p Op FerrierKain-Fritsch Eta-KFD 3260n,pOp FerrierKain-Fritsch with enhanced detrainment PLUS * NMM-WRF control and 1 pert. Pair * ARW-WRF control and 1 pert. pair

The UW Ensemble System Perhaps the highest resolution operational ensemble systems are running at the University of Washington UWME: 8 members at 36 and 12-km UW EnKF system: 60 members at 36 and 4-km

Calibration (Post-Processing) of Ensembles Is Essential

Calibration of Mesoscale Ensemble Systems: The Problem The component models of virtually all ensemble systems have systematic bias that substantially degrade the resulting probabilistic forecasts. Since different models or runs have different systematic bias, this produces forecast variance that DOES NOT represent true forecast uncertainty. Systematic bias reduces sharpness and degrades reliability. Also, most ensemble systems produce forecasts that are underdispersive. Not enough variability!

Example of Bias Correction for UW Ensemble System

Average RMSE (  C) and (shaded) Average Bias Uncorrected + T 2 12 h 24 h 36 h 48 h

Average RMSE (  C) and (shaded) Average Bias Bias-Corrected T 2 12 h 24 h 36 h 48 h

*UW Basic Ensemble with bias correction UW Basic Ensemble, no bias correction *UW Enhanced Ensemble with bias cor. UW Enhanced Ensemble without bias cor Skill for Probability of T 2 < 0°C BSS: Brier Skill Score

The Next Step: Bayesian Model Averaging Although bias correction is useful it is possible to do more. –Optimize the variance of the forecast distributions –Weight the various ensemble members using their previous performance. –An effective way to do this is through Bayesian Model Averaging (BMA).

Bayesian Model Averaging Assumes a gaussian (or other) PDF for each ensemble member. Assumes the variance of each member is the same (in current version). Includes a simple bias correction for each member. Weights each member by its performance during a training period (we are using 25 days) Adds the pdfs from each member to get a total pdf.

Application of BMA-Max 2-m Temperature (all stations in 12 km domain)

Being Able to Create Reliable and Sharp Probabilistic Information is Only Half the Problem! Even more difficult will be communication and getting people and industries to use it.

Deterministic Nature? People seem to prefer deterministic products: “tell me what is going to happen” People complain they find probabilistic information confusing. Many don’t understand POP (probability of precipitation). Media and internet not moving forward very quickly on this.

National Weather Service Icons are not effective in communicating probabilities

And a “slight” chance of freezing drizzle reminds one of a trip to Antarctica

Commercial sector is no better (Weather.Com)

A great deal of research and development is required to develop effective approaches for communicating probabilistic forecasts which will not overwhelm people and allow them to get value out of them.