ECMWF Slide 1Met Op training course – Reading, March 2004 Forecast verification: probabilistic aspects Anna Ghelli, ECMWF.

Slides:



Advertisements
Similar presentations
Verification of Probabilistic Forecast J.P. Céron – Direction de la Climatologie S. Mason - IRI.
Advertisements

Advanced Piloting Cruise Plot.
Climate Modeling LaboratoryMEASNC State University An Extended Procedure for Implementing the Relative Operating Characteristic Graphical Method Robert.
Copyright © 2003 Pearson Education, Inc. Slide 1 Computer Systems Organization & Architecture Chapters 8-12 John D. Carpinelli.
Chapter 1 The Study of Body Function Image PowerPoint
1 Copyright © 2013 Elsevier Inc. All rights reserved. Chapter 1 Embedded Computing.
1 Copyright © 2013 Elsevier Inc. All rights reserved. Appendix 01.
Properties Use, share, or modify this drill on mathematic properties. There is too much material for a single class, so you’ll have to select for your.
UNITED NATIONS Shipment Details Report – January 2006.
1 RA I Sub-Regional Training Seminar on CLIMAT&CLIMAT TEMP Reporting Casablanca, Morocco, 20 – 22 December 2005 Status of observing programmes in RA I.
THOR Annual Meeting - Bergen 9-11 November /25 On the impact of initial conditions relative to external forcing on the skill of decadal predictions:
Federal Department of Home Affairs FDHA Federal Office of Meteorology and Climatology MeteoSwiss Extended range forecasts at MeteoSwiss: User experience.
Page 1© Crown copyright 2004 Seasonal forecasting activities at the Met Office Long-range Forecasting Group, Hadley Centre Presenter: Richard Graham ECMWF.
Page 1© Crown copyright 2004 Presentation to ECMWF Forecast Product User Meeting 16th June 2005.
Slide 1ECMWF forecast User Meeting -- Reading, June 2006 Verification of weather parameters Anna Ghelli, ECMWF.
Slide 1ECMWF forecast products users meeting – Reading, June 2005 Verification of weather parameters Anna Ghelli, ECMWF.
ECMWF User Meeting / 1 Pertti Nurmi Juha Kilpinen Annakaisa Sarkanen ( Finnish Meteorological Institute ) Probabilistic Forecasts.
Page 1 © Crown copyright 2005 ECMWF User Meeting, June 2006 Developments in the Use of Short and Medium-Range Ensembles at the Met Office Ken Mylne.
Severe Weather Forecasts
Training Course 2009 – NWP-PR: Ensemble Verification II 1/33 Ensemble Verification II Renate Hagedorn European Centre for Medium-Range Weather Forecasts.
Jeopardy Q 1 Q 6 Q 11 Q 16 Q 21 Q 2 Q 7 Q 12 Q 17 Q 22 Q 3 Q 8 Q 13
Jeopardy Q 1 Q 6 Q 11 Q 16 Q 21 Q 2 Q 7 Q 12 Q 17 Q 22 Q 3 Q 8 Q 13
FACTORING ax2 + bx + c Think “unfoil” Work down, Show all steps.
Year 6 mental test 10 second questions
Evaluating Provider Reliability in Risk-aware Grid Brokering Iain Gourlay.
1 NCAS SMA presentation 14/15 September 2004 The August 2002 European floods: atmospheric teleconnections and mechanisms Mike Blackburn (1), Brian Hoskins.
1 Understanding Multiyear Estimates from the American Community Survey.
Solve Multi-step Equations
Richmond House, Liverpool (1) 26 th January 2004.
REVIEW: Arthropod ID. 1. Name the subphylum. 2. Name the subphylum. 3. Name the order.
Detection Chia-Hsin Cheng. Wireless Access Tech. Lab. CCU Wireless Access Tech. Lab. 2 Outlines Detection Theory Simple Binary Hypothesis Tests Bayes.
EU Market Situation for Eggs and Poultry Management Committee 21 June 2012.
IP Multicast Information management 2 Groep T Leuven – Information department 2/14 Agenda •Why IP Multicast ? •Multicast fundamentals •Intradomain.
VOORBLAD.
Factor P 16 8(8-5ab) 4(d² + 4) 3rs(2r – s) 15cd(1 + 2cd) 8(4a² + 3b²)
Basel-ICU-Journal Challenge18/20/ Basel-ICU-Journal Challenge8/20/2014.
1..
CONTROL VISION Set-up. Step 1 Step 2 Step 3 Step 5 Step 4.
© 2012 National Heart Foundation of Australia. Slide 2.
Understanding Generalist Practice, 5e, Kirst-Ashman/Hull
Chapter 8: Introduction to Hypothesis Testing. 2 Hypothesis Testing An inferential procedure that uses sample data to evaluate the credibility of a hypothesis.
Model and Relationships 6 M 1 M M M M M M M M M M M M M M M M
25 seconds left…...
H to shape fully developed personality to shape fully developed personality for successful application in life for successful.
Januar MDMDFSSMDMDFSSS
Deutscher Wetterdienst Zentrale Vorhersage DWD Workshop "Use and Verification of LEPS products", Geneve, May, Verification of LEPS products.
Statistical Inferences Based on Two Samples
©Brooks/Cole, 2001 Chapter 12 Derived Types-- Enumerated, Structure and Union.
PSSA Preparation.
Essential Cell Biology
1 Chapter 13 Nuclear Magnetic Resonance Spectroscopy.
1 McGill University Department of Civil Engineering and Applied Mechanics Montreal, Quebec, Canada.
4/4/2015Slide 1 SOLVING THE PROBLEM A one-sample t-test of a population mean requires that the variable be quantitative. A one-sample test of a population.
14 May 2001QPF Verification Workshop Verification of Probability Forecasts at Points WMO QPF Verification Workshop Prague, Czech Republic May 2001.
Verification of probability and ensemble forecasts
Details for Today: DATE:3 rd February 2005 BY:Mark Cresswell FOLLOWED BY:Assignment 2 briefing Evaluation of Model Performance 69EG3137 – Impacts & Models.
ECMWF WWRP/WMO Workshop on QPF Verification - Prague, May 2001 NWP precipitation forecasts: Validation and Value Deterministic Forecasts Probabilities.
Measuring forecast skill: is it real skill or is it the varying climatology? Tom Hamill NOAA Earth System Research Lab, Boulder, Colorado
ECMWF Training Course Reading, 25 April 2006 EPS Diagnostic Tools Renate Hagedorn European Centre for Medium-Range Weather Forecasts.
The Centre for Australian Weather and Climate Research A partnership between CSIRO and the Bureau of Meteorology Verification and Metrics (CAWCR)
Verification of ensemble systems Chiara Marsigli ARPA-SIMC.
Common verification methods for ensemble forecasts
Verifying and interpreting ensemble products
Probabilistic forecasts
Short Range Ensemble Prediction System Verification over Greece
Presentation transcript:

ECMWF Slide 1Met Op training course – Reading, March 2004 Forecast verification: probabilistic aspects Anna Ghelli, ECMWF

ECMWF Slide 2Met Op training course – Reading, March 2004 Why probability forecasts? …the widespread practice of ignoring uncertainty when formulating and communicating forecasts represents an extreme form of inconsistency and generally results in the largest possible reductions in quality and value. --Murphy (1993)

ECMWF Slide 3Met Op training course – Reading, March 2004 Outline 1.Basics 2.Verification measures 3.Performance 4.Signal Detection Theory: Relative Operating Characteristic (ROC) 5.Cost-Loss model 6.Conclusions

ECMWF Slide 4Met Op training course – Reading, March 2004 BASICS Types of forecasts - Completely confident Rain/No rain - Probabilistic Objective (deterministic, statistical, ensemble-based) Subjective P(x) x xoxo

ECMWF Slide 5Met Op training course – Reading, March 2004 Verification framework Observed value x will be 0 if the event has not happened and 1 if the event occurred x = 0 or 1 Forecast probability will vary between 0 and 1.0 f = 0, …, 1.0 Joint distribution: p(f,x), where x = 0, 1

ECMWF Slide 6Met Op training course – Reading, March 2004 Factorization Conditional and marginal probabilities Calibration-Refinement factorization: p(f,x) = p(x|f) p(f) where p(f) is the frequency of use of each forecast probability Likelihood-Base Rate factorization: p(f,x) = p(f|x) p(x) where p(x) is the relative frequency of a Yes observation (e.g., the sample climatology)

ECMWF Slide 7Met Op training course – Reading, March 2004 Verification measures based on calibration- refinement factorization Reliability diagram p(x=1|f i ) vs. f i Plot of the observed relative frequency of an event as function of its forecast probability. It shows the agreement between the mean forecast probability and the observed frequency. Sharpness diagram p(f) It indicates the capability of the system to forecast extreme values, or values close 0 or 1. Attributes diagram Reliability, Resolution, Skill/No-skill

ECMWF Slide 8Met Op training course – Reading, March 2004 Performance measures Brier score: Analogous to MSE; negative orientation; For perfect forecasts: BS=0 Brier skill score: Analogous to MSE skill score

ECMWF Slide 9Met Op training course – Reading, March 2004 Decomposition of the Brier Score ReliabilityResolutionUncertainty Where I is the total number of distinct probability values and resolution tells how informative the probabilistic forecast is; it varies from zero for a system for which all forecast probabilities verify with the same frequency of occurrence to the sample uncertainty for a system for which the frequency of verifying occurrences takes only values 0 or 100% (such a system resolves perfectly the forecast between occurring and non-occurring events); reliability tells how close the frequencies of observed occurrences are from the forecast probabilities (on average, when an event is forecast with probability p, it should occur with the same frequency p); uncertainty varies from 0 to 0.25 and indicates how close to 50% the occurrence of the event was during the sample period (uncertainty is 0.25 when the event is split equally into occurrence and non- occurrence).

ECMWF Slide 10Met Op training course – Reading, March 2004 Reliability and Sharpness (from Wilks 1995) ClimatologyMinimal RESUnderforecasting Good RES, at expense of REL Reliable forecasts of rare event Small sample size

ECMWF Slide 11Met Op training course – Reading, March 2004 Attributes diagram (from Wilks 1995)

ECMWF Slide 12Met Op training course – Reading, March 2004 examples

ECMWF Slide 13Met Op training course – Reading, March 2004 Reliability diagram 24 h accumulated precipitation forecast verified against observed values for different thresholds: 1mm/24h (right) and 5mm/24 h (bottom). The diagrams are relative to Europe. The period is December 2003 to February For the 1mm/24h threshold the model is overconfident. The curve is much closer to the diagonal (perfect forecast) in the 5mm/24h threshold

ECMWF Slide 14Met Op training course – Reading, March 2004 Reliability diagram 24 h accumulated precipitation forecast verified against observed values for different thresholds: 10mm/24h (right) and 20mm/24 h (bottom). The diagrams are relative to Europe for the period December 2003 to February 2004 For the 10mm/24h threshold the model shows a very good match between forecast probability and observed frequencies. The 20mm/24h threshold shows the effect of small sample size!

ECMWF Slide 15Met Op training course – Reading, March 2004 Reliability diagram T850 anomaly greater then 4K (right) and 8K (bottom). The diagrams are relative to Europe for the period June 2003 to July 2003 For both anomalies, the forecast is overconfident.

ECMWF Slide 16Met Op training course – Reading, March 2004 Brier Skill Score (reference is long term climate) for Europe at t+96 (top panel) and t+144 (bottom panel). The variable is the temperature at 850hPa. The curve shows the improvement versus the reference system. Smaller anomalies are better forecast

ECMWF Slide 17Met Op training course – Reading, March 2004 Brier Skill Score (BSS) for different thresholds – Forecast range D+4 Improvements of the EPS in 1999 (increase of vertical resolution and change in cloud scheme) and in Autumn 2000 (change in horizontal resolution)

ECMWF Slide 18Met Op training course – Reading, March 2004 Signal Detection Theory (SDT) Approach that has commonly been applied in medicine and other fields Brought to meteorology by Ian Mason (1982) Evaluates the ability of forecasts to discriminate between occurrence and non-occurrence of an event Summarizes characteristics of the Likelihood-Base Rate decomposition of the framework Tests model performance relative to specific threshold Allows comparison of categorical and probabilistic forecasts

ECMWF Slide 19Met Op training course – Reading, March 2004 ROC -- Basics Based on likelihood-base rate decomposition p(f,x) = p(f|x) p(x) Basic elements : Hit rate (H) H = a/(a+c) -- Estimate of p(f=1|x=1) False Alarm Rate (F) F = b/(b+d) -- Estimate of p(f=1|x=0) Relative Operating Characteristic curve Plot H vs. F Obs YESObs NO FC YESab FC NOcd

ECMWF Slide 20Met Op training course – Reading, March 2004 ROC 24h accumulated precipitation for Europe; DJF > 1mm/24h > 5mm/24h 20%

ECMWF Slide 21Met Op training course – Reading, March 2004 ROC 24h accumulated precipitation for Europe; DJF > 5mm/24h > 1mm/24h

ECMWF Slide 22Met Op training course – Reading, March 2004 ROC 24h accumulated precipitation for Europe; JJA 2002 > 5mm/24h

ECMWF Slide 23Met Op training course – Reading, March 2004 ROC Area Area under the ROC is a measure of forecast skill - Values less than 0.5 indicate negative skill - Area can be underestimated if curve is approximated by straight line segments

ECMWF Slide 24Met Op training course – Reading, March 2004 ROC Area T850 verified against analysis for t+96 (top) and t+144 (bottom). Verification area: Europe

ECMWF Slide 25Met Op training course – Reading, March ROC Area for different thresholds – Forecast range D+4 Sensible improvements of the EPS since Autumn 2000

ECMWF Slide 26Met Op training course – Reading, March ROC Area for different thresholds – Forecast range D+7

ECMWF Slide 27Met Op training course – Reading, March 2004 Verification of ensemble forecasts summary Probabilistic forecasts from ensemble systems can be verified using standard approaches for probabilistic forecasts Common methods Brier score Reliability diagram Brier Skill Score ROC ROC area

ECMWF Slide 28Met Op training course – Reading, March 2004 Bad weather yes Bad weather no Protect yes CC Protect no L0 Event occurs yes Event occurs no Event forecast yes ab Event forecast no cd Using forecast all the time: expense E f =aC+bC+cL Perfect forecast: expense E p = (a+c)C Climate information: expense E c = min(C, (a+b)L) Value of forecast : reduction in expense compared to climate information V= (saving from using forecast)/ (saving from perfect forecast) V= (E c – E f )/(E c -E p ) Cost – Loss Basics

ECMWF Slide 29Met Op training course – Reading, March 2004 Value can be written as follows: Value depends on Forecast quality H and F User through C/L Weather event (a+c) if C/L > if C/L < Quality, value and user

ECMWF Slide 30Met Op training course – Reading, March 2004 Cost-loss model – probabilistic forecast value Known the climatological probability that adverse event happens p clim take action if p clim *L is larger than C P clim > C/L action! P clim < C/L no action

ECMWF Slide 31Met Op training course – Reading, March 2004 Cost – Loss model : probabilistic forecast value Act when probability exceed a certain threshold Choice of probability is user dependent

ECMWF Slide 32Met Op training course – Reading, March 2004 Cost- Loss model: deterministic vs EPS Control forecast: red line EPS: blue line

ECMWF Slide 33Met Op training course – Reading, March 2004 Conclusion Probabilistic forecasts from ensemble systems can be verified using standard approaches for probabilistic forecasts Common methods are: Brier score Reliability diagram Brier Skill Score ROC ROC area The performance of the EPS assessed using probabilistic scores shows improvements We should not forget that not only quality is important, we should look at the value of a forecast to its final user. The forecast has value if it helps the end user to make decisions