Review of verification activities and developments

Slides:



Advertisements
Similar presentations
Robin Hogan Ewan OConnor University of Reading, UK What is the half-life of a cloud forecast?
Advertisements

ECMWF long range forecast systems
Report of the Q2 Short Range QPF Discussion Group Jon Ahlquist Curtis Marshall John McGinley - lead Dan Petersen D. J. Seo Jean Vieux.
1 00/XXXX © Crown copyright Use of radar data in modelling at the Met Office (UK) Bruce Macpherson Mesoscale Assimilation, NWP Met Office EWGLAM / COST-717.
© Crown copyright Met Office Enhanced rainfall services Paul Davies.
Federal Department of Home Affairs FDHA Federal Office of Meteorology and Climatology MeteoSwiss Quantitative precipitation forecasts in the Alps – first.
Ensemble Post-Processing and it’s Potential Benefits for the Operational Forecaster Michael Erickson and Brian A. Colle School of Marine and Atmospheric.
Performance of the MOGREPS Regional Ensemble
Verification has been undertaken for the 3 month Summer period (30/05/12 – 06/09/12) using forecasts and observations at all 205 UK civil and defence aerodromes.
CARPE DIEM Centre for Water Resources Research NUID-UCD Contribution to Area-3 Dusseldorf meeting 26th to 28th May 2003.
© Crown copyright Met Office Operational OpenRoad verification Presented by Robert Coulson.
ECMWF WWRP/WMO Workshop on QPF Verification - Prague, May 2001 NWP precipitation forecasts: Validation and Value Deterministic Forecasts Probabilities.
1 On the use of radar data to verify mesoscale model precipitation forecasts Martin Goeber and Sean Milton Model Diagnostics and Validation group Numerical.
4th Int'l Verification Methods Workshop, Helsinki, 4-6 June Methods for verifying spatial forecasts Beth Ebert Centre for Australian Weather and.
ESA DA Projects Progress Meeting 2University of Reading Advanced Data Assimilation Methods WP2.1 Perform (ensemble) experiments to quantify model errors.
Short-Range Ensemble Prediction System at INM José A. García-Moya & Carlos Santos SMNT – INM COSMO Meeting Zurich, September 2005.
ISDA 2014, Feb 24 – 28, Munich 1 Impact of ensemble perturbations provided by convective-scale ensemble data assimilation in the COSMO-DE model Florian.
Page 1© Crown copyright 2007SRNWP 8-11 October 2007, Dubrovnik SRNWP – Revised Verification Proposal Clive Wilson Presented by Terry Davies at SRNWP Meeting.
Page 1© Crown copyright 2005 SRNWP – Revised Verification Proposal Clive Wilson, COSMO Annual Meeting September 18-21, 2007.
COSMO Annual Meeting September 2005 Zurich (Switzerland) Short-Range Numerical Weather Prediction Programme.
How can LAMEPS * help you to make a better forecast for extreme weather Henrik Feddersen, DMI * LAMEPS =Limited-Area Model Ensemble Prediction.
Federal Department of Home Affairs FDHA Federal Office of Meteorology and Climatology MeteoSwiss Data mining in the joint D- PHASE and COPS archive Mathias.
Pablo Santos WFO Miami, FL Mark DeMaria NOAA/NESDIS David Sharp WFO Melbourne, FL rd IHC St Petersburg, FL PS/DS “HURRICANE CONDITIONS EXPECTED.”
© Crown copyright Met Office Preliminary results using the Fractions Skill Score: SP2005 and fake cases Marion Mittermaier and Nigel Roberts.
EUROBRISA WORKSHOP, Paraty March 2008, ECMWF System 3 1 The ECMWF Seasonal Forecast System-3 Magdalena A. Balmaseda Franco Molteni,Tim Stockdale.
Improving Ensemble QPF in NMC Dr. Dai Kan National Meteorological Center of China (NMC) International Training Course for Weather Forecasters 11/1, 2012,
Latest results in verification over Poland Katarzyna Starosta, Joanna Linkowska Institute of Meteorology and Water Management, Warsaw 9th COSMO General.
© Crown copyright Met Office Probabilistic turbulence forecasts from ensemble models and verification Philip Gill and Piers Buchanan NCAR Aviation Turbulence.
9th COSMO General Meeting September 2007 Athens (Greece) Programme Short-Range NWP Report of the Programme Manager.
Short-Range Ensemble Prediction System at INM José A. García-Moya SMNT – INM 27th EWGLAM & 12th SRNWP Meetings Ljubljana, October 2005.
Page 1© Crown copyright Scale selective verification of precipitation forecasts Nigel Roberts and Humphrey Lean.
EMS 2013 (Reading UK) Verification techniques for high resolution NWP precipitation forecasts Emiel van der Plas Kees Kok Maurice.
U. Damrath, COSMO GM, Athens 2007 Verification of numerical QPF in DWD using radar data - and some traditional verification results for surface weather.
Verification of ensemble systems Chiara Marsigli ARPA-SIMC.
Page 1© Crown copyright 2005 Met Office Verification -status Clive Wilson, Presented by Mike Bush at EWGLAM Meeting October 8- 11, 2007.
Nathalie Voisin 1, Florian Pappenberger 2, Dennis Lettenmaier 1, Roberto Buizza 2, and John Schaake 3 1 University of Washington 2 ECMWF 3 National Weather.
Page 1© Crown copyright 2004 The use of an intensity-scale technique for assessing operational mesoscale precipitation forecasts Marion Mittermaier and.
Verification of ensemble precipitation forecasts using the TIGGE dataset Laurence J. Wilson Environment Canada Anna Ghelli ECMWF GIFS-TIGGE Meeting, Feb.
Trials of a 1km Version of the Unified Model for Short Range Forecasting of Convective Events Humphrey Lean, Susan Ballard, Peter Clark, Mark Dixon, Zhihong.
© Crown copyright Met Office Seasonal forecasting: Not just seasonal averages! Emily Wallace November 2012.
WRF Verification Toolkit Workshop, Boulder, February 2007 Spatial verification of NWP model fields Beth Ebert BMRC, Australia.
NCAR, 15 April Fuzzy verification of fake cases Beth Ebert Center for Australian Weather and Climate Research Bureau of Meteorology.
Federal Department of Home Affairs FDHA Federal Office of Meteorology and Climatology MeteoSwiss Probabilities from COSMO-2 derived with the neighborhood.
Hirlam-Aladin All Staff Meeting Jose A. Garcia-Moya. AEMET. Spain. On behalf of Carlos Santos, Anna Ghelli (ECMWF) & Imanol Guerrero ECMWF & AEMET.
© Crown copyright Met Office EUMETNET- SRNWP verification programme -Met Office Responsible member Clive Wilson – Project manager 31st EWGLAM/16th SRNWP.
New results in COSMO about fuzzy verification activities and preliminary results with VERSUS Conditional Verification 31th EWGLAM &16th SRNWP meeting,
I. Sanchez, M. Amodei and J. Stein Météo-France DPREVI/COMPAS
LEPS VERIFICATION ON MAP CASES
Nigel Roberts Met Reading
Plans for Met Office contribution to SMOS+STORM Evolution
Fuzzy verification using the Fractions Skill Score
Systematic timing errors in km-scale NWP precipitation forecasts
Multi-scale validation of high resolution precipitation products
A New Method for Evaluating Regional Air Quality Forecasts
Verifying and interpreting ensemble products
Nathalie Voisin, Andy W. Wood and Dennis P. Lettenmaier
Verification of multi-model ensemble forecasts using the TIGGE dataset
Application of satellite-based rainfall and medium range meteorological forecast in real-time flood forecasting in the Upper Mahanadi River basin Trushnamayee.
Verification of nowcasting products: Issues and methods
forecasts of rare events
Thomas Gastaldo, Virginia Poli, Chiara Marsigli
Application of a global probabilistic hydrologic forecast system to the Ohio River Basin Nathalie Voisin1, Florian Pappenberger2, Dennis Lettenmaier1,
Quantitative verification of cloud fraction forecasts
Science Objectives contained in three categories
COSMO-LEPS Verification
Deterministic (HRES) and ensemble (ENS) verification scores
Christoph Gebhardt, Zied Ben Bouallègue, Michael Buchhold
Verification of Tropical Cyclone Forecasts
Short Range Ensemble Prediction System Verification over Greece
Ryan Kang, Wee Leng Tan, Thea Turkington, Raizan Rahmat
Presentation transcript:

Review of verification activities and developments Clive Wilson – for Expert Team on diagnostics, validation & verification 31st EWGLAM/16th SRNWP meetings – Athens 29 Sep 2009 © Crown copyright Met Office

Contents Expert team members EUMETNET SRNWP-V – separate presentation tomorrow 4th workshop of WWRP/WGNE working group on verification ,Helsinki, 8-11 June 2009 ECMWF TAC subgroup Consortia activities Plans © Crown copyright Met Office

Expert Team on diagnostics, validation & verification Members Clive Wilson (chair), Joël Stein, Carl Fortelius, Francis Schubiger , Dijana Klaric Dave Richardson (ECMWF contact) Additional members Marek Jerczynski, Alexander Kann , Andrea Raspanti, Ulf Andre, Xiaohua Yang, Lovro Kalin, Nigel Roberts, Marion Mittermaier © Crown copyright Met Office

EUMETNET/SRNWP Verification programme D1: Operational verification comparison of one version of each of the 4 regional models of Europe (available for all the participating members). Responsible Member – Met Office Programme manager-Clive Wilson Deputy PM : Marion Mittermaier Commenced 1 Jan 2009 Ends 31 December 2010 Costs – € 32000 /year © Crown copyright Met Office

4th International verification Methods workshop -JWGFVR Helsinki, 8-11 June 2009 Pertti Nurmi (FMI) local organiser Aims Focus on extremes & severe weather Ensembles/probabilistic verification Uncertainty & Value High resolution forecast verification Promote more focused user-oriented verification © Crown copyright Met Office

4th International verification Methods workshop –SRNWP participation 9 members of ET attended + several others 12 oral, 4 poster Clive Wilson: Do key performance targets work?. Lovro Kalin: Is ETSS Really Equitable? Adriano Raspanti: VERSUS: Unified Verification Package in COSMO Marek Jerczynski: Some robust scale separation methods at work Marion Mittermaier: Time-series analysis of scale-selective verification: Can we use it for operational forecast monitoring? Joel Stein and Marielle Amodei: Another look at the contingency tables: Scores based on Manhattan distances in the error space © Crown copyright Met Office 6

4th International verification Methods workshop –SRNWP participation -2 Clive Wilson: A critical look at the verification of Met Office "Flash" Warnings Marion Mittermaier: Verifying extreme rainfall alerts for surface water flooding Marion Mittermaier: Identifying skillful spatial scales using the Fraction skill Score Ulrich Damrath: Some experiences during verification of precipitation forecasts using fuzzy techniques Kees Kok: Valuing information from high resolution forecasts Chiara Marsigli: QPF Verification of Limited Area Ensemble Systems during the MAP D-PHASE OP Marielle Amodei: Deterministic and fuzzy verification of the cloudiness of High Resolution operational models Francis Schubiger: Verification of precipitation forecasts of the MAP D-PHASE data set with fuzzy methods Sami Niemelä: Verification of High resolution Precipitation forecasts by Using the SAL Method © Crown copyright Met Office

ECMWF TAC Verification Subgroup recommend headline measures that are suitable to complement those in the current ECMWF Strategy (namely anomaly correlation of Z500 for the deterministic forecast and probabilistic scores of T850 for the EPS); recommend verification procedures to aid forecasters’ decision making; recommend measures suitable for validating forecasts of weather associated with high impact events; identify requirements for observational data necessary for this verification; review Member State and Co-operating State requirements for the development of future forecast products. © Crown copyright Met Office

ECMWF TAC Verification Subgroup - meetings March & Sept 2009 Chair Pertti Nurmi FMI Martin Goeber- DWD Carlos Santos AEMET Marielle Amodei Meteo-France Jim Hamilton – Met Eirean Kees Kok- KNMI Marion Mittermaier Met Office Clive Wilson EUMETNET David Stephenson, Chris Ferro, Exeter University ECMWF David Richardson, Mark Rodwell,Walter Zweiflhofer, Erik andersson, Laura Ferranti, Anna Ghelli, Time Hewson,Cristina Primo © Crown copyright Met Office

ECMWF TAC Verification Subgroup-new headline score 24h precipitation new headline currently deterrministic fc Nearest grid-point instead of interpolation ACC 7 RACC sensitivity to outliers 1995-2005 obs Europe 2800  3600 per month ACC sensitive to sample size Zero rainfall inlfuence needs investigation Operational verification different QC Mark Rodwell – new score SLEEPS Semi-linear Equitable Error in Probability Space Potential candidate for monitoring NWP trends Shows reduced errors over time © Crown copyright Met Office

SLEEPS Mark Rodwell NB still being developed 3 category Dry, light and heavy ppn Based on local climatology at stations ~5000 1980-2007 Need at least 150 (~5 years for each month) Scoring matrix for 3 cumulative prob categories Equitablity imposed Not symetric -> semi-linear © Crown copyright Met Office

ECMWF TAC Verification Subgroup –extreme events Extreme dependency score Modified –independent base rate Symmetric SEDS (Hogan et al) All converge for rarer events Theroretical/practcal properties need further investigation before most suitable score can be proposed Influence of bias on EDS-type scores to be investigated © Crown copyright Met Office

ECMWF TAC Verification Subgroup – aid to forecasters New clustering scheme for flow dependent verification Proposed compute prob. Scores verifying analyses & individual clusters – see accuracy at different forecast ranges Investigate transitions between regimes Stratification of standard error measures by regimes for assessing year on year changes Feature tracking Extra tropical cyclones (Hewson) Need to look at specific cyclones cf TCs Emphasis on those associated with extreme weather Strike probs in various classes severity cf analyses © Crown copyright Met Officet

Aladin verification activities 1) Common Aladin verification package operational (Slovenia) No significant change during last year. Against surface stations and radio-sounding European data. Allows comparison of the different versions of Aladin Results can be compared with the inter comparison results from the Met Office. © Crown copyright Met Office

Aladin verification activities 2) “Fuzzy” methods Fuzzy, pattern recognition tests (Poland) Operational in Météo-France compare high and low resolution models (AROME and ALADIN-FRANCE ) Deterministic forecasts --> probabilities forecast frequency in a neighbourhood = PN 2 Brier skill scores against persistence either : compare PN to 0 /1 value observed at the centre of the neighborhood (BSS_SO for single observation) or to the observed frequency in the same neighborhood (BSS_NO for neighborhood observation). Use climatological French raingauges network (Amodei and Stein 2009) and the tables of contingency corresponds to a temporal windows of 3 months. © Crown copyright Met Office

Aladin verification activities Comparison of post-processed forecasted brightness temperatures with observed temperature by Meteosat 9 for ALADIN and AROME models. Classical and probabilistic scores to quantify the double penalty influence on the comparison between 2 models of different resolution Presentation by I. Sanchez in Athens “Theoretical" formulation of the scores relates them to Manhattan distances in the phase space of the possible forecasts when the observations are fixed. poster presented in Helsinki (Stein and Amodei 2009) based on a preliminary version of a paper submitted to Met App. Revised version restricted to 2x2 tables is now under consideration (Stein 2009). The graphical representation deduced from this study of the table of contingency is now implemented on our web site to compare Arome and Aladin. © Crown copyright Met Office

Joël Stein Scaled misses Bias=1 Heidke skill score Equal error distance Scaled false alarms

Joël Stein Heidke skill score Peirce skill score Deterministic limit

COSMO verification activities -separate slides 200910_Clive_EWGLAM-SRNWP.ppt © Crown copyright Met Office

Hirlam verification activities HARMONIE ver package (Andrae 2007) Powerful & flexible compare models & to observations Tables, maps,time-series,vertical profiles,histograms,scatter, du=iurnal & seasonal cycles Standards scores & contingency table scores Includes SAL (Wernli et al, 2008) Recent addition graphs of freq. bias & threat score on hit rate v false alarm ration (Wilson, 2009) © Crown copyright Met Office

aro33h1: AROME 33h1 (2.5km L40) MB71: HIRLAM 7.1.4 (7.5km L60) V72 (RCR): HIRLAM 7.2 (16.5km L60)

aro33h1: AROME 33h1 (2.5km L40) MB71: HIRLAM 7.1.4 (7.5km L60) V72 (RCR): HIRLAM 7.2 (16.5km L60)

Hirlam verification (contd) GLAMEPS calibration & validation at AEMET HPPV ( Santos & Hagel, 2007) Multi-model Rank histograms, PIT histograms, spread-skill, Brier SS, ROC, reliability, sharpness, RV Feature based verifcation : FMI- Finnish radar reflectivity compared to Finnish AROME using radar simulation model (Niemela, 2009 4th Intern. Workshop)) SAL Fuzzy – MOS, traditional scores (Kok et al 2008) © Crown copyright Met Office

SAL features S: Structure -2 … 0 … +2 A: Amplitude -2 … 0 … +2 objects Perfect objects too small or too large or too peaked too flat A: Amplitude -2 … 0 … +2 averaged Perfect averaged QPF under- QPF over- estimated estimated L: Location 0 … +2 Perfect wrong location of Total Center of Mass (TCM) and / or of objects relative to TCM

SAL

Met Office verification activities Operational verification package now extended for ensembles- MOGREPS Reliability, rank histograms, ROC, Brier, value Multimodel ensembles Fractional skill score (FSS) now operational for NAE & UK4 ( soon UK1.5km) Evaluation of new Flood forecasting - Extreme Rainfall Alert service Probabilistic from “fuzzy” UK4 & lagged average Evaluation of new 1.5km Review of warnings (Exeter University, Stephenson & Joliffe, also 4th Intern. workshop) © Crown copyright Met Office

Fractional skill score (FSS) Verification approach (Mittermaier & Roberts) We want to know: How forecast skill varies with neighbourhood size. The smallest neighbourhood size that can be used to give sufficiently accurate forecasts. Does higher resolution provide more accurate forecasts on scales of interest (e.g. river catchments) Compare forecast fractions with fractions from radar over different sized neighbourhoods (squares for convenience) Use rainfall accumulations to apply temporal smoothing

Idealised example

In summary This verification method provides a way of answering some important questions about forecasts from ‘storm-resolving’ NWP models. How does forecast skill vary with spatial scale? At what scales are higher resolution forecasts more skilful (if any)? At what scales are forecasts sufficiently accurate?........... (There are other questions that need different approaches)

See poster by Mittermaier and Thompson How we are using it 12.5 mm 10% threshold FSS 0.5 L(FSS>0.5) 20 km Freq. Bias See poster by Mittermaier and Thompson

THE WINNER 001 - truth 002 - truth 003 - truth > 600 km 200 km

Summary : L(FSS > 0.5) 004 - truth 005 - truth Case > 10 mm 001 200 km 002 > 600 km 003 004 550 km 005 ? 550 km ?

Extreme Rainfall Alerts What are they? Based on sophisticated algorithms to generate first-guess probabilities which can be forecaster-modified Alerts are issued at the county scale. Can be updated or cancelled. Extreme Rainfall Alerts Advisory Early Imminent 30 mm/h or 40 mm/3h or 50 mm/6h Very low but prob >= 10% Low with prob 20-40% Moderate with prob >= 40% Issued 14LT valid for the 24h starting from the next midnight Lead time of 8 – 11h Lead time of 1-3h

Caveat: both of these approaches are inherently deterministic Two approaches have been considered … taking the “event” view, and (did an event occur anywhere in the alert area during the time that the alert was in force) taking the “time series” (continuum) view (comparing the county accumulation totals hour-by-hour during the time that the alert was in force to establish if the threshold was exceeded) Caveat: both of these approaches are inherently deterministic

ET verification plans SRNWP EUMETNET comparison QC-ed Results by end of year Add others next year ? Agree on best (better ?) methods for high resolution forecast verification Link to operations – best methods of presenting forecasts, especially high resolution and EPS © Crown copyright Met Office

Structure,S, Amplitude, A, Location , L (=L1+L2) Wernli, Paulat, Hagen, Frei, 2008 (MWR) © Crown copyright Met Office