By Nicholas Leonardo Advisor: Brian Colle

Slides:



Advertisements
Similar presentations
National Hurricane Center 2008 Forecast Verification James L. Franklin Branch Chief, Hurricane Specialists Unit National Hurricane Center 2009 Interdepartmental.
Advertisements

Munehiko Yamaguchi Typhoon Research Department, Meteorological Research Institute of the Japan Meteorological Agency 9:00 – 12: (Thr) Topic.
Validation of the Ensemble Tropical Rainfall Potential (eTRaP) for Landfalling Tropical Cyclones Elizabeth E. Ebert Centre for Australian Weather and Climate.
HFIP Ensemble Products Subgroup Sept 2, 2011 Conference Call 1.
An Investigation of Cool Season Extratropical Cyclone Forecast Errors Within Operational Models Brian A. Colle 1 and Michael Charles 1,2 1 School of Marine.
United States Coast Guard 1985 Evaluation of a Multi-Model Storm Surge Ensemble for the New York Metropolitan Region Brian A. Colle Tom Di Liberto Stony.
CSTAR Update: New Tools for More Efficient Use of Ensembles in Operations Brian A. Colle, Minghua Zheng, and Edmund K.M. Chang, School of Marine and Atmospheric.
The Relative Contribution of Atmospheric and Oceanic Uncertainty in TC Intensity Forecasts Ryan D. Torn University at Albany, SUNY World Weather Open Science.
Ensemble Post-Processing and it’s Potential Benefits for the Operational Forecaster Michael Erickson and Brian A. Colle School of Marine and Atmospheric.
Forecast Skill and Major Forecast Failures over the Northeastern Pacific and Western North America Lynn McMurdie and Cliff Mass University of Washington.
A. Schumacher, CIRA/Colorado State University NHC Points of Contact: M. DeMaria, D. Brown, M. Brennan, R. Berg, C. Ogden, C. Mattocks, and C. Landsea Joint.
Verification of ensembles Courtesy of Barbara Brown Acknowledgments: Tom Hamill, Laurence Wilson, Tressa Fowler Copyright UCAR 2012, all rights reserved.
How can LAMEPS * help you to make a better forecast for extreme weather Henrik Feddersen, DMI * LAMEPS =Limited-Area Model Ensemble Prediction.
An Improved Wind Probability Program: A Year 2 Joint Hurricane Testbed Project Update Mark DeMaria and John Knaff, NOAA/NESDIS, Fort Collins, CO Stan Kidder,
Verification Approaches for Ensemble Forecasts of Tropical Cyclones Eric Gilleland, Barbara Brown, and Paul Kucera Joint Numerical Testbed, NCAR, USA
Celeste Saulo and Juan Ruiz CIMA (CONICET/UBA) – DCAO (FCEN –UBA)
On the ability of global Ensemble Prediction Systems to predict tropical cyclone track probabilities Sharanya J. Majumdar and Peter M. Finocchio RSMAS.
61 st IHC, New Orleans, LA Verification of the Monte Carlo Tropical Cyclone Wind Speed Probabilities: A Joint Hurricane Testbed Project Update John A.
Short-Range Ensemble Prediction System at INM José A. García-Moya SMNT – INM 27th EWGLAM & 12th SRNWP Meetings Ljubljana, October 2005.
PREDICTABILITY OF WESTERN NORTH PACIFIC TROPICAL CYCLONE EVENTS ON INTRASEASONAL TIMESCALES WITH THE ECMWF MONTHLY FORECAST MODEL Russell L. Elsberry and.
Hurricane Forecast Improvement Project (HFIP): Where do we stand after 3 years? Bob Gall – HFIP Development Manager Fred Toepfer—HFIP Project manager Frank.
Evaluation of Experimental Models for Tropical Cyclone Forecasting in Support of the NOAA Hurricane Forecast Improvement Project (HFIP) Barbara G. Brown,
National Hurricane Center 2010 Forecast Verification James L. Franklin and John Cangialosi Hurricane Specialist Unit National Hurricane Center 2011 Interdepartmental.
Stream 1.5 Runs of SPICE Kate D. Musgrave 1, Mark DeMaria 2, Brian D. McNoldy 1,3, and Scott Longmore 1 1 CIRA/CSU, Fort Collins, CO 2 NOAA/NESDIS/StAR,
Ensemble variability in rainfall forecasts of Hurricane Irene (2011) Molly Smith, Ryan Torn, Kristen Corbosiero, and Philip Pegion NWS Focal Points: Steve.
Page 1. Page 2 German presentations COLIJN Franciscus, GKSS: COSYNA VON STORCH Jin-Song, MPIM: Wind generated power input into the deep ocean VON STORCH.
Predictability of High Impact Weather during the Cool Season: CSTAR Update and the Development of a New Ensemble Sensitivity Tool for the Forecaster Brian.
Impacts of Rossby Wave Packets on Forecast Uncertainties and Errors
Verification of ensemble systems Chiara Marsigli ARPA-SIMC.
Exploring Multi-Model Ensemble Performance in Extratropical Cyclones over Eastern North America and the Western Atlantic Ocean Nathan Korfe and Brian A.
Common verification methods for ensemble forecasts
2015 Production Suite Review: Report from NHC 2015 Production Suite Review: Report from NHC Eric S. Blake, Richard J. Pasch, Andrew Penny NCEP Production.
NCEP CMC ECMWF MEAN ANA BRIAN A COLLE MINGHUA ZHENG │ EDMUND K. CHANG Applying Fuzzy Clustering Analysis to Assess Uncertainty and Ensemble System Performance.
Judith Curry James Belanger Mark Jelinek Violeta Toma Peter Webster 1
VERIFICATION OF A DOWNSCALING SEQUENCE APPLIED TO MEDIUM RANGE METEOROLOGICAL PREDICTIONS FOR GLOBAL FLOOD PREDICTION Nathalie Voisin, Andy W. Wood and.
National Hurricane Center 2009 Forecast Verification James L. Franklin Branch Chief, Hurricane Specialist Unit National Hurricane Center 2009 NOAA Hurricane.
Forecasting Oceanic Cyclones at the NOAA Ocean Prediction Center Joseph M. Sienkiewicz, D. Scott Prosise, and Anthony Crutch NOAA/NWS/NCEP/Ocean Prediction.
Application of the CRA Method Application of the CRA Method William A. Gallus, Jr. Iowa State University Beth Ebert Center for Australian Weather and Climate.
Prediction of Consensus Tropical Cyclone Track Forecast Error ( ) James S. Goerss NRL Monterey March 1, 2011.
Hemispheric Evolution and Origin of Extra-tropical Cyclone Errors Within NCEP Operational Models Brian A. Colle 1 and Michael Charles 1,2 1 School of Marine.
Medium Range Forecasting at the Weather Prediction Center (WPC) –
Notes for Interpretation
A Guide to Tropical Cyclone Guidance
A few examples of heavy precipitation forecast Ming Xue Director
Fuzzy verification using the Fractions Skill Score
Course Evaluation Now online You should have gotten an with link.
Ensemble variability in rainfall forecasts of Hurricane Irene (2011)
Course Evaluation Now online You should have gotten an with link.
Dan Petersen Bruce Veenhuis Greg Carbin Mark Klein Mike Bodner
TC Winds Conference Call
An Analysis of Large Track Error North Atlantic Tropical Cyclones.
Development of an Uncertainty Tool to Assess Model Forecast Parameters
A Review of the CSTAR Ensemble Tools Available for Operations
Storm Surge Forecasting Practices, Tools for Emergency Managers, A Probabilistic Storm Surge Model Based on Ensembles and Past Error Distributions.
Nathalie Voisin, Andy W. Wood and Dennis P. Lettenmaier
Course Evaluation Now online You should have gotten an with link.
Forecast Pressure.
MOS Developed by and Run at the NWS Meteorological Development Lab (MDL) Full range of products available at:
Post Processing.
Motivation Investigation of predictability using ensemble forecasts
Application of a global probabilistic hydrologic forecast system to the Ohio River Basin Nathalie Voisin1, Florian Pappenberger2, Dennis Lettenmaier1,
Michael J. Brennan National Hurricane Center
Science Objectives contained in three categories
Christoph Gebhardt, Zied Ben Bouallègue, Michael Buchhold
Status Report of T-PARC/TCS-08
Tropical storm intra-seasonal prediction
Verification of Tropical Cyclone Forecasts
Short Range Ensemble Prediction System Verification over Greece
The Impact of Moist Singular Vectors and Horizontal Resolution on Short-Range Limited-Area Ensemble Forecasts for Extreme Weather Events A. Walser1) M.
Ryan Kang, Wee Leng Tan, Thea Turkington, Raizan Rahmat
Presentation transcript:

By Nicholas Leonardo Advisor: Brian Colle Verification of Multi-Model Ensemble Forecasts of North Atlantic Tropical Cyclones and an Analysis of Large Track Error Cases By Nicholas Leonardo Advisor: Brian Colle

Motivation Evacuation models are heavily dependent on track forecasts made 3-5 days in advance. On average, deterministic track forecasts have improved over the past decades. There are no comprehensive verification of multi-model ensemble systems spanning several years.

Motivation Isabel (2003) Joaquin (2015) Would need more time to find/make a plot for Mathew (2016) instead... It however seemed more like a >5day issue when it was crossing over the DR region; otherwise I have yet to see if it is a “large error” case (eg. which forecast(s) fit the criteria I’ve set) Some storms are less predictable than others, resulting in large track and/or intensity errors. WRF-ARW initialized 2003 Sept 15th 0000UTC. Every 6 hours plotted up to Sept 19th 1200UTC.

Outline Objectives Data and Methods Model Verification Probabilistic Scores Large Error Cases Conclusions

Objectives Verify 3 global ensembles and several deterministic models for North Atlantic hurricane tracks, focusing on 3-5 day forecasts over the past 8 years. Determine whether the ensembles have probabilistic skill relative to the EC deterministic or the NHC uncertainty cones. Identify the outlier storms in terms of track error. Determine what common patterns might be contributing to these large error cases. Do the sources of error tend to be local or remote relative to the TC?

Data and Methods Forecast cyclone tracks from the ECMWF (51 members), GEFS (21 members), and UKMET (23 members) ensemble prediction systems (http://rda.ucar.edu/datasets/ds330.3/), and those from the deterministic CMC, ECdet, GFSdet, UKMdet, NGPS, GFDL, HWRF, and the NHC’s official (OFCL) forecasts (ftp://ftp.nhc.noaa.gov/atcf/archive/). Verified with the NHC best track data for 109 named storms from 2008 to 2015.

2008-2015 Mean Track Error vs. Lead Time Model Verification Homogenous comparison of track errors as a function of lead time. Included the combination of the 3 ensemble prediction systems or grand ensemble (“GE”). The ECMWF deterministic (“ECdet") does even better than the GE mean. ECdet also has a higher percentage superiority among the models. The grand ensemble is never the worse. The high-resolution coupled HWRF struggles. Lead Time (h) Percentage Best Track Forecasts Percentage Worst Track Forecasts Lead Time (h) Lead Time (h)

Model Verification There appears to be a slow bias in all of the models. The average cross track errors are smaller in magnitude, though the GEFS/GFSdet show a right-of-track bias at 120h 2008-2015 Along-Track Error vs. Lead Time 2008-2015 Cross-Track Error vs. Lead Time

Model Verification Some models have a decrease in track error, especially 2010-2012. This corresponds to an overall improvement in the slow bias. Annual Averages of 72h-120h Total Track Errors Annual Averages of 72h-120h Along-Track Errors

Model Verification Much of the slow bias can be attributed to extratropical transitions (“ET”; eg. Buckingham et al. 2010). Excluding the verification of observed ET events reduces this mean bias in all of the models. Mean Along-Track Error Mean Along-Track Error (w/out ET)

Probabilistic Scores Brier Skill Score vs. Threshold Distance Evaluated the ensembles with the Brier Skill Score: Summed all n 72-120h forecast probabilities ( f ) against observations (o) that a TC will come within a threshold distance from a point. Used the ECdet as the reference. GE and ECMWF show the most skill. UKMET struggles for larger threshold distances. _

Probabilistic Scores BSS (96h-120h) Cone radius Used the OFCL (with climo-based cone) as the reference: a randomly-generated 1000-member ensemble with a Gaussian distribution of distances from the OFCL position in vector space. The standard deviation of these distances is specified by the cone radius. Compared 96h/120h forecasts of the ensembles with OFCL; each forecast pair having the same initialization date. GE and ECMWF show skill over the baseline for distances >200km. UKMET tends to struggle the most.

Probabilistic Scores BSS (72h-96h; ensemble interpolated) Reliability Diagram (d < 500km) Repeated the experiment, but comparing the 72h/96h OFCL forecast with the 84h/108h ensemble forecasts initialized 12h earlier (interpolated). The GE and ECMWF are only comparable to the baseline. The UKMET again has poorer skill. Reliability diagrams imply GEFS and UKMET over-forecasts high hit probabilities.

Probabilistic Scores Not terribly important; delete to save a minute? Rank histograms show all ensembles are under-dispersed in the along-track direction, with a clear slow-bias. ECMWF is slightly over-dispersed in the cross-track, while UKMET and GEFS are under-dispersed.

Large Error Cases Small Error Cases Large Error Cases Nfcst = 30 MAE = 142km Nfcst = 31 MAE = 707km Sorted maximum track errors of the GE mean beyond 72h and took the top (bottom) 25% largest errors. Only included two different forecasts for the same TC if they were separated by >5 days. Initializations of many large error cases appear to be clustered north and east of Puerto Rico and associated with recurving TC’s Most small error cases did not move far throughout the forecast and/or were mostly driven by steady easterlies

Large Error Cases Used ensemble sensitivity analysis (ESA): correlating ensemble gridded fields with peak member track errors (eg. Chang et al., 2013). Also analyzed standardized differences (eg. Torn et al. 2015). Members were sorted by track errors. The mean field of the 10 “good” members is subtracted from that of the 10 “bad” and normalized by the full sample standard deviation Tracked the uncertainty sources related to each large error case to earlier forecast hours. The cases were classified based on the initial proximity of the identified source(s) relative to the TC. The Number of Large Track Error Cases With Uncertainty Source Types Source Type Number Remote Upstream 3 Nearby Upstream 10 Nearby Downstream 9 TC-Local 14 Undefined

Large Error Cases Type 1: Remote Upstream Uncertainty: Associated with some synoptic feature >5000km upstream from the TC (eg. wave-guide-like pattern over the eastern North Pacific) . Ensemble Mean 300mb Heights and ESA with Max Track Error (024h) Step through very quickly... Do so with the other 4 we discussed (>1 minute to explain each)

Large Error Cases Type 1: Remote Upstream Uncertainty: Associated with some synoptic feature >5000km upstream from the TC (eg. wave-guide-like pattern over the eastern North Pacific) . Ensemble Mean 300mb Heights and GOOD10-BAD10 Difference (024h) Step through very quickly... Do so with the other 4 we discussed (>1 minute to explain each)

Large Error Cases Type 1: Remote Upstream Uncertainty: Associated with some synoptic feature >5000km upstream from the TC (eg. wave-guide-like pattern over the eastern North Pacific) . Ensemble Mean 300mb Heights and GOOD10-BAD10 Difference (048h)

Large Error Cases Type 1: Remote Upstream Uncertainty: Associated with some synoptic feature >5000km upstream from the TC (eg. wave-guide-like pattern over the eastern North Pacific) . Ensemble Mean 300mb Heights and GOOD10-BAD10 Difference (060h)

Large Error Cases Type 1: Remote Upstream Uncertainty: Associated with some synoptic feature >5000km upstream from the TC (eg. wave-guide-like pattern over the eastern North Pacific) . Ensemble Mean 300mb Heights and GOOD10-BAD10 Difference (072h)

Large Error Cases Type 1: Remote Upstream Uncertainty: Associated with some synoptic feature >5000km upstream from the TC (eg. wave-guide-like pattern over the eastern North Pacific) . Ensemble Mean 300mb Heights and GOOD10-BAD10 Difference (096h)

Large Error Cases Type 2: Nearby Upstream Uncertainty: Uncertainty attached to an upstream feature that directly steers the TC at a later time. Ensemble Mean 300mb Heights and ESA with Max Track Error (024h)

Large Error Cases Type 2: Nearby Upstream Uncertainty: Uncertainty attached to an upstream feature that directly steers the TC at a later time. Ensemble Mean 300mb Heights and GOOD10-BAD10 Difference (024h)

Large Error Cases Type 2: Nearby Upstream Uncertainty: Uncertainty attached to an upstream feature that directly steers the TC at a later time. Ensemble Mean 300mb Heights and GOOD10-BAD10 Difference (048h)

Large Error Cases Type 2: Nearby Upstream Uncertainty: Uncertainty attached to an upstream feature that directly steers the TC at a later time. Ensemble Mean 300mb Heights and GOOD10-BAD10 Difference (072h)

Large Error Cases Type 2: Nearby Upstream Uncertainty: Uncertainty attached to an upstream feature that directly steers the TC at a later time. Ensemble Mean 300mb Heights and GOOD10-BAD10 Difference (096h)

Large Error Cases Type 3: Nearby Downstream Uncertainty: Uncertainty attached to a downstream (east of the TC’s longitude) feature steering the TC. Ensemble Mean 300mb Heights and ESA with Max Track Error (0h)

Large Error Cases Type 3: Nearby Downstream Uncertainty: Uncertainty attached to a downstream (east of the TC’s longitude) feature steering the TC. Ensemble Mean 300mb Heights and GOOD10-BAD10 Difference (036h)

Large Error Cases Type 3: Nearby Downstream Uncertainty: Uncertainty attached to a downstream (east of the TC’s longitude) feature steering the TC. Ensemble Mean 300mb Heights and GOOD10-BAD10 Difference (048h)

Large Error Cases Type 3: Nearby Downstream Uncertainty: Uncertainty attached to a downstream (east of the TC’s longitude) feature steering the TC. Ensemble Mean 300mb Heights and GOOD10-BAD10 Difference (072h)

Large Error Cases Type 4: TC-related: The signal first appears near the TC or in the immediate TC environment (maybe driven by the TC itself). Ensemble Mean 300mb Heights and ESA with Max Track Error (036h)

Large Error Cases Type 4: TC-related: The signal first appears near the TC or in the immediate TC environment (maybe driven by the TC itself). Ensemble Mean 300mb Heights and GOOD10-BAD10 Difference (048h)

Large Error Cases Type 4: TC-related: The signal first appears near the TC or in the immediate TC environment (maybe driven by the TC itself). Ensemble Mean 300mb Heights and GOOD10-BAD10 Difference (072h)

Large Error Cases Type 4: TC-related: The signal first appears near the TC or in the immediate TC environment (maybe driven by the TC itself). Ensemble Mean 300mb Heights and GOOD10-BAD10 Difference (096h)

Conclusions The ECMWF ensemble has among the smallest MAE’s for TC track. All of the models have a slow bias in along-track at days 3-5, largely due to TC’s undergoing extratropical transitions. Track forecasts have improved in HWRF and GEFS during the last several years, with GEFS MAE’s becoming more comparable to the ECMWF. The ECMWF and GE have the best probabilistic forecasts. Their skill is better than the deterministic ECdet and comparable to the OFCL cone forecast (even after interpolating the ensemble). The largest track error cases appear to be clustered north and east of Puerto Rico, associated with recurving TCs. For many of these cases, the physical sources of uncertainty associated with the track can be traced back to processes occurring in the near-TC environment and to the phase and amplitude of upstream steering trough/ridge systems at earlier lead times.

References Buckingham, C., T. Marchok, I. Ginis, L. Rothstein, and D. Rowe, 2010: Short- and Medium-Range Prediction of Tropical and Transitioning Cyclone Tracks within the NCEP Global Ensemble Forecasting System. Weather and Forecasting, 25, 1741–1742. Chang, E. K. M., M. Zheng, and K. Raeder, 2013: Medium-Range Ensemble Sensitivity Analysis of Two Extreme Pacific Extratropical Cyclones. Mon. Wea. Rev., 141, 211–231, doi:10.1175/MWR-D-11-00304.1. Torn, R. D., J. S. Whitaker, P. Pegion, T. M. Hamill, and G. J. Hakim, 2015: Diagnosis of the Source of GFS Medium-Range Track Errors in Hurricane Sandy (2012). Mon. Wea. Rev., 143, 132–152, doi:10.1175/MWR-D-14-00086.1