Mesoscale Model Evaluation Mike Baldwin Cooperative Institute for Mesoscale Meteorological Studies, University of Oklahoma Also affiliated with NOAA/NSSL.

Slides:



Advertisements
Similar presentations
Introduction to data assimilation in meteorology Pierre Brousseau, Ludovic Auger ATMO 08,Alghero, september 2008.
Advertisements

Report of the Q2 Short Range QPF Discussion Group Jon Ahlquist Curtis Marshall John McGinley - lead Dan Petersen D. J. Seo Jean Vieux.
6th WMO tutorial Verification Martin GöberContinuous 1 Good afternoon! नमस्कार नमस्कार Guten Tag! Buenos dias! до́брый день! до́брыйдень Qwertzuiop asdfghjkl!
Validation of Satellite Precipitation Estimates for Weather and Hydrological Applications Beth Ebert BMRC, Melbourne, Australia 3 rd IPWG Workshop / 3.
The Global Digital Elevation Model (GTOPO30) of Great Basin Location: latitude 38  15’ to 42  N, longitude 118  30’ to 115  30’ W Grid size: 925 m.
WRF Verification: New Methods of Evaluating Rainfall Prediction Chris Davis NCAR (MMM/RAP) Collaborators: Dave Ahijevych, Mike Baldwin, Barb Brown, Randy.
Assessment of Tropical Rainfall Potential (TRaP) forecasts during the Australian tropical cyclone season Beth Ebert BMRC, Melbourne, Australia.
Monitoring the Quality of Operational and Semi-Operational Satellite Precipitation Estimates – The IPWG Validation / Intercomparison Study Beth Ebert Bureau.
Improving Excessive Rainfall Forecasts at HPC by using the “Neighborhood - Spatial Density“ Approach to High Res Models Michael Eckert, David Novak, and.
Validation of the Ensemble Tropical Rainfall Potential (eTRaP) for Landfalling Tropical Cyclones Elizabeth E. Ebert Centre for Australian Weather and Climate.
PROVIDING DISTRIBUTED FORECASTS OF PRECIPITATION USING A STATISTICAL NOWCAST SCHEME Neil I. Fox and Chris K. Wikle University of Missouri- Columbia.
Institut für Physik der Atmosphäre Predictability of precipitation determined by convection-permitting ensemble modeling Christian Keil and George C.Craig.
Raster Data. The Raster Data Model The Raster Data Model is used to model spatial phenomena that vary continuously over a surface and that do not have.
Chapter 1 Ways of Seeing. Ways of Seeing the Atmosphere The behavior of the atmosphere is very complex. Different ways of displaying the characteristics.
Exploring the Use of Object- Oriented Verification at the Hydrometeorological Prediction Center Faye E. Barthold 1,2, Keith F. Brill 1, and David R. Novak.
Univ of AZ WRF Model Verification. Method NCEP Stage IV data used for precipitation verification – Stage IV is composite of rain fall observations and.
The Calibration Process
Warm Season Thunderstorm Patterns Over the New Jersey Area Al Cope Paul Croft National Weather Service Kean University Mount Holly, NJ Union, NJ.
Ensemble Post-Processing and it’s Potential Benefits for the Operational Forecaster Michael Erickson and Brian A. Colle School of Marine and Atmospheric.
Lecture II-2: Probability Review
Elec471 Embedded Computer Systems Chapter 4, Probability and Statistics By Prof. Tim Johnson, PE Wentworth Institute of Technology Boston, MA Theory and.
Tomer Sagi and Avigdor Gal Technion - Israel Institute of Technology Non-binary Evaluation for Schema Matching ER 2012 October 2012, Florence.
WWOSC 2014 Assimilation of 3D radar reflectivity with an Ensemble Kalman Filter on a convection-permitting scale WWOSC 2014 Theresa Bick 1,2,* Silke Trömel.
1 Reading Report 9 Yin Chen 29 Mar 2004 Reference: Multivariate Resource Performance Forecasting in the Network Weather Service, Martin Swany and Rich.
Oceanography 569 Oceanographic Data Analysis Laboratory Kathie Kelly Applied Physics Laboratory 515 Ben Hall IR Bldg class web site: faculty.washington.edu/kellyapl/classes/ocean569_.
On Estimation of Surface Soil Moisture from SAR Jiancheng Shi Institute for Computational Earth System Science University of California, Santa Barbara.
4th Int'l Verification Methods Workshop, Helsinki, 4-6 June Methods for verifying spatial forecasts Beth Ebert Centre for Australian Weather and.
Hydrologic Modeling: Verification, Validation, Calibration, and Sensitivity Analysis Fritz R. Fiedler, P.E., Ph.D.
The Importance of Atmospheric Variability for Data Requirements, Data Assimilation, Forecast Errors, OSSEs and Verification Rod Frehlich and Robert Sharman.
Why Is It There? Getting Started with Geographic Information Systems Chapter 6.
Texture. Texture is an innate property of all surfaces (clouds, trees, bricks, hair etc…). It refers to visual patterns of homogeneity and does not result.
Federal Department of Home Affairs FDHA Federal Office of Meteorology and Climatology MeteoSwiss High-resolution data assimilation in COSMO: Status and.
Development of an object- oriented verification technique for QPF Michael Baldwin 1 Matthew Wandishin 2, S. Lakshmivarahan 3 1 Cooperative Institute for.
Towards an object-oriented assessment of high resolution precipitation forecasts Janice L. Bytheway CIRA Council and Fellows Meeting May 6, 2015.
VERIFICATION OF NDFD GRIDDED FORECASTS IN THE WESTERN UNITED STATES John Horel 1, David Myrick 1, Bradley Colman 2, Mark Jackson 3 1 NOAA Cooperative Institute.
Comparison of convective boundary layer velocity spectra calculated from large eddy simulation and WRF model data Jeremy A. Gibbs and Evgeni Fedorovich.
Model Construction: interpolation techniques 1392.
On the spatial verification of FROST-2014 precipitation forecast fields Anatoly Muraviev (1), Anastasia Bundel (1), Dmitry Kiktev (1), Nikolay Bocharnikov.
Ebert-McBride Technique (Contiguous Rain Areas) Ebert and McBride (2000: Verification of precipitation in weather systems: determination of systematic.
Spatial Interpolation III
Real-time Verification of Operational Precipitation Forecasts using Hourly Gauge Data Andrew Loughe Judy Henderson Jennifer MahoneyEdward Tollerud Real-time.
Sensitivity Analysis of Mesoscale Forecasts from Large Ensembles of Randomly and Non-Randomly Perturbed Model Runs William Martin November 10, 2005.
Data assimilation, short-term forecast, and forecasting error
Assimilation of HF radar in the Ligurian Sea Spatial and Temporal scale considerations L. Vandenbulcke, A. Barth, J.-M. Beckers GHER/AGO, Université de.
A QPE Product with Blended Gage Observations and High-Resolution WRF Ensemble Model Output: Comparison with Analyses and Verification during the HMT-ARB.
Experiences in assessing deposition model uncertainty and the consequences for policy application Rognvald I Smith Centre for Ecology and Hydrology, Edinburgh.
An Object-Based Approach for Identifying and Evaluating Convective Initiation Forecast Impact and Quality Assessment Section, NOAA/ESRL/GSD.
Feature-based (object-based) Verification Nathan M. Hitchens National Severe Storms Laboratory.
Verification of Precipitation Areas Beth Ebert Bureau of Meteorology Research Centre Melbourne, Australia
Chapter 20 Classification and Estimation Classification – Feature selection Good feature have four characteristics: –Discrimination. Features.
Object-oriented verification of WRF forecasts from 2005 SPC/NSSL Spring Program Mike Baldwin Purdue University.
LWG, Destin (Fl) 27/1/2009 Observation representativeness error ECMWF model spectra Application to ADM sampling mode and Joint-OSSE.
U. Damrath, COSMO GM, Athens 2007 Verification of numerical QPF in DWD using radar data - and some traditional verification results for surface weather.
Intermittency Analysis and Spatial Dependence of Magnetic Field Disturbances in the Fast Solar Wind Sunny W. Y. Tam 1 and Ya-Hui Yang 2 1 Institute of.
Nathalie Voisin 1, Florian Pappenberger 2, Dennis Lettenmaier 1, Roberto Buizza 2, and John Schaake 3 1 University of Washington 2 ECMWF 3 National Weather.
Page 1© Crown copyright 2004 The use of an intensity-scale technique for assessing operational mesoscale precipitation forecasts Marion Mittermaier and.
A Random Subgrouping Scheme for Ensemble Kalman Filters Yun Liu Dept. of Atmospheric and Oceanic Science, University of Maryland Atmospheric and oceanic.
WRF Verification Toolkit Workshop, Boulder, February 2007 Spatial verification of NWP model fields Beth Ebert BMRC, Australia.
NCAR, 15 April Fuzzy verification of fake cases Beth Ebert Center for Australian Weather and Climate Research Bureau of Meteorology.
Verification methods - towards a user oriented verification The verification group.
User-Focused Verification Barbara Brown* NCAR July 2006
Application of the CRA Method Application of the CRA Method William A. Gallus, Jr. Iowa State University Beth Ebert Center for Australian Weather and Climate.
Intensity-scale verification technique
Systematic timing errors in km-scale NWP precipitation forecasts
Verifying Precipitation Events Using Composite Statistics
Verifying and interpreting ensemble products
General framework for features-based verification
Composite-based Verification
Hydrologically Relevant Error Metrics for PEHRPP
the performance of weather forecasts
Presentation transcript:

Mesoscale Model Evaluation Mike Baldwin Cooperative Institute for Mesoscale Meteorological Studies, University of Oklahoma Also affiliated with NOAA/NSSL and NOAA/NWS/SPC

NWS – forecasts on hi-res grids What would you suggest that NWS do to verify these forecasts?

Issues in mesoscale verification Validate natural behavior of forecasts Realistic variability, structure of fields Do predicted events occur with realistic frequency? Do characteristics of phenomena mimic those found in nature? Traditional objective verification techniques are not able to address these issues

Outline Problems with traditional verification Solutions: Verify characteristics of phenomena Verify structure/variability Design verification systems that address value of forecasts

Traditional verification Compare a collection of matching pairs of forecast and observed values at the same set of points in space/time Compute various measures of accuracy: RMSE, bias, equitable threat score A couple of numbers may represent the accuracy of millions of model grid points, thousands of cases, hundreds of meteorological events Boiling down that much information into one or two numbers is not very meaningful

Dimensionality of verification info Murphy (1991) and others highlight danger of simplifying complex verification information High-dimension information = data overload Verification information should be easy to understand Need to find ways to measure specific aspects of performance

Quality vs. value Scores typically measure quality, or degree in which forecasts and observations agree Forecast value is benefit of forecast information to decision maker Value is subjective, complex function of quality High-quality forecast may be of low value and vice versa

Forecast #1: smooth OBSERVED FCST #1: smooth FCST #2: detailed OBSERVED

Traditional “measures-oriented” approach to verifying these forecasts Verification MeasureForecast #1 (smooth) Forecast #2 (detailed) Mean absolute error RMS error Bias 0.98 Threat score (>0.45) Equitable threat score (>0.45)

Phase/timing errors High-amplitude, small-scale forecast and observed fields are most sensitive to timing/phase errors

Mean Squared Error (MSE) For 1 point phase error MSE =

Mean Squared Error (MSE) For 1 point phase error MSE = 0.165

Mean Squared Error (MSE) For 1 point phase error MSE = 1.19

Verify forecast “realism” Anthes (1983) paper suggests several ways to verify “realism” Verify characteristics of phenomena Decompose forecast errors as function of spatial scale Verify structure/variance spectra

Characterize the forecast and observed fields Verify the forecast with a similar approach as a human forecaster might visualize the forecast/observed fields Characterize features, phenomena, events, etc. found in forecast and observed fields by assigning attributes to each object Not an unfamiliar concept: “1050 mb high” “category 4 hurricane” “F-4 tornado”

Many possible ways to characterize phenomena Shape, orientation, size, amplitude, location Flow pattern Subjective information (confidence, difficulty) Physical processes in a NWP model Verification information can be stratified using this additional information

“Object-oriented” approach to verification Decompose fields into sets of “objects” that are identified and described by a set of attributes in an automated fashion Using image processing techniques to locate and identify events Produce “scores” or “metrics” based upon the similarity/dissimilarity between forecast and observed events Could also examine the joint distribution of forecast and observed events

Characterization: How? Identify an object Usually involves complex image processing Event #16

Characterization: How? Assign attributes Examples: location, mean, orientation, structure Event #16: Lat=37.3N, Lon=87.8W,  =22.3,  =2.1

Automated rainfall object identification Contiguous regions of measurable rainfall (similar to CRA; Ebert and McBride (2000))

Connected component labeling

Expand area by 15%, connect regions that are within 20km, relabel

Object characterization Compute attributes

Verification of detailed forecasts 12h forecasts of 1h precipitation valid 00Z 24 Apr 2003 observed fcst #1 RMSE = 3.4 MAE = 0.97 ETS = 0.06 RMSE = 1.7 MAE = 0.64 ETS = 0.00 fcst #2

Verification 12h forecasts of 1h precipitation valid 00Z 24 Apr 2003 observed fcst #1 fcst #2  = 3.1 ecc 20 = 2.6 ecc 40 = 2.0 ecc 60 = 2.1 ecc 80 = 2.8  = 1.6 ecc 20 = 10.7 ecc 40 = 7.5 ecc 60 = 4.3 ecc 80 = 2.8  = 7.8 ecc 20 = 3.6 ecc 40 = 3.1 ecc 60 = 4.5 ecc 80 = 3.6

Example of scores produced by this approach f i = (a i, b i, c i, …, x i, y i ) t o k = (a k, b k, c k, …, x k, y k ) t d i,k (f i,o k ) = (f i -o k ) t A (f i -o k ) (Generalized Euclidean distance, measure of dissimilarity) where A is a matrix, different attributes would probably have different weights c i,k (f i,o k ) = cov(f i,o k ) (measure of similarity)

Ebert and McBride (2000) Contiguous Rain Areas Separate errors into amplitude, displacement, shape components

Contour error map (CEM) method Case et al (2003) Phenomena of interest – Florida sea breeze Object identification – sea breeze transition time Contour map of transition time errors Distributions of timing errors Verify post-sea breeze winds

Compositing Nachamkin (2004) Identify events of interest in the forecasts Collect coordinated samples Compare forecast PDF to observed PDF Repeat process for observed events

Decompose errors as a function of scale Bettge and Baumhefner (1980) used band-pass filters to analyze errors at different scales Briggs and Levine (1997) used wavelet analysis of forecast errors

Verify structure Fourier energy spectra Take Fourier transform, multiply by complex conjugate – E(k) Display on log-log plot Natural phenomena often show “power-law” regimes Noise (uncorrelated) results in flat spectrum

Fourier spectra Slope of spectrum indicates degree of structure in the data

Larger absolute values of slope correspond with less structure slope = -1 slope = -3 slope = -1.5 noise

Multiscale statistical properties (Harris et al 2001) Fourier energy spectrum Generalized structure function: spatial correlation Moment-scale analysis: intermittency of a field, sparseness of sharp intensities Looking for “power law”, much like in atmospheric turbulence (–5/3 slope) FIG. 3. Isotropic spatial Fourier power spectral density (PSD) for forecast RLW (qr; dotted line) and radar-observed qr (solid line). Comparison of the spectra shows reasonable agreement at scales larger than 15 km. For scales smaller than 15 km, the forecast shows a rapid falloff in variability in comparison with the radar. The estimated spectral slope with fit uncertainty is = 3.0 ± 0.1

Example log[E(k)] log[wavenumber] Obs_4Eta_12Eta_8 WRF_22WRF_10KF_22 3-6h forecasts from 04 June UTC

June z runs 12, 24, 36, 48h fcsts

Comparing forecasts that contain different degrees of structure Obs=black Detailed = blue Smooth = green MSE detailed = 1.57 MSE smooth = 1.43

Common resolved scales vs. unresolved Filter other forecasts to have same structure MSE “detailed” = 1.32 MSE smooth = 1.43

Lack of detail in analyses Methods discussed assume realistic analysis of observations Problems: Relatively sparse observations Operational data assimilation systems Smooth first guess fields from model forecasts Smooth error covariance matrix Smooth analysis fields result

True mesoscale analyses Determine what scales are resolved Mesoscale data assimilation Frequent updates All available observations Hi-res NWP provides first guess Ensemble Kalman filter Tustison et al. (2002) scale-recursive filtering takes advantage of natural “scaling”

Design verification systems that address forecast value Value measures the benefits of forecast information to users Determine what aspects of forecast users are most sensitive to If possible, find out users “cost/loss” situation Are missed events or false alarms more costly?

Issues How to distill the huge amount of verification information into meaningful “nuggets” that can be used effectively? How to elevate verification from an annoyance to an integral part of the forecast process? What happens when conflicting information from different verification approaches is obtained?

Summary Problems with traditional verification techniques when used with forecasts/observations with structure Verify realism Issues of scale Work with forecasters/users to determine most important aspects of forecast information

References Good books Papers mentioned in this presentation Beth Ebert’s website

Scores based on similarity/dissimilarity matrices D = [d i,j ] euclidean distance matrix C = [c i,j ] covariance matrix Scores could be: tr[D] = trace of matrix, for euclidean distance this equates to  (f i – o i ) 2 ~ RMSE det[D] = determinant of matrix, a measure of the magnitude of a matrix

Fourier power spectra Compare 3h accumulated QPF to radar/gage analyses Forecasts were linearly interpolated to same 4km grid as “Stage IV” analysis Errico (1985) Fourier analysis code used. 2-d Fourier transform converted to 1-d by annular average Fixed grid used for analysis located away from complex terrain of Western U.S. Want to focus on features generated by model physics and dynamics, free from influence of orographically forced circulations