Download presentation
Presentation is loading. Please wait.
Published byHollie Wilkins Modified over 9 years ago
1
Mesoscale Model Evaluation Mike Baldwin Cooperative Institute for Mesoscale Meteorological Studies, University of Oklahoma Also affiliated with NOAA/NSSL and NOAA/NWS/SPC
2
NWS – forecasts on hi-res grids What would you suggest that NWS do to verify these forecasts?
3
Issues in mesoscale verification Validate natural behavior of forecasts Realistic variability, structure of fields Do predicted events occur with realistic frequency? Do characteristics of phenomena mimic those found in nature? Traditional objective verification techniques are not able to address these issues
4
Outline Problems with traditional verification Solutions: Verify characteristics of phenomena Verify structure/variability Design verification systems that address value of forecasts
5
Traditional verification Compare a collection of matching pairs of forecast and observed values at the same set of points in space/time Compute various measures of accuracy: RMSE, bias, equitable threat score A couple of numbers may represent the accuracy of millions of model grid points, thousands of cases, hundreds of meteorological events Boiling down that much information into one or two numbers is not very meaningful
6
Dimensionality of verification info Murphy (1991) and others highlight danger of simplifying complex verification information High-dimension information = data overload Verification information should be easy to understand Need to find ways to measure specific aspects of performance
7
Quality vs. value Scores typically measure quality, or degree in which forecasts and observations agree Forecast value is benefit of forecast information to decision maker Value is subjective, complex function of quality High-quality forecast may be of low value and vice versa
8
Forecast #1: smooth OBSERVED FCST #1: smooth FCST #2: detailed OBSERVED
9
Traditional “measures-oriented” approach to verifying these forecasts Verification MeasureForecast #1 (smooth) Forecast #2 (detailed) Mean absolute error 0.157 0.159 RMS error 0.254 0.309 Bias 0.98 Threat score (>0.45) 0.214 0.161 Equitable threat score (>0.45) 0.170 0.102
10
Phase/timing errors High-amplitude, small-scale forecast and observed fields are most sensitive to timing/phase errors
11
Mean Squared Error (MSE) For 1 point phase error MSE = 0.0016
12
Mean Squared Error (MSE) For 1 point phase error MSE = 0.165
13
Mean Squared Error (MSE) For 1 point phase error MSE = 1.19
14
Verify forecast “realism” Anthes (1983) paper suggests several ways to verify “realism” Verify characteristics of phenomena Decompose forecast errors as function of spatial scale Verify structure/variance spectra
15
Characterize the forecast and observed fields Verify the forecast with a similar approach as a human forecaster might visualize the forecast/observed fields Characterize features, phenomena, events, etc. found in forecast and observed fields by assigning attributes to each object Not an unfamiliar concept: “1050 mb high” “category 4 hurricane” “F-4 tornado”
16
Many possible ways to characterize phenomena Shape, orientation, size, amplitude, location Flow pattern Subjective information (confidence, difficulty) Physical processes in a NWP model Verification information can be stratified using this additional information
17
“Object-oriented” approach to verification Decompose fields into sets of “objects” that are identified and described by a set of attributes in an automated fashion Using image processing techniques to locate and identify events Produce “scores” or “metrics” based upon the similarity/dissimilarity between forecast and observed events Could also examine the joint distribution of forecast and observed events
18
Characterization: How? Identify an object Usually involves complex image processing Event #16
19
Characterization: How? Assign attributes Examples: location, mean, orientation, structure Event #16: Lat=37.3N, Lon=87.8W, =22.3, =2.1
20
Automated rainfall object identification Contiguous regions of measurable rainfall (similar to CRA; Ebert and McBride (2000))
21
Connected component labeling
22
Expand area by 15%, connect regions that are within 20km, relabel
23
Object characterization Compute attributes
24
Verification of detailed forecasts 12h forecasts of 1h precipitation valid 00Z 24 Apr 2003 observed fcst #1 RMSE = 3.4 MAE = 0.97 ETS = 0.06 RMSE = 1.7 MAE = 0.64 ETS = 0.00 fcst #2
25
Verification 12h forecasts of 1h precipitation valid 00Z 24 Apr 2003 observed fcst #1 fcst #2 = 3.1 ecc 20 = 2.6 ecc 40 = 2.0 ecc 60 = 2.1 ecc 80 = 2.8 = 1.6 ecc 20 = 10.7 ecc 40 = 7.5 ecc 60 = 4.3 ecc 80 = 2.8 = 7.8 ecc 20 = 3.6 ecc 40 = 3.1 ecc 60 = 4.5 ecc 80 = 3.6
26
Example of scores produced by this approach f i = (a i, b i, c i, …, x i, y i ) t o k = (a k, b k, c k, …, x k, y k ) t d i,k (f i,o k ) = (f i -o k ) t A (f i -o k ) (Generalized Euclidean distance, measure of dissimilarity) where A is a matrix, different attributes would probably have different weights c i,k (f i,o k ) = cov(f i,o k ) (measure of similarity)
27
Ebert and McBride (2000) Contiguous Rain Areas Separate errors into amplitude, displacement, shape components
28
Contour error map (CEM) method Case et al (2003) Phenomena of interest – Florida sea breeze Object identification – sea breeze transition time Contour map of transition time errors Distributions of timing errors Verify post-sea breeze winds
29
Compositing Nachamkin (2004) Identify events of interest in the forecasts Collect coordinated samples Compare forecast PDF to observed PDF Repeat process for observed events
30
Decompose errors as a function of scale Bettge and Baumhefner (1980) used band-pass filters to analyze errors at different scales Briggs and Levine (1997) used wavelet analysis of forecast errors
33
Verify structure Fourier energy spectra Take Fourier transform, multiply by complex conjugate – E(k) Display on log-log plot Natural phenomena often show “power-law” regimes Noise (uncorrelated) results in flat spectrum
34
Fourier spectra Slope of spectrum indicates degree of structure in the data
35
Larger absolute values of slope correspond with less structure slope = -1 slope = -3 slope = -1.5 noise
36
Multiscale statistical properties (Harris et al 2001) Fourier energy spectrum Generalized structure function: spatial correlation Moment-scale analysis: intermittency of a field, sparseness of sharp intensities Looking for “power law”, much like in atmospheric turbulence (–5/3 slope) FIG. 3. Isotropic spatial Fourier power spectral density (PSD) for forecast RLW (qr; dotted line) and radar-observed qr (solid line). Comparison of the spectra shows reasonable agreement at scales larger than 15 km. For scales smaller than 15 km, the forecast shows a rapid falloff in variability in comparison with the radar. The estimated spectral slope with fit uncertainty is = 3.0 ± 0.1
37
Example log[E(k)] log[wavenumber] Obs_4Eta_12Eta_8 WRF_22WRF_10KF_22 3-6h forecasts from 04 June 2002 1200 UTC
39
June 2002 00z runs 12, 24, 36, 48h fcsts
40
Comparing forecasts that contain different degrees of structure Obs=black Detailed = blue Smooth = green MSE detailed = 1.57 MSE smooth = 1.43
41
Common resolved scales vs. unresolved Filter other forecasts to have same structure MSE “detailed” = 1.32 MSE smooth = 1.43
42
Lack of detail in analyses Methods discussed assume realistic analysis of observations Problems: Relatively sparse observations Operational data assimilation systems Smooth first guess fields from model forecasts Smooth error covariance matrix Smooth analysis fields result
43
True mesoscale analyses Determine what scales are resolved Mesoscale data assimilation Frequent updates All available observations Hi-res NWP provides first guess Ensemble Kalman filter Tustison et al. (2002) scale-recursive filtering takes advantage of natural “scaling”
44
Design verification systems that address forecast value Value measures the benefits of forecast information to users Determine what aspects of forecast users are most sensitive to If possible, find out users “cost/loss” situation Are missed events or false alarms more costly?
45
Issues How to distill the huge amount of verification information into meaningful “nuggets” that can be used effectively? How to elevate verification from an annoyance to an integral part of the forecast process? What happens when conflicting information from different verification approaches is obtained?
46
Summary Problems with traditional verification techniques when used with forecasts/observations with structure Verify realism Issues of scale Work with forecasters/users to determine most important aspects of forecast information
48
References Good books Papers mentioned in this presentation Beth Ebert’s website
49
Scores based on similarity/dissimilarity matrices D = [d i,j ] euclidean distance matrix C = [c i,j ] covariance matrix Scores could be: tr[D] = trace of matrix, for euclidean distance this equates to (f i – o i ) 2 ~ RMSE det[D] = determinant of matrix, a measure of the magnitude of a matrix
51
Fourier power spectra Compare 3h accumulated QPF to radar/gage analyses Forecasts were linearly interpolated to same 4km grid as “Stage IV” analysis Errico (1985) Fourier analysis code used. 2-d Fourier transform converted to 1-d by annular average Fixed grid used for analysis located away from complex terrain of Western U.S. Want to focus on features generated by model physics and dynamics, free from influence of orographically forced circulations
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.