Shaocai Yu *, Brian Eder* ++, Robin Dennis* ++, Shao-hang Chu**, Stephen Schwartz*** *Atmospheric Sciences Modeling Division, National Exposure Research.

Slides:



Advertisements
Similar presentations
Office of Research and Development National Exposure Research Laboratory, Atmospheric Modeling and Analysis Division Changes in U.S. Regional-Scale Air.
Advertisements

A PERFORMANCE EVALUATION OF THE ETA - CMAQ AIR QUALITY FORECAST MODEL FOR THE SUMMER OF 2004 CMAS Workshop Chapel Hill, NC 20 October, 2004.
Regional Haze Modeling: Recent Modeling Results for VISTAS and WRAP October 27, 2003, CMAS Annual Meeting, RTP, NC University of California, Riverside.
N emissions and the changing landscape of air quality Rob Pinder US EPA Office of Research and Development Atmospheric Modeling & Analysis Division.
Photochemical Model Performance for PM2.5 Sulfate, Nitrate, Ammonium, and pre-cursor species SO2, HNO3, and NH3 at Background Monitor Locations in the.
Perspectives in Designing and Operating a Regional Ammonia Monitoring Network Gary Lear USEPA Clean Air Markets Division.
An Assessment of CMAQ with TEOM Measurements over the Eastern US Michael Ku, Chris Hogrefe, Kevin Civerolo, and Gopal Sistla PM Model Performance Workshop,
Three-State Air Quality Study (3SAQS) Three-State Data Warehouse (3SDW) 2008 CAMx Modeling Model Performance Evaluation Summary University of North Carolina.
Havala O. T. Pye 1, Rob Pinder 1, Ying Xie 1, Deborah Luecken 1, Bill Hutzell 1, Golam Sarwar 1, Jason Surratt 2 1 US Environmental Protection Agency 2.
Does ozone model performance vary as a function of synoptic meteorological type? Pat Dolwick, Christian Hogrefe, Mark Evangelista, Chris Misenis, Sharon.
Tanya L. Otte and Robert C. Gilliam NOAA Air Resources Laboratory, Research Triangle Park, NC (In partnership with U.S. EPA National Exposure Research.
StateDivision Mean Winter Temperature CT 1 - Northwest26.9 +/ Central29.5 +/ Coastal31.9 +/ MA 1 - Western24.9.
The concept of SPR is introduced in (Beatty, Petersen & Swindale, 1979) p The National Soils Handbook § (U.S. Department of Agriculture,
Air Resources Laboratory Yunsoo Choi 12, Daewon Byun 1, Pius Lee 1, Rick Saylor 1, Ariel Stein 12, Daniel Tong 12, Hyun-Cheol Kim 12, Fantine Ngan 13,
Office of Research and Development National Exposure Research Laboratory, Atmospheric Modeling and Analysis Division S.T. Rao Director, Atmospheric Modeling.
PM Model Performance Goals and Criteria James W. Boylan Georgia Department of Natural Resources - VISTAS National RPO Modeling Meeting Denver, CO May 26,
Research Design. Research is based on Scientific Method Propose a hypothesis that is testable Objective observations are collected Results are analyzed.
Calculating Statistics: Concentration Related Performance Goals James W. Boylan Georgia Department of Natural Resources PM Model Performance Workshop Chapel.
Five-year Progress in the Performance of Air Quality Forecast Models: Analysis on Categorical Statistics for the National Air Quality Forecast Capacity.
Lecture 9. The organizational structure of management of enterprise.
MODELS3 – IMPROVE – PM/FRM: Comparison of Time-Averaged Concentrations R. B. Husar S. R. Falke 1 and B. S. Schichtel 2 Center for Air Pollution Impact.
Modeling Studies of Air Quality in the Four Corners Region National Park Service U.S. Department of the Interior Cooperative Institute for Research in.
Comparison of three photochemical mechanisms (CB4, CB05, SAPRC99) for the Eta-CMAQ air quality forecast model for O 3 during the 2004 ICARTT study Shaocai.
O. Russell Bullock, Jr. National Oceanic and Atmospheric Administration (NOAA) Atmospheric Sciences Modeling Division (in partnership with the U.S. Environmental.
PM2.5 Model Performance Evaluation- Purpose and Goals PM Model Evaluation Workshop February 10, 2004 Chapel Hill, NC Brian Timin EPA/OAQPS.
Chapter 1 Introduction to Statistics. Statistical Methods Were developed to serve a purpose Were developed to serve a purpose The purpose for each statistical.
Fine scale air quality modeling using dispersion and CMAQ modeling approaches: An example application in Wilmington, DE Jason Ching NOAA/ARL/ASMD RTP,
On the Model’s Ability to Capture Key Measures Relevant to Air Quality Policies through Analysis of Multi-Year O 3 Observations and CMAQ Simulations Daiwen.
A comparison of PM 2.5 simulations over the Eastern United States using CB-IV and RADM2 chemical mechanisms Michael Ku, Kevin Civerolo, and Gopal Sistla.
Using CMAQ-AIM to Evaluate the Gas-Particle Partitioning Treatment in CMAQ Chris Nolte Atmospheric Modeling Division National Exposure Research Laboratory.
Prakash Bhave, Shawn Roselle, Frank Binkowski, Chris Nolte, Shaocai Yu, Jerry Gipson, & Ken Schere CMAS Conference – Paper #6.8 Chapel Hill, NC, October.
A detailed evaluation of the WRF-CMAQ forecast model performance for O 3, and PM 2.5 during the 2006 TexAQS/GoMACCS study Shaocai Yu $, Rohit Mathur +,
Model Evaluation Comparing Model Output to Ambient Data Christian Seigneur AER San Ramon, California.
Application of the CMAQ-UCD Aerosol Model to a Coastal Urban Site Chris Nolte NOAA Atmospheric Sciences Modeling Division Research Triangle Park, NC 6.
Prakash V. Bhave, Ph.D. Physical Scientist PM Model Performance Workshop February 10, 2004 Postprocessing Model Output for Comparison to Ambient Data.
Time-Resolved & In-Depth Evaluation of PM and PM Precursors using CMAQ Robin L. Dennis Atmospheric Modeling Division U.S. EPA/ORD:NOAA/ARL PM Model Performance.
Copyright © Cengage Learning. All rights reserved. 12 Analysis of Variance.
Use of space-based tropospheric NO 2 observations in regional air quality modeling Robert W. Pinder 1, Sergey L. Napelenok 1, Alice B. Gilliland 1, Randall.
William G. Benjey* Physical Scientist NOAA Air Resources Laboratory Atmospheric Sciences Modeling Division Research Triangle Park, NC Fifth Annual CMAS.
The Impact of Short-term Climate Variations on Predicted Surface Ozone Concentrations in the Eastern US 2020 and beyond Shao-Hang Chu and W.M. Cox US Environmental.
Evaluation of Models-3 CMAQ I. Results from the 2003 Release II. Plans for the 2004 Release Model Evaluation Team Members Prakash Bhave, Robin Dennis,
Diagnostic Study on Fine Particulate Matter Predictions of CMAQ in the Southeastern U.S. Ping Liu and Yang Zhang North Carolina State University, Raleigh,
Seasonal Modeling of the Export of Pollutants from North America using the Multiscale Air Quality Simulation Platform (MAQSIP) Adel Hanna, 1 Rohit Mathur,
Office of Research and Development National Exposure Research Laboratory, Atmospheric Modeling and Analysis Division 16 October 2012 Integrating source.
Evaluating temporal and spatial O 3 and PM 2.5 patterns simulated during an annual CMAQ application over the continental U.S. Evaluating temporal and spatial.
Extending Size-Dependent Composition to the Modal Approach: A Case Study with Sea Salt Aerosol Uma Shankar and Rohit Mathur The University of North Carolina.
Robert W. Pinder, Alice B. Gilliland, Robert C. Gilliam, K. Wyat Appel Atmospheric Modeling Division, NOAA Air Resources Laboratory, in partnership with.
Office of Research and Development National Exposure Research Laboratory, Atmospheric Modeling and Analysis Division October 21, 2009 Evaluation of CMAQ.
Evaluation of CMAQ Driven by Downscaled Historical Meteorological Fields Karl Seltzer 1, Chris Nolte 2, Tanya Spero 2, Wyat Appel 2, Jia Xing 2 14th Annual.
Sampling Design and Analysis MTH 494 Lecture-21 Ossam Chohan Assistant Professor CIIT Abbottabad.
Measurements and Their Analysis. Introduction Note that in this chapter, we are talking about multiple measurements of the same quantity Numerical analysis.
PM Methods Update and Network Design Presentation for WESTAR San Diego, CA September 2005 Peter Tsirigotis Director Emissions, Monitoring, and Analysis.
W. T. Hutzell 1, G. Pouliot 2, and D. J. Luecken 1 1 Atmospheric Modeling Division, U. S. Environmental Protection Agency 2 Atmospheric Sciences Modeling.
A Comparison Study of CMAQ Aerosol Prediction by Two Thermodynamic Modules: UHAERO V.S. ISORROPIA Case study for January 2002 episode Fang-Yi Cheng 1,
U.S. Environmental Protection Agency Office of Research and Development Implementation of an Online Photolysis Module in CMAQ 4.7 Christopher G. Nolte.
A study of process contributions to PM 2.5 formation during the 2004 ICARTT period using the Eta-CMAQ forecast model over the eastern U.S. Shaocai Yu $,
Office of Research and Development National Risk Management Research Laboratory, Air Pollution Prevention and Control Division Photo image area measures.
Robin L. Dennis, Jesse O. Bash, Kristen M. Foley, Rob Gilliam, Robert W. Pinder U.S. Environmental Protection Agency, National Exposure Research Laboratory,
Daiwen Kang 1, Rohit Mathur 2, S. Trivikrama Rao 2 1 Science and Technology Corporation 2 Atmospheric Sciences Modeling Division ARL/NOAA NERL/U.S. EPA.
Preliminary Evaluation of the June 2002 Version of CMAQ Brian Eder Shaocai Yu Robin Dennis Jonathan Pleim Ken Schere Atmospheric Modeling Division* National.
Simulation of PM2.5 Trace Elements in Detroit using CMAQ
Statistical Methods for Model Evaluation – Moving Beyond the Comparison of Matched Observations and Output for Model Grid Cells Kristen M. Foley1, Jenise.
Simulation of primary and secondary (biogenic and anthropogenic) organic aerosols over the United States by US EPA Models-3/CMAQ: Evaluation and regional.
Predicting Future-Year Ozone Concentrations: Integrated Observational-Modeling Approach for Probabilistic Evaluation of the Efficacy of Emission Control.
Shaocai Yu*, Brian Eder*++, Robin Dennis*++,
J. Burke1, K. Wesson2, W. Appel1, A. Vette1, R. Williams1
WRAP Modeling Forum, San Diego
REGIONAL AND LOCAL-SCALE EVALUATION OF 2002 MM5 METEOROLOGICAL FIELDS FOR VARIOUS AIR QUALITY MODELING APPLICATIONS Pat Dolwick*, U.S. EPA, RTP, NC, USA.
Evaluation of Models-3 CMAQ Annual Simulation Brian Eder, Shaocai Yu, Robin Dennis, Alice Gilliland, Steve Howard,
Presentation transcript:

Shaocai Yu *, Brian Eder* ++, Robin Dennis* ++, Shao-hang Chu**, Stephen Schwartz*** *Atmospheric Sciences Modeling Division, National Exposure Research Laboratory, ** Office of Air Quality Planning and Standards, U.S. EPA, NC *** Atmospheric Sciences Division, Brookhaven National Laboratory, Upton, NY Air Resources Laboratory, National Oceanic and Atmospheric Administration, RTP, NC In this study, we propose new metrics to solve the symmetrical problem between overprediction and underprediction following the concept of factor. Theoretically, factor is defined as ratio of model prediction to observation if the model prediction is higher than the observation, whereas it is defined as ratio of observation to model prediction if the observation is higher than the model prediction. Following this concept, the mean normalized factor bias (M NFB ), mean normalized gross factor error (M NGFE ), normalized mean bias factor (N MBF ) and normalized mean error factor (N MEF ) are proposed and defined as follows: where M i and O i are values of model (prediction) and observation at time and/or location i, respectively, N is number of samples (by time and/or location). The values of M NFB, and N MBF are linear and not bounded (range from -  to +  ). Like M NB and M NGE, M NFB and M NGFE can have another general problem when some observation values (denomination) are trivially low and they can significantly influence the values of those metrics. N MBF and N MEF can avoid this problem because the sum of the observations is used to normalize the bias and error,. The above formulas of N MBF and N MEF can be rewritten for case as follows: N MBF and N MEF are actually the results of summaries of normalized bias (M NB ) and error (M NGE ) with the observational concentrations as a weighting function, respectively. N MBF and N MEF have both advantages of avoiding dominance by the low values of observations in normalization like N MB and N ME and maintaining adequate evaluation symmetry. The meanings of N MBF can be interpreted as follows: if N MBF  0, for example, N MBF =1.2, this means that the model overpredicts the observation by a factor of 2.2 (i.e., N MBF +1=1.2+1=2.2); if N MBF  0, for example, N MBF =-0.2, this means that the model underpredicts the observation by a factor of 1.2 (i.e., N MBF -1=-0.2-1=-1.2). Table 3 and Figures 3 and 4: (1) For weekly data from CASTNet, both N MBF (0.03 to 0.08) and N MEF (0.24 to 0.27) for SO 4 2- are lower than the performance criteria. (2) For 24-hour data from IMPROVE, SEARCH and STN, both N MBF (-0.19 to 0.22) and N MEF (0.42 to 0.46) for SO 4 2- are slightly higher than the performance criteria. The model performed better on the weekly data of CASTNet and better over the eastern region than western region for SO This is consistent with the results of Dennis et al. [1993]. (3) For PM 2.5 NO 3 -, both N MBF (-0.96 to 0.59) and N MEF (0.80 to 1.70) for SEARCH, CASTNet, and IMPROVE data are larger than the performance criteria. Although the N MBF (0.01) for STN data in 2002 is very small, its N MEF (0.77) is still larger than the performance criteria. (4) CMAQ performed well over the Northeastern region but overpredicted most of observed NO 3 - over the mid-Atlantic region by more than a factor of 2 and underpredicted most of observed NO 3 - over the west region and southeast by more than a factor of 2. Both N MBF and N MEF of winter of 2002 are smaller than those of summer of This is because the NO 3 - concentrations in winter of 2002 were 7 times higher than those in summer of 1999 although M B values of winter of 2002 were higher than those of summer of (5) One of major reasons for poor performance of model on PM 2.5 NO 3 - is that the PM 2.5 NO 3 - concentrations are very low and are very sensitive to the errors in PM 2.5 SO 4 2- and NH 4 +, and temperature and RH when thermodynamic model (such as ISSORROPIA model here) partitions total nitrate (i.e., aerosol NO 3 - +gas HNO 3 ) between the gas and aerosol phases. 4. Summary and conclusions A set of new statistical quantitative metrics on the basis of concept of factor has been proposed that can provide a rational operational evaluation of air quality model for the relative difference between modeled and observed results. The new metrics (i.e., N MBF, N MEF ) have advantages of both avoiding dominance by the low values of observations and maintaining adequate evaluation symmetry. The application of these new quantitative metrics in operational evaluation of CMAQ performance on PM 2.5 SO 4 2- and NO 3 - shows that the new quantitative metrics are useful and their meanings are also very clear and easy to explain. Figure 1: a dataset for a real case of model and observation for aerosol NO 3 - was separated into four regions, (1) region 1 for model/observation  0.5, (2) region 2 for 0.5  model/observation  1.0, (3) region 3 for 1.0<model/observation  2.0, (4) region 4 for 2.0  model/observation. Results in Table 2: (1) for the only data in region 1 with model/observation  0.5, M NB, N MB, F B, N MFB, M NFB and N MBF are – 0.82, -0.78, -1.43, -1.28, , and – 3.58, respectively. Obviously, only normalized mean bias factor (N MBF ) gives reasonable description of model performance, i.e., the model underpredicted the observations by a factor of 4.58 in this case. (2) For the only data in region 4 with model/observation>2, M NB, N MB, F B, N MFB, M NFB and N MBF are 4.27, 2.25, 1.12, 1.06, 4.27 and 2.25, respectively. The results of N MBF and N MB reasonably indicate that the model overpredicted the observations by a factor of (3) For the results of each metrics on combination case of regions 1 and 4 data, M NB, N MB, F B, N MFB, M NFB and N MBF are 1.50, 0.06, -0.27, 0.06, and 0.06, respectively. Both N MB and N MBF show that the model slightly overpredicted the observations by a factor of 1.06, while F B (-0.27) shows that the model underpredicted the observations. This shows that the value of F B can sometimes result in a misleading conclusion as well. This specific case shows that it is not wise to use F B as an evaluation metric. N ME and N MEF : gross error between observations and model results is 1.19 times of mean observation. The good model performance can be concluded only under the condition that both relative bias (N MBF ) and relative gross error (N MEF ) meet the certain performance standards. (4) For the all data in Figure 1 (combination case in Table 2), M NB, N MB, F B, N MFB, M NFB and N MBF are 0.96, 0.09, -0.13, 0.09, and 0.09, respectively. Both N MB and N MBF show that the mean model only overpredicted the mean observation by a factor of However, the gross error (N MGE ) between the model and observation is 0.77 times as high as observation. See Figure 1. On the basis of the above analyses and test, it can be concluded that our proposed new statistical metrics (i.e., N MBF and N MEF ) on the basis of concept of factor can show the model performance reasonably. These new metrics use observational data as only reference for the model evaluation and their meanings are also very clear and easy to explain. 3. Applications of new metrics over the US CMAQ model on PM 2.5 SO 4 2- and NO 3 - over the US: 6/15 to 7/17, 1999 and 1/8 to 2/18, PM 2.5 SO 4 2- and NO 3 - observational data: (1) at 61 rural sites from IMPROVE. Two 24-hour samples are collected on quartz filters each week, on Wednesday and Saturday, beginning at midnight local time. (2) at 73 rural sites from CASNet. Weekly (Tuesday to Tuesday) samples are collected on Teflon filters. (3) at 8 sites from SEARCH (Southeastern Aerosol Research and Characterization project). Daily samples are collected on quartz filters. (4) at 153 urban sites from STN (Speciated Trends Network). 24-hour samples are usually taken once every six days. The recommended performance criteria for O 3 by US EPA (1991): normalized bias (M NB )  5 to  15%; normalized gross error (M NGE ) 30% to 35%. +15% of M NB corresponds to +15% of N MBF while – 15% of M NB corresponds to – 18% of N MBF. New unbiased symmetric metrics for evaluation of the air quality model 1. INTRODUCTION Model performance evaluation is a problem of longstanding concern, and has two important objectives, i.e., (1) determine a model’s degree of acceptability and usefulness for a specific task, and (2) establish that one is getting good results for the right reason (Russell and Dennis, 2000). These objectives are accomplished by two different types of model evaluation, i.e., operational evaluation and diagnostic evaluation (or model physics evaluation) (Fox, 1981; EPA, 1991; Weil et al., 1992; Russell and Dennis, 2000), respectively. Although the operational evaluations for different air quality models have been intensively performed for regulatory purposes in the past years, the resulting array of statistical metrics are so diverse and numerous that it is difficult to judge the overall performance of the models (EPA, 1991; Seigneur et al., 2000; Yu et al., 2003). Some statistical metrics can cause misleading conclusions about the model performance (Seigneur et al., 2000). In this paper, a new set of quantitative metrics for the operational evaluation is proposed and applied in real evaluation cases. The other quantitative metrics frequently used in the operational evaluation are tested and their shortcomings are examined. 2. Quantitative metrics related to the operational evaluation and their examinations Table 1: traditional quantitative metrics,mathematical expressions used in the operational evaluation. Differences between the model and observations: mean bias (M B ) and mean absolute gross error (M AGE ) (or root mean square error). Relative differences between the model and observations: The traditional metrics (such as M NB, M NGE, N MB,and N ME ) two problems that may mislead conclusions with this approach: (1) the values of M NB and N MB can grow disproportionately for overpredictions and underpredicitons because both values of M NB and N MB are bounded by –100% for underprediction; (2) the values of M NB and M NGE can be significantly influenced by some points with trivially low values of observations (denomination). To solve this asymmetric evaluation problem between overpredictions and underpredictions,  fractional bias (F B ) and fractional gross error (F GE ) (Irwin and Smith, 1984; Kasibhatla et al., 1997; Seigneur et al., 2000).  Problems for F B and F GE: (1)what the metrics F B and F GE measure is not clear because the model prediction is not evaluated against observation but average of observation and model prediction, which is contradict to the traditional definition of evaluation. (2)the scales of F B and F GE are not linear and are seriously compressed beyond  1 as F B and F GE are bounded by  2 and +2, respectively. Acknowledgements The authors wish to thank other members at ASMD of EPA for their contributions to the 2002 release version of EPA Models-3/CMAQ during the development and evaluation. This work has been subjected to US Environmental Protection Agency peer review and approved for publication. Mention of trade names or commercial products does not constitute endorsement or recommendation for use. REFERENCES Cox, W.M., Tikvart, J.A., Atmospheric Environment 24, Eder, B., Shaocai, Yu, etc., Atmospheric Environment (in preparation). EPA, Guideline for regulatory application of the urban airshed model. USEPA Report No. EPA-450/ U.S. EPA, Office of Air Quality Planning and Standards, Research Triangle Park, North Carolina. Dennis, R.L., J.N. McHenry, W.R.Barchet, F.S. Binkowski, and D.W. Byun, Atmos. Environ., 26 A(6), , Fox, D.G., Bulletin American Meteorological Society 62, Irwin, J., and M. Smith, 1984, Bulletin American Meteorological Society 65, Russell, A., and R. Dennis, Atmos. Environ., 34, , Seigneur, C., et al.,., Journal of the Air & Waste Management Association 50, Weil, J.C., R.I. Sykes, and A. Venkatram, 1992, Journal of Applied Meteorology, 31, Yu, S.C., Kasibhatla, P.S., Wright, D.L., Schwartz, S.E., McGraw, R., Deng, A., 2003.Journal of Geophysical Research 108(D12), 4353, doi: /2002JD