Tim Smyth and Jamie Shutler Assessment of analysis and forecast skill Assessment using satellite data.

Slides:



Advertisements
Similar presentations
ECMWF long range forecast systems
Advertisements

Wavelets Fast Multiresolution Image Querying Jacobs et.al. SIGGRAPH95.
Validation of Satellite Precipitation Estimates for Weather and Hydrological Applications Beth Ebert BMRC, Melbourne, Australia 3 rd IPWG Workshop / 3.
A Global Daily Gauge-based Precipitation Analysis, Part I: Assessing Objective Techniques Mingyue Chen & CPC Precipitation Working Group CPC/NCEP/NOAA.
An Assessment of CMAQ with TEOM Measurements over the Eastern US Michael Ku, Chris Hogrefe, Kevin Civerolo, and Gopal Sistla PM Model Performance Workshop,
Reliability Trends of the Global Forecast System Model Output Statistical Guidance in the Northeastern U.S. A Statistical Analysis with Operational Forecasting.
Assessing cognitive models What is the aim of cognitive modelling? To try and reproduce, using equations or similar, the mechanism that people are using.
Investigating Representation Errors in Inversions of Satellite CO 2 Retrievals K.D. Corbin, A.S. Denning, N.C. Parazoo Department of Atmospheric Science.
Part II – TIME SERIES ANALYSIS C2 Simple Time Series Methods & Moving Averages © Angel A. Juan & Carles Serrat - UPC 2007/2008.
Berkeley Parlab 1. INTRODUCTION A Comparison of Error Metrics for Learning Model Parameters in Bayesian Knowledge Tracing 2. CORRELATIONS TO THE GROUND.
Seamless precipitation prediction skill in a global model: Actual versus potential skill Matthew Wheeler 1, Hongyan Zhu 1, Adam Sobel 2, and Debra Hudson.
1 NOAA’s National Climatic Data Center April 2005 Climate Observation Program Blended SST Analysis Changes and Implications for the Buoy Network 1.Plans.
Slides 13b: Time-Series Models; Measuring Forecast Error
Time-series InSAR with DESDynI: Lessons from ALOS PALSAR Piyush Agram a, Mark Simons a and Howard Zebker b a Seismological Laboratory, California Institute.
Kening Wang, Charles Stegman, Sean W. Mulvenon, and Yanling Xia University of Arkansas, Fayetteville, AR, Using Kriging and Interactive Graphics.
ERCOT 2003 UFE ANALYSIS By William Boswell & Carl Raish AEIC Load Research Conference July 13, 2005.
1 GOES-R AWG Hydrology Algorithm Team: Rainfall Probability June 14, 2011 Presented By: Bob Kuligowski NOAA/NESDIS/STAR.
Peritoneal Dialysis for Elderly Patients: A Review Source: Tesar V. Peritoneal dialysis in the elderly—is its underutilization justified? Nephrol Dial.
© Crown copyright Met Office Climate Projections for West Africa Andrew Hartley, Met Office: PARCC national workshop on climate information and species.
1 The Venzke et al. * Optimal Detection Analysis Jeff Knight * Venzke, S., M. R. Allen, R. T. Sutton and D. P. Rowell, The Atmospheric Response over the.
PM2.5 Model Performance Evaluation- Purpose and Goals PM Model Evaluation Workshop February 10, 2004 Chapel Hill, NC Brian Timin EPA/OAQPS.
Gridding Daily Climate Variables for use in ENSEMBLES Malcolm Haylock, Climatic Research Unit Nynke Hofstra, Mark New, Phil Jones.
Spatio-Temporal Surface Vector Wind Retrieval Error Models Ralph F. Milliff NWRA/CoRA Lucrezia Ricciardulli Remote Sensing Systems Deborah K. Smith Remote.
Combining CMORPH with Gauge Analysis over
Center for Radiative Shock Hydrodynamics Fall 2011 Review Assessment of predictive capability Derek Bingham 1.
Validation Dr Andy Evans. Preparing to model Verification Calibration/Optimisation Validation Sensitivity testing and dealing with error.
DEVELOPING HIGH RESOLUTION AOD IMAGING COMPATIBLE WITH WEATHER FORECAST MODEL OUTPUTS FOR PM2.5 ESTIMATION Daniel Vidal, Lina Cordero, Dr. Barry Gross.
Objective Data  The outlined square marks the area of the study arranged in most cases in a coarse 24X24 grid.  Data from the NASA Langley Research Center.
Error & Uncertainty: II CE / ENVE 424/524. Handling Error Methods for measuring and visualizing error and uncertainty vary for nominal/ordinal and interval/ratio.
Regional Climate Model Evaluation System based on satellite and other observations for application to CMIP/AR downscaling Peter Lean 1, Jinwon Kim 1,3,
VIP: Finding Important People in Images Clint Solomon Mathialagan Andrew C. Gallagher Dhruv Batra CVPR
Evaluation of the VISTAS 2002 CMAQ/CAMx Annual Simulations T. W. Tesche & Dennis McNally -- Alpine Geophysics, LLC Ralph Morris -- ENVIRON Gail Tonnesen.
Verification of Precipitation Areas Beth Ebert Bureau of Meteorology Research Centre Melbourne, Australia
1. Analysis and Reanalysis Products Adrian M Tompkins, ICTP picture from Nasa.
Presented at the 7th Annual CMAS Conference, Chapel Hill, NC, October 6-8, 2008 Identifying Optimal Temporal Scale for the Correlation of AOD and Ground.
Biostatistics Case Studies 2006 Peter D. Christenson Biostatistician Session 2: Correlation of Time Courses of Simultaneous.
Improving MISR-retrieved Aerosol Properties Using GOCART Simulations Yang Liu, PhD June 3, 2015 St. Louis, MO.
Exposure Assessment for Health Effect Studies: Insights from Air Pollution Epidemiology Lianne Sheppard University of Washington Special thanks to Sun-Young.
Introduction GOES-R ABI will be the first GOES imaging instrument providing observations in both the visible and the near infrared spectral bands. Therefore.
Model Fusion and its Use in Earth Sciences R. Romero, O. Ochoa, A. A. Velasco, and V. Kreinovich Joint Annual Meeting NSF Division of Human Resource Development.
Robert W. Pinder, Alice B. Gilliland, Robert C. Gilliam, K. Wyat Appel Atmospheric Modeling Division, NOAA Air Resources Laboratory, in partnership with.
WCRP Extremes Workshop Sept 2010 Detecting human influence on extreme daily temperature at regional scales Photo: F. Zwiers (Long-tailed Jaeger)
A Temporal Filtering Algorithm to Reconstruct Daily Albedo Series Based on GLASS Albedo product Nanfeng Liu 1,2, Qiang Liu 1,2, Lizhao Wang 2, Jianguang.
Page 1© Crown copyright 2005 Met Office Verification -status Clive Wilson, Presented by Mike Bush at EWGLAM Meeting October 8- 11, 2007.
Page 1© Crown copyright 2004 The use of an intensity-scale technique for assessing operational mesoscale precipitation forecasts Marion Mittermaier and.
Critical Assessment of TOMS-derived Tropospheric Ozone: Comparisons with Other Measurements and Model Evaluation of Controlling Processes M. Newchurch.
Validation of Satellite-derived Clear-sky Atmospheric Temperature Inversions in the Arctic Yinghui Liu 1, Jeffrey R. Key 2, Axel Schweiger 3, Jennifer.
1 Spatio-temporal Distribution of Latent Heating in the Southeast Asian Monsoon Region School of Earth and Atmospheric Sciences Georgia Institute of Technology.
MGS3100_03.ppt/Feb 11, 2016/Page 1 Georgia State University - Confidential MGS 3100 Business Analysis Time Series Forecasting Feb 11, 2016.
Learning to Rank: From Pairwise Approach to Listwise Approach Authors: Zhe Cao, Tao Qin, Tie-Yan Liu, Ming-Feng Tsai, and Hang Li Presenter: Davidson Date:
MODIS Atmosphere Products: The Importance of Record Quality and Length in Quantifying Trends and Correlations S. Platnick 1, N. Amarasinghe 1,2, P. Hubanks.
Predictability of Monthly Mean Temperature and Precipitation: Role of Initial Conditions Mingyue Chen, Wanqiu Wang, and Arun Kumar Climate Prediction Center/NCEP/NOAA.
Exposure Prediction and Measurement Error in Air Pollution and Health Studies Lianne Sheppard Adam A. Szpiro, Sun-Young Kim University of Washington CMAS.
CARPE DIEM 4 th meeting Critical Assessment of available Radar Precipitation Estimation techniques and Development of Innovative approaches for Environmental.
Daiwen Kang 1, Rohit Mathur 2, S. Trivikrama Rao 2 1 Science and Technology Corporation 2 Atmospheric Sciences Modeling Division ARL/NOAA NERL/U.S. EPA.
Tushar Sinha Assistant Professor
Systematic timing errors in km-scale NWP precipitation forecasts
Statistical Methods for Model Evaluation – Moving Beyond the Comparison of Matched Observations and Output for Model Grid Cells Kristen M. Foley1, Jenise.
4th Joint EU-OECD Workshop on BCS, Brussels, October 12-13
Aquarius SSS space/time biases with respect to Argo data
Eun-Pa Lim and Harry H. Hendon Science to Services
A Review in Quality Measures for Halftoned Images
University of Washington Center for Science in the Earth System
Quantitative verification of cloud fraction forecasts
Fig. 1 The temporal correlation coefficient (TCC) skill for one-month lead DJF prediction of 2m air temperature obtained from 13 coupled models and.
Validation for TPW (PGE06)
Modeling spatially-dependent, non-stationary bias in GEOS-CHEM
GloSea4: the Met Office Seasonal Forecasting System
Seasonal Forecasting Using the Climate Predictability Tool
REGIONAL AND LOCAL-SCALE EVALUATION OF 2002 MM5 METEOROLOGICAL FIELDS FOR VARIOUS AIR QUALITY MODELING APPLICATIONS Pat Dolwick*, U.S. EPA, RTP, NC, USA.
Presentation transcript:

Tim Smyth and Jamie Shutler Assessment of analysis and forecast skill Assessment using satellite data

Rationale – what are we trying to achieve ? Original approaches and web portal. New metrics and approaches –Annual SST comparisons –Annual chl-a comparisons –Weekly chl-a comparisons ( 2 examples) Conclusions. Overview

Rationale Sat in too many conferences, squinting from the back, where models and satellite data were said to be “in good agreement”: –Highly dissatisfactory as is subjective; –Authors frequently play with scales to suit; –Not necessarily comparing like with like –SOLUTION: come up with comparison web site …

Advantages Immediately shows up where discrepancy between satellite and model; Gives potential for models to be improved over shorter time-cycles (need to discuss this at the workshop); The shame factor! Disadvantages Not necessarily comparing like with like; Only surface measurements Problems with cloud coverage and averaging; “What is truth?”

Some immediate results On short time-scales: poor agreement; –Spring bloom timing is crucial: time of dramatic changes. On seasonal time-scales: better agreement; (see Allen et al. (2008)); –Not surprising as our models are based on trying to predict things like seasonal succession. There is some model skill in predicting the likelihood of blooms. However still need to produce more useful metrics to judge different aspects of the model …

New metrics Towards determining reasons for differences: Principal component analysis (or EOF) of time series. Kappa coefficient –Method traditionally employed to look at land change. –Measure of ‘difference’ between two sets of data. Receiver operator curves (ROC). –Used extensively in pattern recognition. –Employed in precipitation analyses (Met Office). –Performance graphing technique. Wavelet decomposition method. –Employed in precipitation analyses (Met Office). –Spatial decomposition. –Provides a measure of skill and mean squared error.

New metrics General approach: Metrics are automatically generated every week. Comparing weekly composite data. All data mapped to same scale and domain. Common cloud masking (of composite). Satellite data spatially averaged to approximate spatial scale of model data. Two sets of results: Annual analyses (February to October) Weekly analyses

Annual : SST metric results Visual observations: spatially and temporally similar Satellite PC1 Satellite PC2 model PC1 model PC2 Satellite temporal weighting Model temporal weighting

Annual : SST metric results Visual observations: spatially and temporally similar Satellite PC1 Satellite PC2 model PC1 model PC2 Satellite temporal weighting Model temporal weighting

Annual : chlorophyll metric results Visual observations: spatially and temporally different Satellite PC1 Satellite PC2 model PC1 model PC2 Satellite temporal weighting Model temporal weighting

Weekly : 5 July chlorophyll metric results Satellite MRCS model %age difference 5 th July 2008 Visual observations: model data appear to contain a large bias some structures appear similar (changes in gradient) Same geophysical scale

Weekly : 5 July chlorophyll metric results Satellite MRCS model %age difference 5 th July 2008 Visual observations: model data appear to contain a large bias some structures appear similar (changes in gradient). Same geophysical scale Similar gradients (with bias)

Weekly : 5 July chlorophyll metric results Satellite MRCS model %age difference 5 th July 2008 Visual observations: model data appear to contain a large bias. some structures appear similar (changes in gradient). Same geophysical scale Similar gradients (with bias)

Weekly : 5 July chlorophyll metric results Kappa MSE Skill score ROC

Weekly : 5 July chlorophyll metric results Kappa ROC MSE Skill score Data diverge At ~0.5mg m -3 Errors at all scales across range of chlorophyll -ve skill score for high chlorophyll, particularly at lower spatial scales Better than random performance

Weekly : 11 October chlorophyll metric results Satellite MRCS model %age difference 11 th October 2008 Visual observations: mixture of similar and dissimilar structures model data inverse of satellite data in some regions.

Weekly : 11 October chlorophyll metric results Satellite MRCS model %age difference 11 th October 2008 Visual observations: mixture of similar and dissimilar structures model data inverse of satellite data in some regions. opposite responses

Weekly : 11 October chlorophyll metric results Satellite MRCS model %age difference 11 th October 2008 Visual observations: mixture of similar and dissimilar structures model data inverse of satellite data in some regions. Feature in model data not apparent in satellite

Weekly : 11 October chlorophyll metric results Kappa ROC MSE Skill score

Weekly : 11 October chlorophyll metric results Data diverge At ~0.1mg m -3 Performance closer to random Errors at all scales Across range of chlorophyll -ve skill score for high chlorophyll, at low spatial scales ROC Kappa MSE Skill score

Conclusions Presented operational framework for evaluating MRCS vs satellite data Apparent need to use multiple metric approach. Seasonal signal captured well by model SST. Chlorophyll comparisons show the biggest differences. Performance of model chlorophyll varies (when compared with satellite data). However, satellite chlorophyll signal in winter months is likely to be incorrect or biased (algorithm performance reduces during this time). Both datasets have uncertainties. Still not comparing like with like (e.g. temporal differences in input data, possible that binning the satellite data to match model data will improve results).