Experiences Developing a Semantic Representation of Product Quality, Bias, and Uncertainty for a Satellite Data Product Patrick West 1, Gregory Leptoukh.

Slides:



Advertisements
Similar presentations
Characterization of radiance uncertainties for SeaWiFS and Modis-Aqua Introduction The spectral remote sensing reflectance is arguably the most important.
Advertisements

Experiences Developing a User-centric Presentation of Provenance for a Web- based Science Data Analysis Tool Stephan Zednik 1, Gregory Leptoukh 2, Peter.
GEOS-5 Simulations of Aerosol Index and Aerosol Absorption Optical Depth with Comparison to OMI retrievals. V. Buchard, A. da Silva, P. Colarco, R. Spurr.
A Tutorial on MODIS and VIIRS Aerosol Products from Direct Broadcast Data on IDEA Hai Zhang 1, Shobha Kondragunta 2, Hongqing Liu 1 1.IMSG at NOAA 2.NOAA.
Gregory Leptoukh, David Lary, Suhung Shen, Christopher Lynnes What’s in a day?
Data Quality Screening Service Christopher Lynnes, Bruce Vollmer, Richard Strub, Thomas Hearty Goddard Earth Sciences Data and Information Sciences Center.
Retrieval of smoke aerosol loading from remote sensing data Sean Raffuse and Rudolf Husar Center for Air Pollution Impact and Trends Analysis Washington.
Constraining aerosol sources using MODIS backscattered radiances Easan Drury - G2
Quantifying aerosol direct radiative effect with MISR observations Yang Chen, Qinbin Li, Ralph Kahn Jet Propulsion Laboratory California Institute of Technology,
Data Quality Screening Service Christopher Lynnes, Richard Strub, Thomas Hearty, Bruce Vollmer Goddard Earth Sciences Data and Information Sciences Center.
European Geosciences Union General Assembly 2006 Vienna, Austria, 02 – 07 April 2006 Paper’s objectives: 1. Contribute to the validation of MODIS aerosol.
Informatics takes on data and information quality, uncertainty and bias (in atmospheric science) Peter Fox (TWC/RPI), and … Stephan Zednik 1, Gregory Leptoukh.
ESTEC July 2000 Estimation of Aerosol Properties from CHRIS-PROBA Data Jeff Settle Environmental Systems Science Centre University of Reading.
MODIS Regional and Global Cloud Variability Brent C. Maddux 1,2 Steve Platnick 3, Steven A. Ackerman 1,2, Paul Menzel 1, Kathy Strabala 1, Richard Frey.
Satellite Remote Sensing of Surface Air Quality
Experiences Developing a User- centric Presentation of A Domain- enhanced Provenance Data Model Cynthia Chang 1, Stephan Zednik 1, Chris Lynnes 2, Peter.
1 Satellite Remote Sensing of Particulate Matter Air Quality ARSET Applied Remote Sensing Education and Training A project of NASA Applied Sciences Pawan.
Determining the accuracy of MODIS Sea- Surface Temperatures – an Essential Climate Variable Peter J. Minnett & Robert H. Evans Meteorology and Physical.
Dust Detection in MODIS Image Spectral Thresholds based on Zhao et al., 2010 Pawan Gupta NASA Goddard Space Flight Center GEST/University of Maryland Baltimore.
2010 CEOS Field Reflectance Intercomparisons Lessons Learned K. Thome 1, N. Fox 2 1 NASA/GSFC, 2 National Physical Laboratory.
Aircraft spiral on July 20, 2011 at 14 UTC Validation of GOES-R ABI Surface PM2.5 Concentrations using AIRNOW and Aircraft Data Shobha Kondragunta (NOAA),
AeroStat: Online Platform for the Statistical Intercomparison of Aerosols Gregory Leptoukh, NASA/GSFC (P.I.) Christopher Lynnes, NASA/GSFC (Co-I.) Robert.
Spatially Complete Global Surface Albedos Derived from MODIS Data
VALIDATION OF SUOMI NPP/VIIRS OPERATIONAL AEROSOL PRODUCTS THROUGH MULTI-SENSOR INTERCOMPARISONS Huang, J. I. Laszlo, S. Kondragunta,
Addressing and Presenting Quality of Satellite Data Gregory Leptoukh ESIP Information Quality Cluster.
Developing a High Spatial Resolution Aerosol Optical Depth Product Using MODIS Data to Evaluate Aerosol During Large Wildfire Events STI-5701 Jennifer.
Giovanni for AQ Gregory Leptoukh NASA Goddard Space Flight Center Goddard Earth Sciences Data and Information Services Center (GES DISC)
Occurrence of TOMS V7 Level-2 Ozone Anomalies over Cloudy Areas Xiong Liu, 1 Mike Newchurch, 1,2 and Jae Kim 1,3 1. Department of Atmospheric Science,
Orbit Characteristics and View Angle Effects on the Global Cloud Field
Surface Reflectivity from OMI: Effects of snow on OMI NO 2 retrievals Gray O’Byrne 1, Randall Martin 1,2, Joanna Joiner 3, Edward A. Celarier 3 1 Dalhousie.
1 Satellite Remote Sensing of Particulate Matter Air Quality ARSET Applied Remote Sensing Education and Training A project of NASA Applied Sciences Pawan.
EUMETSAT METEOROLOGICAL SATELLITE CONFERENCE 15/09/2013 – 20/09/2013, VIENNA EUMETSAT METEOROLOGICAL SATELLITE CONFERENCE 15/09/2013 – 20/09/2013, VIENNA.
IAAR Seminar 21 May 2013 AOD trends over megacities based on space monitoring using MODIS and MISR Pinhas Alpert 1,2, Olga Shvainshtein 1 and Pavel Kishcha.
Collection 6 update: MODIS ‘Deep Blue’ aerosol Andrew M. Sayer, N. Christina Hsu, Corey Bettenhausen, Myeong-Jae Jeong, Jaehwa Lee.
Surface Reflectivity from OMI: Effects of Snow on OMI NO 2 Gray O’Byrne 1, Randall Martin 1,2, Aaron van Donkelaar 1, Joanna Joiner 3, Edward A. Celarier.
Advances in Applying Satellite Remote Sensing to the AQHI Randall Martin, Dalhousie and Harvard-Smithsonian Aaron van Donkelaar, Akhila Padmanabhan, Dalhousie.
MAPSS and AeroStat: integrated analysis of aerosol measurements using Level 2 Data and Point Data in Giovanni Maksym Petrenko Charles Ichoku (with the.
1 of 26 Characterization of Atmospheric Aerosols using Integrated Multi-Sensor Earth Observations Presented by Ratish Menon (Roll Number ) PhD.
Satellite observations of AOD and fires for Air Quality applications Edward Hyer Naval Research Laboratory AQAST June, Madison, Wisconsin 15 June.
The Second TEMPO Science Team Meeting Physical Basis of the Near-UV Aerosol Algorithm Omar Torres NASA Goddard Space Flight Center Atmospheric Chemistry.
14 ARM Science Team Meeting, Albuquerque, NM, March 21-26, 2004 Canada Centre for Remote Sensing - Centre canadien de télédétection Geomatics Canada Natural.
QA filtering of individual pixels to enable a more accurate validation of aerosol products Maksym Petrenko Presented at MODIS Collection 7 and beyond Retreat.
Provenance in Earth Science Gregory Leptoukh NASA GSFC.
Fog- and cloud-induced aerosol modification observed by the Aerosol Robotic Network (AERONET) Thomas F. Eck (Code 618 NASA GSFC) and Brent N. Holben (Code.
As components of the GOES-R ABI Air Quality products, a multi-channel algorithm similar to MODIS/VIIRS for NOAA’s next generation geostationary satellite.
Data was collected from various instruments. AOD values come from our ground Radiometer (AERONET) The Planetary Boundary Layer (PBL) height is collected.
Satellite Aerosol Validation Pawan Gupta NASA ARSET- AQ – GEPD & SESARM, Atlanta, GA September 1-3, 2015.
Ambiguity of Quality in Remote Sensing Data Christopher Lynnes, NASA/GSFC Greg Leptoukh, NASA/GSFC Funded by : NASA’s Advancing Collaborative Connections.
Synergy of MODIS Deep Blue and Operational Aerosol Products with MISR and SeaWiFS N. Christina Hsu and S.-C. Tsay, M. D. King, M.-J. Jeong NASA Goddard.
Characterization of Aerosol Data Quality from MODIS for Coastal Regions Jacob Anderson Mentor: Gregory Leptoukh.
Experiences Developing a Semantic Representation of Product Quality, Bias, and Uncertainty for a Satellite Data Product Patrick West 1, Gregory Leptoukh.
Uncertainty in aerosol retrievals: interaction with the community Adam Povey 1, Thomas Holzer-Popp 2, Gareth Thomas 3, Don Grainger 1, Gerrit de Leeuw.
Estimating PM 2.5 from MODIS and MISR AOD Aaron van Donkelaar and Randall Martin March 2009.
Integrating satellite fire and aerosol data into air quality models: recent improvements and ongoing challenges Edward Hyer Naval Research Laboratory 6.
SCIAMACHY TOA Reflectance Correction Effects on Aerosol Optical Depth Retrieval W. Di Nicolantonio, A. Cacciari, S. Scarpanti, G. Ballista, E. Morisi,
Initial Analysis of the Pixel-Level Uncertainties in Global MODIS Cloud Optical Thickness and Effective Particle Size Retrievals Steven Platnick 1, Robert.
GEWEX Aerosol Assessment Panel members Sundar Christopher, Rich Ferrare, Paul Ginoux, Stefan Kinne, Jeff Reid, Paul Stackhouse Program Lead : Hal Maring,
INTEGRATING SATELLITE AND MONITORING DATA TO RETROSPECTIVELY ESTIMATE MONTHLY PM 2.5 CONCENTRATIONS IN THE EASTERN U.S. Christopher J. Paciorek 1 and Yang.
Direct aerosol radiative effects based on combined A-Train observations Jens Redemann, Y. Shinozuka, J. Livingston, M. Vaughan, P. Russell, M.Kacenelenbogen,
Rationale for a Global Geostationary Fire Product by the Global Change Research Community Ivan Csiszar - UMd Chris Justice - UMd Louis Giglio –UMd, NASA,
Interannual Variability and Decadal Change of Solar Reflectance Spectra Zhonghai Jin Costy Loukachine Bruce Wielicki (NASA Langley research Center / SSAI,
MODIS Atmosphere Products: The Importance of Record Quality and Length in Quantifying Trends and Correlations S. Platnick 1, N. Amarasinghe 1,2, P. Hubanks.
International Ocean Color Science Meeting, Darmstadt, Germany, May 6-8, 2013 III. MODIS-Aqua normalized water leaving radiance nLw III.1. R2010 vs. R2012.
A comparison of AMSR-E and AATSR SST time-series A preliminary investigation into the effects of using cloud-cleared SST data as opposed to all-sky SST.
Incorporating Satellite Time-Series data into Modeling Watson Gregg NASA/GSFC/Global Modeling and Assimilation Office Topics: Models, Satellite, and In.
MODIS Atmosphere Group Summary Summary of modifications and enhancements in collection 5 Summary of modifications and enhancements in collection 5 Impacts.
Fourth TEMPO Science Team Meeting
Ambiguity of Quality in Remote Sensing Data
OVERVIEW OF THE AEROSTAT PROJECT
Using dynamic aerosol optical properties from a chemical transport model (CTM) to retrieve aerosol optical depths from MODIS reflectances over land Fall.
Presentation transcript:

Experiences Developing a Semantic Representation of Product Quality, Bias, and Uncertainty for a Satellite Data Product Patrick West 1, Gregory Leptoukh 2, Stephan Zednik 1, Chris Lynnes 2, Suraiya Ahmad 3, Jianfu Pan 4, Peter Fox 1 (1)Tetherless World Constellation (2)NASA Goddard Space Flight Center (3)NASA Goddard Space Flight Center/Innovim (4)NASA Goddard Space Flight Cetner/Adnet Systems, Inc. EGU

Issue Climate model and various environmental monitoring and protection applications have begun to increasingly rely on satellite measurements. Research application users seek good quality satellite data, with uncertainties and biases provided for each data point Remote-sensing quality issues are addressed rather inconsistently and differently by different communities. 19/14/2015EGU 2011

Where are we in respect to this data challenge? “The user cannot find the data; If he can find it, cannot access it; If he can access it, ; he doesn't know how good they are; if he finds them good, he can not merge them with other data” The Users View of IT, NAS /14/20152EGU 2011

Challenges in dealing with Data Quality Q: Why now? What has changed? A: With the recent revolutionary progress in data systems, dealing with data from many different sensors finally has become a reality. Only now, a systematic approach to remote sensing quality is on the table. NASA is beefing up efforts on data quality. ESA is seriously addressing these issues. QA4EO: an international effort to bring communities together on data quality. GeoVique 9/14/20153EGU 2011

Data from multiple sources to be used together: Current sensors/missions: MODIS, MISR, GOES, OMI. Future missions: ACE, NPP, JPSS, Geo-CAPE European and other countries’ satellites Models Harmonization needs: It is not sufficient just to have the data from different sensors and their provenances in one place Before comparing and fusing data, things need to be harmonized: Metadata: terminology, standard fields, units, scale Data: format, grid, spatial and temporal resolution, wavelength, etc. Provenance: source, assumptions, algorithm, processing steps Quality: bias, uncertainty, fitness-for-purpose, validation Quality: bias, uncertainty, fitness-for-purpose, validation Dangers of easy data access without proper assessment of the joint data usage - It is easy to use data incorrectly Intercomparison of data from multiple sensors 9/14/20154EGU 2011

Three projects with data quality flavor We have three projects where different aspects of data quality are addressed. We mostly deal with aerosol data I’ll briefly describe them and then show why they are related 59/14/2015EGU 2011

Data Quality Screening Service for Remote Sensing Data The DQSS filters out bad pixels for the user Default user scenario –Search for data –Select science team recommendation for quality screening (filtering) –Download screened data More advanced scenario –Search for data –Select custom quality screening parameters –Download screened data 69/14/2015EGU 2011

DQSS Ontology (The Whole Enchilada) 79/14/2015EGU 2011

DQSS Ontology (Zoom) 89/14/2015EGU 2011

AeroStat: Online Platform for the Statistical Intercomparison of Aerosols 9 Explore & Visualize Level 3 Compare Level 3 Correct Level 2 Compare Level 2 Before and After Merge Level 2 to new Level 3 Level 3 are too aggregated Switch to high-res Level 2 9/14/2015 Explore & Visualize Level 2 9 EGU 2011

IQ Qurator Model Qurator info model used our assessment of existing quality models. Describes a model whereby: – data annotated with quality annotation metadata QA metadata can be associated with data of varying degrees of granularity –ex: products, collections, arrays, specific values, etc. –this supports our interest in associated data with a product Quality evidence, a measurable quantity, provides a 'clue' into the quality –ex: hit-ratio, standard deviation, etc. –common examples associated with statistical analysis –often computed in QC –would global coverage, scatter plots, etc. fit? Quality assertions are domain-specific functions based on quality evidence –good, bad, ugly –No confidence, marginal, good, best Quality property (aka quality dimensions) –accuracy, completeness, currency –many dimensions of quality to consider, each with different evidence 10

IQ Curator Model Application to our Project 11

Application to Focus Area 129/14/2015EGU 2011

Multi-Sensor Data Synergy Advisor (MDSA) Goal: Provide science users with clear, cogent information on salient differences between data candidates for fusion, merging and intercomparison –Enable scientifically and statistically valid conclusions Develop MDSA on current missions: –Terra, Aqua, (maybe Aura) Define implications for future missions 139/14/2015EGU 2011

Reference: Hyer, E. J., Reid, J. S., and Zhang, J., 2011: An over-land aerosol optical depth data set for data assimilation by filtering, correction, and aggregation of MODIS Collection 5 optical depth retrievals, Atmos. Meas. Tech., 4, , doi: /amt Title: MODIS Terra C5 AOD vs. Aeronet during Aug-Oct Biomass burning in Central Brazil, South America (General) Statement: Collection 5 MODIS AOD at 550 nm during Aug- Oct over Central South America highly over-estimates for large AOD and in non-burning season underestimates for small AOD, as compared to Aeronet; good comparisons are found at moderate AOD. Region & season characteristics: Central region of Brazil is mix of forest, cerrado, and pasture and known to have low AOD most of the year except during biomass burning season (Example) : Scatter plot of MODIS AOD and AOD at 550 nm vs. Aeronet from ref. (Hyer et al, 2011) (Description Caption) shows severe over-estimation of MODIS Col 5 AOD (dark target algorithm) at large AOD at 550 nm during Aug- Oct over Brazil. (Constraints) Only best quality of MODIS data (Quality =3 ) used. Data with scattering angle > 170 deg excluded. (Symbols) Red Lines define regions of Expected Error (EE), Green is the fitted slope Results: Tolerance= 62% within EE; RMSE=0.212 ; r2=0.81; Slope=1.00 For Low AOD ( 1.4) Slope=1.54 (Dominating factors leading to Aerosol Estimate bias): 1.Large positive bias in AOD estimate during biomass burning season may be due to wrong assignment of Aerosol absorbing characteristics. (Specific explanation) a constant Single Scattering Albedo ~ 0.91 is assigned for all seasons, while the true value is closer to ~ [ Notes or exceptions: Biomass burning regions in Southern Africa do not show as large positive bias as in this case, it may be due to different optical characteristics or single scattering albedo of smoke particles, Aeronet observations of SSA confirm this] 2. Low AOD is common in non burning season. In Low AOD cases, biases are highly dependent on lower boundary conditions. In general a negative bias is found due to uncertainty in Surface Reflectance Characterization which dominates if signal from atmospheric aerosol is low Aeronet AOD Central South America * Mato Grosso * Santa Cruz * Alta Floresta

FACETS OF DATA QUALITY 159/14/2015EGU 2011

Quality Control vs. Quality Assessment Quality Control (QC) flags in the data (assigned by the algorithm) reflect “happiness” of the retrieval algorithm, e.g., all the necessary channels indeed had data, not too many clouds, the algorithm has converged to a solution, etc. Quality assessment is done by analyzing the data “after the fact” through validation, intercomparison with other measurements, self-consistency, etc. It is presented as bias and uncertainty. It is rather inconsistent and can be found in papers, validation reports all over the place. 9/14/201516EGU 2011

Different kinds of reported and perceived data quality Pixel-level Quality (reported): algorithmic guess at usability of data point (some say it reflects the algorithm “happiness”) –Granule-level Quality: statistical roll-up of Pixel-level Quality Product-level Quality (wanted/perceived): how closely the data represent the actual geophysical state Record-level Quality: how consistent and reliable the data record is across generations of measurements Different quality types are often erroneously assumed having the same meaning Different focus and action at these different levels to ensure Data Quality 17EGU 20119/14/2015

Percent of Biased Data in MODIS Aerosols Over Land Increase as Confidence Flag Decreases *Compliant data are within  Aeronet Statistics from Hyer, E. J., Reid, J. S., and Zhang, J., 2011: An over-land aerosol optical depth data set for data assimilation by filtering, correction, and aggregation of MODIS Collection 5 optical depth retrievals, Atmos. Meas. Tech., 4, , doi: /amt /14/201518EGU 2011

General Level 2 Pixel-Level Issues How to extrapolate validation knowledge about selected Level 2 pixels to the Level 2 (swath) product? How to harmonize terms and methods for pixel-level quality? 19EGU 2011 AIRS Quality Indicators MODIS Aerosols Confidence Flags 0 Best Data Assimilation 1Good Climatic Studies 2Do Not Use Use these flags in order to stay within expected error bounds 3Very Good 2 Good 1 Marginal 0 Bad 3Very Good 2 Good 1 Marginal 0 Bad Ocean Land ±0.05 ± 0.15 t ±0.03 ± 0.10 t Ocean Land Purpose Match up the recommendations? 9/14/2015

Spatial and temporal sampling – how to quantify to make it useful for modelers? MODIS Aqua AOD July 2009MISR Terra AOD July 2009 Completeness: MODIS dark target algorithm does not work for deserts Representativeness: monthly aggregation is not enough for MISR and even MODIS Spatial sampling patterns are different for MODIS Aqua and MISR Terra: “pulsating” areas over ocean are oriented differently due to different direction of orbiting during day-time measurement  Cognitive bias 20EGU 20119/14/2015

Addressing Level 3 data “quality” Terminology: Quality, Uncertainty, Bias, Error budget, etc. Quality aspects (examples): –Completeness: Spatial (MODIS covers more than MISR) Temporal (Terra mission has been longer in space than Aqua) Observing Condition (MODIS cannot measure over sun glint while MISR can) –Consistency: Spatial (e.g., not changing over sea-land boundary) Temporal (e.g., trends, discontinuities and anomalies) Observing Condition (e.g., exhibit variations in retrieved measurements due to the viewing conditions, such as viewing geometry or cloud fraction) –Representativeness: Neither pixel count nor standard deviation fully express representativeness of the grid cell value 21EGU 20119/14/2015

Some differences in L3 are due to difference processing Spatial and temporal binning (L2  L3 daily) leads to Aggregation bias: –Measurements (L2 pixels) from one or more orbits can go into a single grid cell  different within-grid variability –Different weighting: pixel counts, quality –Thresholds used, i.e., > 5 pixels Data aggregation (L3D  L3monthly  regional  global): –Weighting by pixel counts or quality –Thresholds used, i.e., > 2 days While these algorithms have been documented in ATBD, reports and papers, the typical data user is not immediately aware of how a given portion of the data has been processed, and what is the resulting impact EGU /14/2015

Case 1: MODIS vs. MERIS Same parameterSame space & time Different results – why? MODIS MERIS A threshold used in MERIS processing effectively excludes high aerosol values. Note: MERIS was designed primarily as an ocean-color instrument, so aerosols are “obstacles” not signal.

Case 2: Aggregation Mishchenko et al., 2007 The AOD difference can be up to 40% due to differences in aggregation Levy, Leptoukh, et al., 2009 AOD difference between sensors MODIS Terra only AOD: difference between diff. aggregations

EGU Case 3: DataDay definition MODIS-Terra vs. MODIS-Aqua: Map of AOD temporal correlation, 2008 MODIS Level 3 dataday definition leads to artifact in correlation 9/14/2015

Conclusion Quality is very hard to characterize, different groups will focus on different and inconsistent measures of quality. Products with known Quality (whether good or bad quality) are more valuable than products with unknown Quality. –Known quality helps you correctly assess fitness-for-use Harmonization of data quality is even more difficult that characterizing quality of a single data product 269/14/2015EGU 2011

References Levy, R. C., Leptoukh, G. G., Zubko, V., Gopalan, A., Kahn, R., & Remer, L. A. (2009). A critical look at deriving monthly aerosol optical depth from satellite data. IEEE Trans. Geosci. Remote Sens., 47, Zednik, S., Fox, P., & McGuinness, D. (2010). System Transparency, or How I Learned to Worry about Meaning and Love Provenance! 3rd International Provenance and Annotation Workshop, Troy, NY. P. Missier, S. Embury, M.Greenwood, A. Preece, and B. Jin. Quality views: capturing and exploiting the user perspective on data quality. Procs VLDB, (PDF) _vldb2006.pdfhttp://users.cs.cf.ac.uk/A.D.Preece/qurator/resources/qurator _vldb2006.pdf 279/14/2015EGU 2011