NPOESS Preparatory Project (NPP) Science Data Segment (SDS) Ocean Product Evaluation and Analysis Tool Element (PEATE) Peer Review November 30, 2005.

Slides:



Advertisements
Similar presentations
Network II.5 simulator ..
Advertisements

ASIAES Project Overview Satellite Image Network for Natural Hazard Management in ASEAN+3 region Pakorn Apaphant Geo-Informatics and Space Technology Development.
Module 13: Performance Tuning. Overview Performance tuning methodologies Instance level Database level Application level Overview of tools and techniques.
Global Processing of MODIS for Operational SST, Ocean Color, and GHRSST Bryan Franz and the NASA Ocean Biology Processing Group 8th GHRSST-PP Science Team.
Test Case Management and Results Tracking System October 2008 D E L I V E R I N G Q U A L I T Y (Short Version)
OVERVIEW TEAM5 SOFTWARE The TEAM5 software manages personnel and test data for personal ESD grounding devices. Test and personnel data may be viewed/reported.
VIIRS Land PEATE Masuoka, Wolfe, Isaacman and Devadiga.
Validata Release Coordinator Accelerated application delivery through automated end-to-end release management.
The new The new MONARC Simulation Framework Iosif Legrand  California Institute of Technology.
Linux Networking CIS Why Linux/Unix? Configurability ▫Customizable System to satisfy unique needs. Scalability ▫Able to serve an increasing number.
Check Disk. Disk Defragmenter Using Disk Defragmenter Effectively Run Disk Defragmenter when the computer will receive the least usage. Educate users.
Web-based Portal for Discovery, Retrieval and Visualization of Earth Science Datasets in Grid Environment Zhenping (Jane) Liu.
Suomi National Polar-orbiting Partnership (SNPP) Data Access NOAA Satellite Conference April 8-12, 2013 Kevin Berberich NESDIS/OSD NDE Project Photographs.
System Design/Implementation and Support for Build 2 PDS Management Council Face-to-Face Mountain View, CA Nov 30 - Dec 1, 2011 Sean Hardman.
Microsoft ® Application Virtualization 4.6 Infrastructure Planning and Design Published: September 2008 Updated: February 2010.
Grid Computing Meets the Database Chris Smith Platform Computing Session #
Satellite Data Provided by the NASA Ocean Biology Processing Group (OBPG)
The SAM-Grid Fabric Services Gabriele Garzoglio (for the SAM-Grid team) Computing Division Fermilab.
The MODIS online archive and on-demand processing Edward Masuoka NASA Goddard Space Flight Center, Greenbelt, MD, USA.
Overview Introduction Missions Evolution Philosophy Software Components and Subsystems: > RDBMS > Ingest > VDC/Scheduler > Distribution Scientific Support.
15 Maintaining a Web Site Section 15.1 Identify Webmastering tasks Identify Web server maintenance techniques Describe the importance of backups Section.
DCS Overview MCS/DCS Technical Interchange Meeting August, 2000.
MASSACHUSETTS INSTITUTE OF TECHNOLOGY NASA GODDARD SPACE FLIGHT CENTER ORBITAL SCIENCES CORPORATION NASA AMES RESEARCH CENTER SPACE TELESCOPE SCIENCE INSTITUTE.
1 Next Generation of Operational Earth Observations From the National Polar-Orbiting Operational Environmental Satellite System (NPOESS): Program Overview.
Aquarius Project Status 7 th Aquarius/SAC-D Science Team Meeting 11 April 2012.
Page 1 Ocean PEATE Fred Patt NPP SDS Peer MOR Building 8 Auditorium August 27, 2008.
OOI CI LCA REVIEW August 2010 Ocean Observatories Initiative OOI Cyberinfrastructure Architecture Overview Michael Meisinger Life Cycle Architecture Review.
Silberschatz, Galvin and Gagne  Operating System Concepts Chapter 3: Operating-System Structures System Components Operating System Services.
ASI-Eumetsat Meeting Matera, 4-5 Feb CNM Context Matera, February 4-5, 20092ASI-Eumetsat Meeting.
Module 13 Implementing Business Continuity. Module Overview Protecting and Recovering Content Working with Backup and Restore for Disaster Recovery Implementing.
Lecture 22: Client-Server Software Engineering
VIPER Quality Assessment Overview Presenter: Sathyadev Ramachandran, SPS Inc.
NA-MIC National Alliance for Medical Image Computing UCSD: Engineering Core 2 Portal and Grid Infrastructure.
CLASS Information Management Presented at NOAATECH Conference 2006 Presented by Pat Schafer (CLASS-WV Development Lead)
Diane E. Wickland NPP Program Scientist NPP Science: HQ Perspective on VIIRS May 18, 2011.
March 2004 At A Glance autoProducts is an automated flight dynamics product generation system. It provides a mission flight operations team with the capability.
Page 1 Ocean PEATE Fred Patt. Page 2 Agenda Introduction Design Overview Design Description Systems Engineering Closing Remarks SDS Data Depository and.
Development of GHRSST L2Pcore for MODIS Bryan Franz NASA Ocean Biology Processing Group Goddard Space Flight Center.
06-1L ASTRO-E2 ASTRO-E2 User Group - 14 February, 2005 Astro-E2 Archive Lorella Angelini/HEASARC.
Robert Wolfe NASA Goddard Space Flight Center Code 614.5, Greenbelt, MD Robert Wolfe NASA Goddard Space Flight Center Code 614.5,
VIIRS Product Evaluation at the Ocean PEATE Frederick S. Patt Gene C. Feldman IGARSS 2010 July 27, 2010.
TSS Database Inventory. CIRA has… Received and imported the 2002 and 2018 modeling data Decided to initially store only IMPROVE site-specific data Decided.
SNPP Ocean SIPS status SNPP Applications Workshop 18 November 2014 Bryan Franz and the Ocean Biology Processing Group.
August 2003 At A Glance The IRC is a platform independent, extensible, and adaptive framework that provides robust, interactive, and distributed control.
National Aeronautics and Space Administration Aquarius Validation Data System Overview and Status Algorithm Workshop March 21-22, 2007 John Gunn Earth.
Aquarius Mission Simulation A realistic simulation is essential for mission readiness preparations This requires the ability to produce realistic data,
Lecture 4 Mechanisms & Kernel for NOSs. Mechanisms for Network Operating Systems  Network operating systems provide three basic mechanisms that support.
Integrating the Scientific Community with a Measurement Based, Multi-Sensor Data Processing and Distribution System Bryan Franz, Donna Thomas, Sean Bailey,
Page 1 Land PEATE support for CERES processing Ed Masuoka Gang Ye August 26, 2008 CERES Delta Design Review.
ADPS Science Software Development Bryan Franz NASA Ocean Biology Processing Group Aquarius Data Processing Workshop, NASA/GSFC, March 2007.
MODIS SDST, STTG and SDDT MODIS Science Team Meeting (Land Discipline Breakout Session) July 13, 2004 Robert Wolfe Raytheon NASA GSFC Code 922.
IPS Infrastructure Technological Overview of Work Done.
Module 6: Administering Reporting Services. Overview Server Administration Performance and Reliability Monitoring Database Administration Security Administration.
Page 1 Ocean PEATE Fred Patt. Page 2 Ocean PEATE Agenda Science Team Introduction Ocean PEATE Design ODPS Design Overview Implementation Plan and Schedule.
From Missions to Measurements: an Ocean Discipline Experience.
March 2004 At A Glance The AutoFDS provides a web- based interface to acquire, generate, and distribute products, using the GMSEC Reference Architecture.
WebScan: Implementing QueryServer 2.0 Karl Geiger, Amgen Inc. BRS NA UG August 1999.
Physical Oceanography Distributed Active Archive Center THUANG June 9-13, 20089th GHRSST-PP Science Team Meeting GHRSST GDAC and EOSDIS PO.DAAC.
Building Preservation Environments with Data Grid Technology Reagan W. Moore Presenter: Praveen Namburi.
V7 Foundation Series Vignette Education Services.
Page 1 June 26, 2016 NPOESS Preparatory Project (NPP) Science Data Segment (SDS) Ocean PEATE Status and Plans January 27, 2010 Ocean PEATE Team.
AIRS Meeting GSFC, February 1, 2002 ECS Data Pool Gregory Leptoukh.
Architecture Review 10/11/2004
Fred Patt NPP SDS Peer MOR Building 8 Auditorium August 27, 2008
Threads vs. Events SEDA – An Event Model 5204 – Operating Systems.
MODIS SST Processing and Support for GHRSST at OBPG
Design Unit 26 Design a small or home office network
Cloud computing mechanisms
Database System Concepts and Architecture
Presentation transcript:

NPOESS Preparatory Project (NPP) Science Data Segment (SDS) Ocean Product Evaluation and Analysis Tool Element (PEATE) Peer Review November 30, 2005 Ocean PEATE Team

Agenda Introduction Design Overview Ocean PEATE Implementation Schedule and Resources Issues and Conclusion

Primary Tasks of the Ocean PEATE Acquire VIIRS RDRs, SDRs, and Ocean EDRs from the SD3E and ADS/CLASS Assess the quality of the NPP Ocean EDRs for accomplishing NASA’s climate research requirements Provide suggested algorithm improvements to the IDPS via the Project Science Working Group (PSWG) Process selected data subsets in support of Evaluation and Validation activities

Requirements Overview Number Requirement Description 3.4 PRODUCT EVALUATION AND ANALYSIS TOOL ELEMENTS (PEATEs) 3.4.1 PEATEs Ingest Data 3.4.1.1 The PEATEs shall have the capability of ingesting xDRs from the SD3E. 3.4.1.2 The PEATEs shall have the capability of ingesting IPs from the SD3E. 3.4.1.3 The PEATEs shall have the capability of ingesting official ancillary data from the SD3E. 3.4.1.4 The PEATEs shall have the capability of ingesting and storing calibration products from the SD3E. 3.4.1.5 The PEATEs shall have the capability of requesting: a listing of current products in SD3E, a listing of products for SD3E to re-acquire, and a listing of products for SD3E to retransmit 3.4.1.6 The PEATEs shall provide the capability of submitting product subscriptions to the ADS as a request for data delivery. 3.4.1.7 The PEATEs shall be capable of submitting ad-hoc requests to the ADS. 3.4.1.8 The PEATEs shall have the capability of ingesting xDRs from the ADS.

Requirements Overview Number Requirement Description 3.4.1.9 The PEATEs shall have the capability of ingesting official ancillary data from the ADS. 3.4.1.10 The PEATEs shall have the capability of ingesting and storing calibration products from ADS. 3.4.1.11 The PEATEs shall have the capability of ingesting IP from the ADS. 3.4.1.12 The PEATEs shall have the capability of ingesting status information about queries made to the ADS. 3.4.1.13 The PEATEs shall have the capability of ingesting additional ancillary data from other sources (e.g., NOAA/NESDIS, NASA, USGS, JPL, NOAA Space Environment Center, USNAVO) as deemed appropriate by each PEATE. 3.4.1.14 The PEATEs shall have the capability of ingesting pre-launch algorithms, calibration, test data files, and instrument parameters from the NEXT system. 3.4.1.15 The PEATEs shall have the capability of ingesting engineering reports from the C3S

Requirements Overview Number Requirement Description 3.4.2 Store/Catalog Data 3.4.2.1 The PEATEs shall have the capability of storing and cataloging all data needed for EDR evaluation. 3.4.3 Manage Software Configuration 3.4.3.1 The PEATEs shall have the capability of performing source code configuration control for all software and documentation developed or enhanced.

Requirements Overview Number Requirement Description 3.4.4 Process Science Data 3.4.4.1 The PEATEs shall have the capability of ingesting the operational algorithms for processing from the SD3E. 3.4.4.2 The PEATEs shall have the capability of receiving management directions (e.g., priority changes) from the PSOE. 3.4.4.3 The PEATEs shall have the capability of receiving instrument service reports from the PSOE. 3.4.4.4 The PEATEs shall have the capability of receiving Cal LUTs from the NICSE. 3.4.4.5 The PEATEs shall have the capability of receiving algorithm status from the I&TSE. 3.4.4.6 The PEATEs shall have the capability of supporting the Science Team in pre-launch assessment of RDR, SDR, EDR, and Intermediate Products. 3.4.4.7 The PEATE shall support the Science Team in assessment of xDR and IP science algorithm functionality and implementation. 3.4.4.8 The PEATEs shall have the capability of supporting the Science Team in obtaining, examining, and running EDR operational algorithm from the IDPS.

Requirements Overview Number Requirement Description 3.4.4.9 The PEATEs shall have the capability of supporting the Science Team in testing and characterizing candidate improvements/enhancements made to SDR and EDR science or operational algorithm software. 3.4.4.10 The PEATEs shall have the capability of supporting the Science Team in developing comparison tools (e.g., IDL programming, FORTRAN and C programming, etc.). 3.4.4.11 The PEATEs shall have the capability of supporting the Science Team in comparing products against data from other satellite missions and ground truth measurement campaigns. 3.4.4.12 The PEATE shall have the capability of supporting the Science Team in running comparisons between standard and alternate products and in analyzing comparison results. 3.4.4.13 The PEATEs shall have the capability of supporting the Science Team in analyzing comparison results and developing candidate science algorithm improvements and enhancements. 3.4.4.14 The PEATEs shall have the capability of supporting processing of RDRs into SDRs using software used by IDPS with alternate LUTs provided by the Science Team.

Requirements Overview Number Requirement Description 3.4.4.15 The PEATEs shall have the capability of supporting processing of SDRs into EDRs using software used by IDPS with alternate LUTs generated by the Science Team. 3.4.4.16 The PEATEs shall have the capability of supporting the Science Team in their development and generation of data for testing candidate science algorithms. 3.4.4.17 The PEATE shall have the capability of supporting the Science Team in developing candidate SDR (L1B) science algorithm improvements/enhancements. 3.4.4.18 The PEATE shall have the capability of supporting the Science Team in developing candidate EDR (L2 and L3) science algorithm improvements/ enhancements. 3.4.4.19 The PEATE shall have the capability of supporting the Science Team in testing and evaluating candidate improvements/enhancements made to SDR and EDR operational algorithm. 3.4.4.20 The PEATEs shall have the capability of supporting the Science Team in documenting software that was developed in the course of the mission. 3.4.4.21 The PEATEs shall have the capability of providing a status report to the PSOE.

Requirements Overview Number Requirement Description 3.4.4.22 The PEATEs shall have the capability of requesting instrument service reports from the PSOE. 3.4.4.23 The PEATEs shall have the capability of processing calibration quality control requests for calibration validation from NICSE. 3.4.4.24 The PEATEs shall have the capability of providing NICSE with the results of calibration evaluation.

Requirements Overview Number Requirement Description 3.4.5 PEATEs Export Data 3.4.5.1 The PEATEs shall have the capability of exporting all science data collected to the Science Team 3.4.5.2 The PEATEs shall have the capability of permitting the Science Team to place standing orders for xDRs and ancillary data stored in the PEATEs. 3.4.5.3 The PEATEs shall have the capability of pushing science data to the Science Team members. 3.4.5.4 The PEATEs shall have the capability of delivering data to Science Team members in response to ad-hoc browse and order sessions. 3.4.5.5 The PEATEs shall have the capability of compressing data that they deliver to users.

Requirements Overview Number Requirement Description 3.4.8 Ocean PEATE 3.4.8.1 The NPP Ocean PEATE shall assess short-term and long-term quality, through independent means, of the NPP Ocean Color and Sea Surface Temperature EDR for climate research. 3.4.8.2 The NPP Ocean PEATE shall provide to the NPP Science Team the capabilities necessary to validate NPP Ocean Product as listed in the EDR allocations Table in Appendix D, or Enhanced EDRs developed by the Science Team in cases where the EDR is insufficient for climate research.

Agenda Introduction Design Overview Ocean PEATE Implementation Schedule and Resources Issues and Conclusion

Ocean PEATE External Interfaces • SDS Science Data Distribution and Depository Element (SD3E) Provides NRT access to raw data Primary source of RDRs Provides selected SDRs and EDRs SDS Integration and Test System Element (I&TSE) Test updates to operational code prior to delivery to IDPS Archive Distribution Segment (ADS) Primary source for archived data xDRs, IPs, Ancillary Data, Operational Algorithm/Source Code and Calibration Products Ancillary Data Providers (ADP) Provides alternate ancillary data sets (e.g., ozone, meteorological data sets) CasaNOSA Serves as the NPP pre-flight repository of Government held data for distribution to Government user teams Place to acquire pre-launch NPP algorithms and supported data files NASA VIIRS Ocean Science Team (VOST) Coordinate activities with PEATEs and PSOE on xDR and recommended algorithm improvements. Supports Independent Calibration Validation Activities NPP Instrument Calibration Support Element (NICSE) Provides alternative calibration LUTs and recommended improvements to calibration algorithms PEATE provides results of LUT and algorithm tests Project Science Office Element (PSOE) Provides management direction Accepts algorithm update recommendations Ocean Science Community Relies on Ocean PEATE to provide evaluation products and results

Ocean PEATE Interface Diagram SD3E VOST Software, Data xDR Eval. Results, Algorithm Updates RDRs, IPs, Ancillary Data CLASS (ADS) Casa- NOSA Processing Request xDRs, IPs, Ancillary Data (if unavailable from SD3E) Ocean PEATE Pre-flight Algorithms, Data, Info Ancillary Data Providers I&TSE Ancillary Data Algorithm Updates NICSE Calibration Updates and Evaluations Ocean Science Community Interaction Management Direction Analysis Results, Proposed Algorithm Updates PSOE

ODPS Design This section provides an overview of the history, architecture, design philosophy, configuration management, features, components, and lessons learned for the current Ocean Color Data Acquisition, Processing, Quality-Control, Archive, and Distribution system. This system has been successfully supporting operational, satellite-based remote-sensing missions since 1996, and its capabilities continue to evolve and expand to meet the demands and challenges of future missions.

Design Overview Fully automated, distributed data system for acquiring, processing, archiving, and distributing scientific data Highly scalable Easily adaptable to support multiple concurrent missions Graphical user interfaces for controlling and monitoring system functions and activity Non-platform specific

Design Philosophy Building-Block approach Programs usually small and do one thing well Programs are less complex and subsequently easy to maintain Promotes reuse Programs loosely coupled so testing and production can be done in the same environment Adopt basic standards ANSI, POSIX, C9x Use existing technology when possible Exit statuses indicate successful or failure conditions

Architecture: Hardware Processing Servers Intel-based dual Xeon / AMD-based dual Opteron 8 GB RAM 5 72-GB SCSI drives Storage Servers Intel-based P4 / AMD-based single Opteron 2 GB RAM 1.2 TB IDE RAID 5 (3ware) / 5.1 TB SATA RAID 6 (Areca) 2 hot spare drives Database Server Sun V880 9 70-GB SCSI HDD

ODPS Data Processing System Current Components Distribution Servers (ftp) 6 storage nodes 7.2 TB Distribution Servers (web)1 large server – 3 TB 58 storage nodes 69.6 TB Ingest Servers 2 SeaSpace ground stations 5 storage nodes 6 TB Cal/Val & QC Systems Processing Cluster 34 processing nodes 1.5 TB Extreme Networks Black Diamond 6816 Gigabit Ethernet switch Mission Operations Systems Database Server 1 large server 876 GB Development Servers 1 processing node 2 storage nodes 2.4 TB Testing Cluster 13 test nodes 2 TB Backup Servers 1 large server 876 GB+2 TB tape 5 storage nodes 6 TB User Desktops Network Support Systems

Architecture: Software System Software C Perl Shell (sh/csh) SQL Additional Software Required Sybase Adaptive Server Enterprise (12.5.3) Open-Client CT-Library Perl DBI Module X/Motif (Open Motif 2.2) GMT 3.4.1 ImageMagick 5.5.6 Netpbm 9.24

Software Configuration Management Science software maintained separately from system software Separate development, testing, and operational environments allow new features to be added with no impact on operational activities All environments are password protected to restrict access to authorized developers and operators Science software is delivered as precompiled binary executables to eliminate the possibility of changes being introduced by a different build environment

Software Configuration Management Science software versions are stored in the product- catalog database tables to provide a trace back to the specific build that created a product Major changes to scientific and system software configuration are recorded in a Mission Events and Major Changes log, which can be publicly accessed via the Ocean Color web site System and science code maintained under source-code version-control system, currently Subversion (subversion.tigris.org) in use

System Features Graphical user interfaces with color-coded displays allow system status to be determined at a glance and provide a mechanism for controlling system functions Distributed architecture provides high level of scalability Automated active and passive data-acquisition mechanisms Dynamic allocation of user-defined processing resources Easily configurable report manager provides additional operator “eyes” Token-based network load balancing

System Features (cont.) Innovative disk-resource management for supporting logical archive pools Host and disk monitors poll system resources and toggle their availability for use within the system Prioritized processing capability Host-based processing constraints Data-specific pre-processing rule enforcement Event- and time-based file migration and management Back end for web-based data ordering and distribution

System Adaptability Scalable architecture Generic core components are non-data specific Scalable architecture Non-platform dependent, currently SGI IRIX and Linux (ix86) supported Shell wrappers allow new processing algorithms to be quickly adapted into production Support for ADEOS OCTS mission in less than two weeks and ported a fully functional system to NOAA in one week

Components and Subsystems

Components and Subsystems: RDBMS Primary element that manages all system activity Generic core databases support system infrastructure and non-mission-specific functions Mission databases catalogue products and house mission- specific data and procedures High level of reuse possible for similar missions e.g. MODIS Aqua/Terra, SeaWiFS, and OCTS are all ocean-color missions and have similar product suites and requirements

Components and Subsystems Relational Database Management System (RDBMS) supports all of the system components (subsystems) Scheduler is the primary controlling module within the system, supporting both time- and event-based tasks Other subsystems are independent modules, yet rely on the Scheduler for some their functions Scheduler Visual Database Cookbook (VDC) Archive Device Manager (ADM) Data acquisition and ingest Level-3 Scheduler File migration and management Data distribution

Agenda Introduction Design Overview Ocean PEATE Implementation Schedule and Resources Issues and Conclusion

Ocean PEATE-Specific Tasks Acquire and ingest xDRs from the SD3E: VIIRS RDRs VIIRS CaDRs (calibration RDRS) VIIRS SDRs OCC EDRs SST EDRs VIIRS IPs Acquire and ingest xDRs from ADS/CLASS (same as above except maybe not RDRs) Acquire and ingest NPP ancillary data sets Catalog and manage NPP data sets (all of the above)

Ocean PEATE-Specific Tasks (cont.) Support evaluation processing of VIIRS data using IDPS operational code: SDRs, OCC EDRs, SST EDRs Support VIIRS calibration analysis using CaDRs (solar and lunar calibrations) Perform matchups of VIIRS data with SeaBASS data Support cross-comparison of VIIRS ocean data with concurrent sensor data sets Support cross-comparison of VIIRS ocean data with climatological data sets Support internal consistency evaluation of VIIRS ocean data - Interannual repeatability in deep and clear water

Data Acquisition and Ingest Insert DB records for new data-source servers and data types Compose data-specific post-ingest scripts Configure ingest daemons if that method is going to be used for any of the new data types Define archive-device pools for product storage

Data Cataloging Insert record into the core tables (catalog DB) that describe the mission and products Create mission specific database and objects, reusing objects from existing mission databases where applicable Compose a program to provide geographical L1 meta-data information including granule start and stop times, day-night flag, and geographic coordinates, e.g.: MSl1info Compose functions for DB-metaload program, so meta-data files can be read and product tables can be populated Configure file migration and management actions for new mission data

Data Processing Streams Define a recipe for each distinct serial processing stream Insert records for each recipe and each recipe step Compose a job template for each recipe Compose a ancillary-selection procedure for each recipe that requires ancillary data Compose wrapper scripts for each science program associated with the new mission's data Insert record in recipe-constraints table for each processing host allowed to run a recipe Compose AP-load procedure for each base data type that can be processed with a recipe Update the reproc program to support each base data type that has an AP-load procedure Configure L3-Scheduler for desired composite processing

Data Distribution Update Subscription CGI to support new mission data Compose match-subscription procedure If data extraction and mapping is to be supported: Compose extraction and mapping programs Compose match-XM-requests procedure Modify XM CGI to support new mission products Compose browser capabilities for new mission products Provide FTP access to new mission products

Calibration Analyses NICSE has primary responsibility for radiometic calibration of VIIRS VOST will provide support and supplemental analyses to achieve radiometric accuracy needed for Ocean products Temporal radiometric stability has been achieved for SeaWiFS over the 8 year mission using lunar calibrations Vicarious calibration using surface measurements gives constant gain correction

Matchup Analysis Ocean data granules in ODPS catalog are automatically matched with in situ data SeaWiFS Bio-optical Archive and Storage System (SeaBASS) stores and manages in situ holdings from field programs and supported investigators. Ocean staff acquire, QC and analyze new data samples Over 300,000 in situ samples stored

Matchup Process Flow Chart

Sensor Cross-Comparisons Level-3 parameters (e.g., nLw) compared for common spectral bands Common bins extracted and compared over the period of overlap between the sensors Comparisons are performed globally (deep water, clear water, coastal), zonally and for specified regions.

Improvement in MODIS-SeaWiFS Comparisons

Internal Consistency Analysis Global averages from successive years are overplotted to determine interannual repeatability.

Level-3 Product Generation Sensor cross-comparisons and interannual comparisons require Level-3 binned (equal-area) products. The Ocean PEATE will implement software to process VIIRS EDRs to Level 3 binned products in current ODPS (SeaWiFS-like) format. Use current binning code with new input functions to read EDRs This will automatically provide the additional capabilities to produce multi-temporal composites, standard mapped image (SMI) products and Level-3 browse files.

Agenda Introduction Design Overview Operational Scenarios Schedule and Resources Issues and Conclusion

Schedule Initial Capability (L-18 months) All interfaces fully implemented and tested Initial (build 1.3?) versions of operational code ported and running in ODPS L-3 product code developed and tested Data storage capacity for ? months Initial test products generated for review by VIIRS Ocean Science Team Full Mission Capability (L-12 months) Routine exercise of interfaces to acquire proxy, surrogate (Aqua?) and/or simulated data Prelaunch (build 1.4?) versions of operational code running in ODPS Browse and distribution capability developed and tested Test products routinely generated based on simulated data and posted for access by VIIRS Ocean Science Team Data storage for ? years

Data Storage Estimate Data Type Daily 1 Year 5 Years RDR 150 GB 53.5 TB 267.5 TB SDR (M-band) 194 GB 7 TB (*) 35 TB (*) OCC EDR 38 GB 1.4 TB (*) SST EDR 17 GB 0.6 TB (*) 3 TB (*) Inter. Products 70 GB N/A Ancillary Data 0.1 GB .04 TB .2 TB Total 469 GB 62.5 TB 312.7 TB Assumption: (*) Long-term storage is sized for 100% of RDRs and 10% of SDRs and EDRs; packaged without geolocation.

Ocean PEATE Testing Initial Capability Testing Data acquisition and ingest testing will depend upon availability of external interfaces and convergence of data product formats. Product generation testing generation will depend upon completion of IDPS-independent operational software that is compatible with external data product formats, and availability of test data sets to reasonably exercise algorithms and software logic. Full Mission Capability Testing Objective is to continuously exercise interfaces, systems and software in NRT mode using realistic simulated data. Support for functional tests, instrument data flows and mission simulations as defined by Project test schedule.

Agenda Introduction Design Overview Operational Scenarios Schedule and Resources Issues and Conclusion

Issues Operational code versions to run in the PEATEs ADS/CLASS Porting efforts by IPO CalVal, Direct Readout Lab, PEATEs Consistency of ported versions – data formats, O/S, compilers, runtime interfaces ADS/CLASS Archived versions of xDRs (e.g., aggregations) Bandwidth for bulk data transfers Reliability and stability of external data product formats and documentation Availability of useful test data sets Radiometric and geometric fidelity for testing algorithms Instrument format fidelity for RDRs (compression, housekeeping, calibration) Ongoing simulations to provide reasonable level of operational testing Schedule and plans for interface and mission tests Objectives vs. test data sets Instrument test data flows – before and after S/C integration, duration, etc. Current test schedule w.r.t. launch date Processes and mechanisms for getting algorithm changes approved and into the IDPS operational code.

Conclusion Ocean PEATE requirements will be supported using the proven capabilities of the ODPS, which will support EDR evaluation strategies successfully employed on current missions. ODPS has well-established processes for adding the data acquisition, management, processing and distribution tasks needed by the Ocean PEATE. Additional capacity (roughly equivalent to combined Terra and Aqua MODIS) will be readily supported by technology refresh by readiness date of L-1 year. Additional development effort (Level-3 products) leverages existing software. Evaluation methodologies and tools are already established for data sets cataloged within the ODPS. Main issues pertain to IDPS software porting effort and system-level testing.

Backup Slides

Components and Subsystems: RDBMS Goal: Isolate RDBMS from system software To use a different RDBMS vendor, swap in a new Database Services Layer RDBMS Vendor Library Module Vendor Client Library Database Services Layer Perl DBI Module C Interface Functions Perl Scripts C Programs

Components and Subsystems: RDBMS Generic Core Databases Admin Catalogue Dataflow Processing MODIS Aqua MODIS Terra OCTS SeaWiFS Aquarius Mission-Specific Databases

Subsystems: Scheduler C program with supporting database procedures Runs in a daemon-like state Primary system element responsible for coordinating most of the system activity Monitors task records in a to-do list database table Runs tasks according to defined task attributes Standard job-shell interface allows new programs to be quickly adapted for Scheduler control

Subsystems: Scheduler Daily Tasks To-do List Status updated Tasks for the current day Operating System Environment Task Shell Daily Task Scheduler Scheduler

Subsystems: VDC Highly scalable, distributed infrastructure for concurrent processing of serial streams (e.g. L0 -> L1A -> L1B -> L2) Suite of C programs with supporting database procedures Uses recipes to encapsulate data-specific processing schemes, parameters, and pre-processing rules Virtual Processing Units (VPUs) serve as distinct distributed processing resources VPUs dynamically allocated based on available time and current OS load Comprehensive, user-defined processing priorities

VDC: Ancillary Data Stager Runs in a daemon-like state Monitors entries in the processing queue and runs the ancillary-select procedure for each entry's recipe Updates queue-entry status when ancillary data are available Governed by currently configured processing priorities

VDC: MakeVDC Selects processing-queue entries that have met pre-processing requirements Generate VDC job files according to configured priorities Runs as a Scheduler task, so it can easily be configured to run as often as needed to keep the VDC queue full

VDC: Engine Runs in a daemon-like state on each processing server Each instance of the VDC Engine actively competes for jobs that it is allowed to run, based on priority and length of time in the queue Monitors and manages processing resources Initializes processing streams Invokes recipe steps and monitors step-execution time Handles operator-requested stream actions

VDC: Diagram VDC Engine Processing Queue Streams Ancillary Recipe- Status updated Status updated Streams Status updated Status updated Ancillary Data Stager Recipe- step Shell VDC Queue MakeVDC VDC Engine Job file Deletes claimed job entry Operating System Environment Instantiated on each processing server VDC Inputs Input files Recipes Recipe Steps

Subsystems: Archive Device Manager (ADM) User defines logical pools of storage devices Processes request a device in a specific pool ADM returns information for a storage device in the requested pool If auto cycling is enabled, the ADM time-stamps the record for the selected device, so a different device within the pool will be selected for the next request Disk-monitor process polls all devices periodically to record usage statistics and invoke threshold handlers

Subsystems: Data Acquisition and Ingest Data types and sources are described in the database Active, passive, and periodic notification methods Active method scans remote systems for new files and populates the ingest queue Passive method waits for arrival of E-mail messages describing type and location of new file and populates the ingest queue Periodic method schedules transfers of files at user- specified intervals

Subsystems: Data Acquisition and Ingest File transfers handled by ingest daemons and Scheduler tasks FTP, RCP, and SCP transfer protocols supported Generic script handles the file transfer and then hands off to data-specific post-ingest scripts

Subsystems: Data Acquisition and Ingest Types and Sources Processing Queue Product Catalog Scheduler Transfer Manager Transfer Task Generic File- Transfer Script Post- Transfer Script Site Scanner Transfer Queue Transfer Status Records for new files Dedicated Transfer Manager Polling for new files Archive Server Remote Data Source

Subsystems: Level-3 Scheduler Responsible for scheduling processing for level-3 composite products Runs as a Scheduler task Configuration is database driven Mission-specific stored procedures are invoked to identify input files for a composite product

Subsystems: File Migration and Management Responsible for compressing files and migrating them to their various destinations Event- or time-based actions Queries associated with each action are run periodically by a Scheduler task to select files that are eligible for some type of migratory action and populate a migration queue Command-line queuing for file removal and delayed copies Migration daemons query the migration queue, perform registered actions on the files, and update catalog tables

Subsystems: File Migration and Management Migrate Manager Migration Events Migration Queue Status updated Product Catalog Migrate Engine Location and status updated

Subsystems: Data Distribution Interactive, web-based Data Ordering System, currently supporting SeaWiFS and MODIS Aqua Data Subscription System, currently supporting MODIS Aqua, allows users to define region and products of interest Order and Subscription Manager Daemons monitor the order and subscription queues and stage files on FTP servers (stage rate ~12 GBs / hr) Near-real-time data extraction and image support Web-CGI applications that allow users to view and update their orders and subscriptions

Subsystems: Data Distribution Orders Order Manager Local Distribution Servers U s e r U s e r Data Sub- scriptions Subscription Manager Extraction and Mapping Recipe Regional Extraction and Map Requests Data and images optionally pushed to users

Operational SeaWiFS Data Flow Software process Hardware system Database table Ancillary input data GPS Elements SeaWiFS L1 table L0-L1 (SWl01) SeaWiFS L0 table seawifs2 Wallops mantaray GSFC/28 Archive - Distrib Server User Community MET MET/Ozone data are dynamically selected for each L1A granule Each ground station sends a standard text e-mail notification Ingest process Ozone SeaWiFS L2 table Project Mail Server L1A-L2 (MSl12) Ocean Color Web Server Sensor CAL Archive - Distrib Server Sensor attribs fetchmail procmail Ingest queue table Browser CGI / httpd Atm corr SeaWiFS L3-bin table L3BIN (l2bin) Selects best of three down links for ingest Select downlink Archive - Distrib Server MySQL DB Incoming Mail Directory SeaWiFS L3-map table Parses and loads DB records for each down link L3MAP (smigen) Product meta data are populated from production DB Proc SW-GAC Mail Downlink manager table Archive - Distrib Server

Operational MODIS-Aqua Data Flow (NRTPE) NASA EDOS System Software process Hardware system Database table Ancillary input data NOAA Realtime System User Community MET Full-resolution day- and nighttime via SEN (~60 GB per day) MET, Ozone, and OISST data are dynamically selected for each L1A granule Ozone Ocean Color Ingest MODISA L2 table L1B-L2 (MSl12) OISST Ocean Color Web Server Archive - Distrib Server Sensor CAL Ingest queue table L1A-L1B (MOD_PR02) Sensor attribs Browser CGI / httpd MODISA L3-bin table Atm corr L3BIN (l2bin) Archive - Distrib Server ATT/EPH Ingest process Archive - Distrib Server MySQL DB Geo-Location (MOD_PR03) MODISA atteph table MODISA L3-map table L3MAP (smigen) Product meta data are populated from production DB Archive - Distrib Server L0 Ingest process Archive - Distrib Server Archive - Distrib Server L0-L1A (MOD_PR01) MODISA L0 table MODISA L1 table

SCHEDMON Used to monitor and control the Scheduler activity April 22, 2017

VDCMON Used to monitor and control the VDC activity