HMI JSOC Science Data Processing Art Amezcua

Slides:



Advertisements
Similar presentations
HMI Data Analysis Software Plan for Phase-D. JSOC - HMI Pipeline HMI Data Analysis Pipeline Doppler Velocity Heliographic Doppler velocity maps Tracked.
Advertisements

MUNIS Platform Migration Project WELCOME. Agenda Introductions Tyler Cloud Overview Munis New Features Questions.
HMI - Page 1HMI Team Meeting – Jan 26, 2005 JSOC Implementation HMI Team Meeting 26 Jan 2005 Jim Aloise System Programmer
Page 1JSOC Peer Review – AIA Science Center – March 17, 2005 AIA Science Data Processing Infrastructure Neal Hurlburt AIA Data Scientist
JSOC Overview - 1HMI/AIA Team Meeting – Feb 2006 JSOC Summary- Phil Scherrer Pipeline Processing - Rasmus Larsen Data Access - Rick Bogart Data Visualization.
Experimental Facilities Division ANL-ORNL SNS Experimental Data Standards (Status) Richard Riedel SNS Data Acquisition Group Leader.
Page 1JSOC Review – 17 March 2005 Database Servers Challenges A very large database on the order of a few TB -- We can't copy the whole database in real.
Page 1HMI Team Meeting – January 26, 2005 HMI Mission Operations Rock Bush HMI Stanford Program Manager Stanford University
25 September 2007eSDO and the VO, ADASS 2007Elizabeth Auden Accessing eSDO Solar Image Processing and Visualisation through AstroGrid Elizabeth Auden ADASS.
Advanced Technology Center 1 HMI Rasmus Larsen / Processing Modules Stanford University HMI Team Meeting – May 2003 Processing Module Development Rasmus.
November 2009 Network Disaster Recovery October 2014.
ESDO Algorithms, Visualization and Data Access Elizabeth Auden 21 September 2006 AHM 2006 Nottingham.
Data Processing and Display Challenges for Solar Dynamics Observatory Using the Heliophysics Event Knowledgebase Ralph Seguin Ankur Somani Lockheed Martin.
F. I. Suárez-Sol á 1, E. González-Suárez 1, I. González-Hernández 1, A.R. Davey 2,J. Hourcl é 3, VSO Team 1 National Solar Observatory, Tucson AZ – 2 Harvard-Smithsonian.
Chapter 8 Implementing Disaster Recovery and High Availability Hands-On Virtual Computing.
MASSACHUSETTS INSTITUTE OF TECHNOLOGY NASA GODDARD SPACE FLIGHT CENTER ORBITAL SCIENCES CORPORATION NASA AMES RESEARCH CENTER SPACE TELESCOPE SCIENCE INSTITUTE.
Data Management Subsystem Jeff Valenti (STScI). DMS Context PRDS - Project Reference Database PPS - Proposal and Planning OSS - Operations Scripts FOS.
GONG data and pipelines: Present & future. Present data products 800x800 full-disk images, one per minute, continuous (0.87 average duty cycle) Observables:
ASI-Eumetsat Meeting Matera, 4-5 Feb CNM Context Matera, February 4-5, 20092ASI-Eumetsat Meeting.
Module 13 Implementing Business Continuity. Module Overview Protecting and Recovering Content Working with Backup and Restore for Disaster Recovery Implementing.
Page 1LWS Teams Day JSOC Overview HMI-AIA Joint Science Operations Center Science Data Processing a.k.a. JSOC-SDP Overview.
Page 1GS CDR May 2005 JSOC SDP Agenda Significant Level 4 requirements SDP Architecture Decomposition –Datacapture System –Datacapture System Components.
CLASS Information Management Presented at NOAATECH Conference 2006 Presented by Pat Schafer (CLASS-WV Development Lead)
Page 1JSOC Peer Review 17Mar2005 HMI & AIA JSOC Architecture Science Team Forecast Centers EPO Public Catalog Primary Archive HMI & AIA Operations House-
HMI Major Science Objectives The primary goal of the Helioseismic and Magnetic Imager (HMI) investigation is to study the origin of solar variability and.
Page 1JSOC Peer Review 17Mar2005 HMI & AIA JSOC Architecture Science Team Forecast Centers EPO Public Catalog Primary Archive HMI & AIA Operations House-
Page 1JSOC Overview August 2007 HMI Status HMI is virtually done. –Virtually  similar to but not in fact –Front window issue resolved, flight window now.
06-1L ASTRO-E2 ASTRO-E2 User Group - 14 February, 2005 Astro-E2 Archive Lorella Angelini/HEASARC.
Super Computing 2000 DOE SCIENCE ON THE GRID Storage Resource Management For the Earth Science Grid Scientific Data Management Research Group NERSC, LBNL.
Heliophysics Projects Division SDO Flight Operations Readiness Review (FORR) July 21-22, 2009 SOC Operations Readiness HMI-AIA JSOC Science Data.
Page 1SDO Teams Meeting, March 2008 Status and Overview of HMI–AIA Joint Science Operations Center (JSOC) Science Data Processing (SDP) P. Scherrer Science.
1 SUZAKU HUG 12-13April, 2006 Suzaku archive Lorella Angelini/HEASARC.
ESDO: UK Access to the Solar Dynamics Observatory 1. Mullard Space Science Laboratory, University College London. Holmbury St. Mary, Dorking, Surrey SM2.
AIRS Meeting GSFC, February 1, 2002 ECS Data Pool Gregory Leptoukh.
A Solution for Maintaining File Integrity within an Online Data Archive Dan Scholes PDS Geosciences Node Washington University 1.
Data Acquisition, Diagnostics & Controls (DAQ)
Compute and Storage For the Farm at Jlab
Architecture Review 10/11/2004
RHEV Platform at LHCb Red Hat at CERN 17-18/1/17
Software Overview Sonja Vrcic
AGENDA JSOC-SDP Overview JSOC-SDP Status (H/W, S/W)
WP18, High-speed data recording Krzysztof Wrona, European XFEL
JSOC Status at SDO PSR-1 January 21-22, 2009 GSFC
E.C. Auden1, J.L. Culhane1, Y. P. Elsworth2, A. Fludra3, M. Thompson4
U.S. ATLAS Grid Production Experience
Installation, setup and use - Ferro Backup – V1.1
Existing Perl/Oracle Pipeline
Virtualization, Cloud Computing and Big Data
Chapter 10: Device Discovery, Management, and Maintenance
LQCD Computing Operations
Workshop Summary Dirk Duellmann.
Current SOC-Ground System Integration Schedule (Jan )
HMI Science Investigation Overview
JDAT Production Hardware
Chapter 10: Device Discovery, Management, and Maintenance
HMI Data Analysis Pipeline
Status and Overview of HMI–AIA Joint Science Operations Center (JSOC) Science Data Processing (SDP) April 22, 2008 P. Scherrer.
GLAST Large Area Telescope
Operational Dataset Update Functionality Included in the NCAR Research Data Archive Management System Zaihua Ji Doug Schuster Steven Worley Computational.
Intermountain West Data Warehouse
Launch and On-orbit Checkout
JSOC Pipeline Processing System Components
Cloud computing mechanisms
Click to add title Planning for LSST Verification George Angeli LSST All Hands Meeting Tucson August 15, 2016.
Science Data Capture Jim Aloise Nov , 2004 System Programmer
HMI – AIA Joint Science Operations Center “Peer” Overview, 17 March AM HEPL Conference Room, Stanford Overview of JSOC Parts (Phil, 5 min.) JSOC.
HMI Data Analysis Pipeline
Overview of Workflows: Why Use Them?
ATLAS DC2 & Continuous production
Integration & Test Instrument Operations Coordination
Presentation transcript:

HMI JSOC Science Data Processing Art Amezcua AGENDA JSOC-SDP Overview JSOC-SDP Status (H/W, S/W) Pipeline processing Database Level-0, Level-1 and higher levels JSOC-SDP Maintenance and CM Documentation Staffing Summary 1

JSOC Science Data Processing (SDP) Status JSOC-SDP supports both HMI and AIA through Level-1 and HMI through Level-2 science data products JSO-SDP Infrastructure is complete JSOC-SDP Hardware is complete, upgrades in process Database systems – Warm standby system online in September 2009 Web Server – Upgrade online in September 2009 JSOC-SDP Software Data Record Management System and Storage Unit Management System (DRMS/ SUMS) complete as of March 2009 JSOC-SDP Archive System is fully operational Software Components needed to support commissioning Level-0 Image processing for both AIA and HMI is ready and was used to support Observatory I&T Level-0 HK, FDS and other metadata merge – complete as of May 2009 Level-1 (science observables) – will be completed during commissioning HMI Doppler and LOS Magnetic – 95% complete HMI Vector Field Observables – 90% complete AIA Level-1.5 Images – 50% complete

JSOC-SDP Status (continued) Software components needed to support science mission: Production Pipeline Manager – In development, expected during commissioning HMI Level-2 (Version 1 of science data products) Local Helioseismology – Work in parallel on “rings”, “time-distance”, and “holography” proceeding with basic capability, expected to be ready during commissioning Global Helioseismology – Ready for testing during commissioning Magnetic Field standard products – Ready for testing during commissioning Vector Field disambiguation – 80% complete with preliminary product ready by end of commissioning (requires real data to proceed) Export and Catalog Browse Tools Functional but needs work (http://jsoc.stanford.edu/ajax/lookdata.html) Refinements will continue All science products need flight data during commissioning to complete development AIA Visualization Center (AVC) at Lockheed Martin Higher-level AIA processing and science product generation Heliophysics Event Knowledgebase (HEK) Stanford Summary: On schedule for L – 4 and Phase E – 6 months

HMI and AIA JSOC Overview Science Team Forecast Centers EPO Public Catalog Primary Archive MOC DDS Redundant Data Capture System 19-Day Archive Offsite LMSAL Offline HMI JSOC Pipeline Processing System Data Export & Web Service JSOC-SDP High-Level Data Import AIA Analysis System Local Archive HMI & AIA Operations House- keeping Database Quicklook Viewing housekeeping GSFC White Sands World JSOC-IOC Stanford JSOC-AVC

JSOC-SDP Data Center Facility Components Located in the climate-controlled basement of Physics and Astrophysics building at Stanford Important components on UPS and building back-up power; databases auto-shutdown on power outage Critical components monitored for failure (telephone, email, webpage notification of issues) Components 3 data-capture machines (plus one at MOC) 1 data-processing cluster (512 CPU cores, 64 nodes, queuing system) 1 file- and tape-server machine 3 database machines 2 gateway machines (to MOC, to SDP) 1 web-server machine 2 LMSAL real-time (housekeeping) machines OC3 lines from DDS, ethernet (1 Gbps) connects all components, high-speed (20 Gbps) interconnect between data-processing cluster and file- and tape-server, 10 Gbps link between file- and tape-server and LMSAL.

Pipeline Cluster 512 cores JSOC-SDP Major Components DDS GSFC Access (FDS, L0 HK) DRMS & SUMS Database Hosts Web Server OutsideWorld OC3 lines 16 10 TB 16 10 TB 16 10 TB 4 10 TB 1G Link 8 HMIDB HMIDB2 Web/Export Data Capture System ethernet (1 G) 4 13 TB 50 Slots 4 13 TB 50 Slots 4 13 TB 50 Slots Pipeline Cluster 512 cores File/Tape Server RT Mon 10G Link HMI Spare AIA 2 Quad Core X86-64 Processors 8 SPARC 0.5 TB 400 TB Plus 150 TB py 2200-Slot Tape Library Plus 12 LTO-4 Drives HMISDP-mon Local Science Workstations SPARC 0.5 TB Firewall LMSAL AIASDP-mon 20-G interconnect 10-G ethernet MOC link for real-time housekeeping X86-64 Processor Cores Disk LTO-4 Tape Library

Data Capture Telemetry files transmitted to data-capture (DCS) machines via two OC3 lines One line for AIA (data from four cameras over two virtual channels) One line for HMI (data from two cameras over two virtual channels) Three sets of telemetry: DCS machines archive telemetry files to tape, driven to LMSAL twice a week and stored in cabinet Production processes on dedicated cluster node ingest raw telemetry from DCS disk into hmi.tlm and aia.tlm Dedicated cluster node creates Level 0 data from telemetry and stores in DRMS/SUMS as hmi.lev0 and aia.lev0 DCS acks (once per day) DDS when offsite tape in place and verified and records in hmi.tlm created

Building Power w/ Generator Backup Data Capture Details HMI Monitoring AIA Monitoring DDS HMI AIA OC3 MOC link Switches C I S O Y T E M C I S O Y T E M Pipeline Processing HMI DCS Spare DCS AIA DCS I S S Y T S ethernet Back End C I S O Y T E M 10gig private link Spare DCS C I S O Y T E M Tape Archive Disk array 13 TB (19 days) Tape Robot Operator hand net UPS UPS UPS Building Power w/ Generator Backup LMSAL MOC

Data Capture Status Data capture machines Online as of January 2008 Each is capable of caching 19 days telemetry Tape drive is used to generate offsite copy stored at LMSAL Pipeline system is used to generate tlm copy and Level-0 data series All tapes are LTO-4 (800 GB)

JSOC-SDP Pipeline Hardware Data-processing cluster, file- and tape-server machine, and T950 tape library [COMPLETE as of July 2008]

Data-Processing Pipeline Status All machines fully tested for integrity with simulated data at realistic data rate and volumes All machines/components under service warranties with vendors Database machines have been online for four years (during DRMS development) Data-processing cluster, file- and tape-server machine, T950 tape robot, and tape systems went online in July 2008 Upgrades (new machines onsite) MOC web-access machine solarport (gateway to SDP) ftp server web server two database machines in service in September 2009

Pipeline Software – DRMS/SUMS Data Series Related images and metadata stored in “data series” Rows are data records (e.g., one record per time step) Columns are keywords, pointers to data files (in SUMS), pointers to other data series Storage Unit Management System (SUMS) Image files (e.g., FITS files) stored in SUMS Uses PostgreSQL database Sole client is DRMS Data Record Management System (DRMS) Data series minus image files Implemented as C library which wraps PostgreSQL database Has a FORTRAN interface Scientist interact directly with DRMS NetDRMS Network of DRMS/SUMS sites that share DRMS/SUMS data DRMS data shared via RemoteDRMS, which uses Slony-1 to make data logs that are ingested at remote site Data files residing in SUMS shared via RemoteSUMS, which uses scp; integrates with VSO so that data are obtained from least-congested NetDRMS sites

Database Configuration

image.fits image_sm.fits image.png image.fits image_sm.fits image.png Level-0 Processing VC02*.tlm VC02*.qac VC05*.tlm VC05*.qac NFS from dcs1 VC01*.tlm VC01*.qac VC04*.tlm VC04*.qac NFS from dcs0 ingest_lev0 VC02 cl1n001 HMI VC05 hmi.tlm filename | SUDIR *.tlm *.qac SUMS hmi.lev0 fsn | lev0 keys | SUDIR image.fits image_sm.fits image.png VC01 cl1n001 AIA VC04 aia.tlm *.tlm *.qac aia.lev0 image.fits image_sm.fits image.png AIA ~24 images per VC01 or VC04 .tlm file. HMI ~16 images per VC02 or VC05 .tlm file. Reconstructs images from tlm; no modification of CCD pixels [COMPLETE as of August 2008]

Level-1 Processing COMPLETE as of September 2009 fetch level-0 keywords & segments get readout mode corrections read flatfield arguments interpolate predicted orbit vectors interpolate spacecraft pointing vectors hmi.flatfield fetch level-0 image remove overscan rows & cols correct for gain & offset sdo.fds_orbit_vectors sdo.lev0_asd_003 ID bad pixels calculate image center set quality hmi.lev0 ancillary-data input ancillary-data processing image-data input image-data processing hmi.lev1 COMPLETE as of September 2009

HMI Observables – “Level-1” Products

HMI Higher-Level Processing Status Higher-Level Science Products Internal Rotation Ω(r,Θ) [estimated complete as of August 2009] Internal Sound Speed cs(r,Θ) [estimated complete as of August 2009] Full-Disk Velocity v(r,Θ,Φ) [estimated complete as of December 2009] Sound Speed cs(r,Θ,Φ) [estimated complete as of December 2009] Carrington Synoptic Velocity Maps [estimated complete as of December 2009] Carrington Synoptic Speed Maps [estimated complete as of December 2009] High-Resolution Velocity Maps [estimated complete as of December 2009] High-Resolution Speed Maps [estimated complete as of December 2009] Deep Focus Maps [estimated complete as of July 2010] Far-Side Activity Maps [estimated complete as of December 2009] Line-of-Sight Magnetic Field Maps [COMPLETE as of July 2009] Vector field inversion and direction disambiguation [estimated complete as of March 2010] Vector Magnetic Field Maps [estimated complete as of April 2010] Coronal Magnetic Field Extrapolations [COMPLETE as of July 2009] Coronal and Solar Wind Models [estimated complete as of April 2010] Brightness Images [estimated complete as of August 2009]

Data Distribution and Export Scope AIA – Level-0 and Level-1 data HMI – Level-0 data through Level-2 data Web Export http://jsoc.stanford.edu/ajax/lookdata.html Query for desired data, then download via web Supports several data formats (internal files, FITS files, tar files, compressed files) Provides support for special processing (such as extracting regions) Other developers can expand on this export method by writing javascript that is allowed to access our web cgi programs Functional now; enhancements estimated complete as of August 2009 NetDRMS Network of DRMS sites Can share DRMS data (not just data files) among sites using RemoteDRMS and RemoteSUMS Scientists can request the same data from one of many sites Virtual Solar Observatory (VSO) Integration Provides UI that allows uniform search of disparate types of data Obtains metadata and data files from NetDRMS sites experiencing the least congestion Estimated complete as of December 2009

Maintenance and Expansion During Mission Hardware Each hardware component is covered under a vendor service plan; as plans expire, they are renewed Planned phased replacement/upgrades throughout Phase-E Software Lead software developers are part of continuing team for Phase-E Storage File Server – 150 TB per year disk Tape Library – filled tapes stored in Data Center, replaced with new tapes as needed; library expansion entails new 1300-slot cabinet when needed Functionality Anticipate continued development of science processing and distribution tools during Phase-E

Documentation Wiki - http://jsoc.stanford.edu/jsocwiki Overview of JSOC Data Series, DRMS, and SUMS JSOC Series Definition Files DRMS Names JSOC User’s Guide JSOC Developer’s Guide Doxygen - http://jsoc.stanford.edu/doxygen_html Manual describing DRMS API functions and modules Provides synopsis and describes input parameters, output, return values To date, ~ 1/2 of functions/modules have documentation Flow Diagrams Tree diagrams illustrating connections between various programs, data series, data tables, etc. Diagrammatic view of pipeline processes links to documentation Note stages of development (A – E) and estimated completion date Procedures documented Database maintenance DCS operations Level-0 processing management RemoteDRMS/SUMS installation and maintenance Procedure documentation in progress Calibration Processing (filters, flat fields, etc.) Pipeline Dataflow Management System Export management Weekly data product report generation

Summary The JSOC-SDP can support: Archive and distribution functions now Analysis for instrument commissioning now Initial science data processing by launch

Backup Slides

JSOC-SDP Stages of Software Development Stage A – Code specification exists, but working code does not exist Stage B – Prototype code exists, but not necessarily on HMI data and not necessarily in the correct language Stage C – Working code exists but cannot run inside JSOC pipeline Stage D – Working code capable of running in JSOC pipeline, but undergoing final testing and not released for general use Stage E – Working code complete and integrated into JSOC pipeline Following dataflow charts show status as of FORR with estimated months to complete to stage E shown

JSOC-SDP Dataflow – Data Capture

JSOC-SDP Dataflow HMI Level-0 and Level-1

JSOC-SDP Dataflow HMI Level-1 – Detail

JSOC-SDP Dataflow HMI Level-2 Helioseismology

JSOC-SDP Dataflow HMI Level-2 – LOS Mag

JSOC-SDP Dataflow HMI Level-2 – Vector Mag

AVC Dataflow – Data Distribution Minutes Days Months 30