Presentation is loading. Please wait.

Presentation is loading. Please wait.

HMI JSOC Science Data Processing Art Amezcua

Similar presentations


Presentation on theme: "HMI JSOC Science Data Processing Art Amezcua"— Presentation transcript:

1 HMI JSOC Science Data Processing Art Amezcua
AGENDA JSOC-SDP Overview JSOC-SDP Status (H/W, S/W) Pipeline processing Database Level-0, Level-1 and higher levels JSOC-SDP Maintenance and CM Documentation Staffing Summary 1

2 JSOC Science Data Processing (SDP) Status
JSOC-SDP supports both HMI and AIA through Level-1 and HMI through Level-2 science data products JSO-SDP Infrastructure is complete JSOC-SDP Hardware is complete, upgrades in process Database systems – Warm standby system online in September 2009 Web Server – Upgrade online in September 2009 JSOC-SDP Software Data Record Management System and Storage Unit Management System (DRMS/ SUMS) complete as of March 2009 JSOC-SDP Archive System is fully operational Software Components needed to support commissioning Level-0 Image processing for both AIA and HMI is ready and was used to support Observatory I&T Level-0 HK, FDS and other metadata merge – complete as of May 2009 Level-1 (science observables) – will be completed during commissioning HMI Doppler and LOS Magnetic – 95% complete HMI Vector Field Observables – 90% complete AIA Level-1.5 Images – 50% complete

3 JSOC-SDP Status (continued)
Software components needed to support science mission: Production Pipeline Manager – In development, expected during commissioning HMI Level-2 (Version 1 of science data products) Local Helioseismology – Work in parallel on “rings”, “time-distance”, and “holography” proceeding with basic capability, expected to be ready during commissioning Global Helioseismology – Ready for testing during commissioning Magnetic Field standard products – Ready for testing during commissioning Vector Field disambiguation – 80% complete with preliminary product ready by end of commissioning (requires real data to proceed) Export and Catalog Browse Tools Functional but needs work ( Refinements will continue All science products need flight data during commissioning to complete development AIA Visualization Center (AVC) at Lockheed Martin Higher-level AIA processing and science product generation Heliophysics Event Knowledgebase (HEK) Stanford Summary: On schedule for L – 4 and Phase E – 6 months

4 HMI and AIA JSOC Overview
Science Team Forecast Centers EPO Public Catalog Primary Archive MOC DDS Redundant Data Capture System 19-Day Archive Offsite LMSAL Offline HMI JSOC Pipeline Processing System Data Export & Web Service JSOC-SDP High-Level Data Import AIA Analysis System Local Archive HMI & AIA Operations House- keeping Database Quicklook Viewing housekeeping GSFC White Sands World JSOC-IOC Stanford JSOC-AVC

5 JSOC-SDP Data Center Facility Components
Located in the climate-controlled basement of Physics and Astrophysics building at Stanford Important components on UPS and building back-up power; databases auto-shutdown on power outage Critical components monitored for failure (telephone, , webpage notification of issues) Components 3 data-capture machines (plus one at MOC) 1 data-processing cluster (512 CPU cores, 64 nodes, queuing system) 1 file- and tape-server machine 3 database machines 2 gateway machines (to MOC, to SDP) 1 web-server machine 2 LMSAL real-time (housekeeping) machines OC3 lines from DDS, ethernet (1 Gbps) connects all components, high-speed (20 Gbps) interconnect between data-processing cluster and file- and tape-server, 10 Gbps link between file- and tape-server and LMSAL.

6 Pipeline Cluster 512 cores
JSOC-SDP Major Components DDS GSFC Access (FDS, L0 HK) DRMS & SUMS Database Hosts Web Server OutsideWorld OC3 lines 16 10 TB 16 10 TB 16 10 TB 4 10 TB 1G Link 8 HMIDB HMIDB2 Web/Export Data Capture System ethernet (1 G) 4 13 TB 50 Slots 4 13 TB 50 Slots 4 13 TB 50 Slots Pipeline Cluster 512 cores File/Tape Server RT Mon 10G Link HMI Spare AIA 2 Quad Core X86-64 Processors 8 SPARC 0.5 TB 400 TB Plus 150 TB py 2200-Slot Tape Library Plus 12 LTO-4 Drives HMISDP-mon Local Science Workstations SPARC 0.5 TB Firewall LMSAL AIASDP-mon 20-G interconnect 10-G ethernet MOC link for real-time housekeeping X86-64 Processor Cores Disk LTO-4 Tape Library

7 Data Capture Telemetry files transmitted to data-capture (DCS) machines via two OC3 lines One line for AIA (data from four cameras over two virtual channels) One line for HMI (data from two cameras over two virtual channels) Three sets of telemetry: DCS machines archive telemetry files to tape, driven to LMSAL twice a week and stored in cabinet Production processes on dedicated cluster node ingest raw telemetry from DCS disk into hmi.tlm and aia.tlm Dedicated cluster node creates Level 0 data from telemetry and stores in DRMS/SUMS as hmi.lev0 and aia.lev0 DCS acks (once per day) DDS when offsite tape in place and verified and records in hmi.tlm created

8 Building Power w/ Generator Backup
Data Capture Details HMI Monitoring AIA Monitoring DDS HMI AIA OC3 MOC link Switches C I S O Y T E M C I S O Y T E M Pipeline Processing HMI DCS Spare DCS AIA DCS I S S Y T S ethernet Back End C I S O Y T E M 10gig private link Spare DCS C I S O Y T E M Tape Archive Disk array 13 TB (19 days) Tape Robot Operator hand net UPS UPS UPS Building Power w/ Generator Backup LMSAL MOC

9 Data Capture Status Data capture machines Online as of January 2008
Each is capable of caching 19 days telemetry Tape drive is used to generate offsite copy stored at LMSAL Pipeline system is used to generate tlm copy and Level-0 data series All tapes are LTO-4 (800 GB)

10 JSOC-SDP Pipeline Hardware
Data-processing cluster, file- and tape-server machine, and T950 tape library [COMPLETE as of July 2008]

11 Data-Processing Pipeline Status
All machines fully tested for integrity with simulated data at realistic data rate and volumes All machines/components under service warranties with vendors Database machines have been online for four years (during DRMS development) Data-processing cluster, file- and tape-server machine, T950 tape robot, and tape systems went online in July 2008 Upgrades (new machines onsite) MOC web-access machine solarport (gateway to SDP) ftp server web server two database machines in service in September 2009

12 Pipeline Software – DRMS/SUMS
Data Series Related images and metadata stored in “data series” Rows are data records (e.g., one record per time step) Columns are keywords, pointers to data files (in SUMS), pointers to other data series Storage Unit Management System (SUMS) Image files (e.g., FITS files) stored in SUMS Uses PostgreSQL database Sole client is DRMS Data Record Management System (DRMS) Data series minus image files Implemented as C library which wraps PostgreSQL database Has a FORTRAN interface Scientist interact directly with DRMS NetDRMS Network of DRMS/SUMS sites that share DRMS/SUMS data DRMS data shared via RemoteDRMS, which uses Slony-1 to make data logs that are ingested at remote site Data files residing in SUMS shared via RemoteSUMS, which uses scp; integrates with VSO so that data are obtained from least-congested NetDRMS sites

13 Database Configuration

14 image.fits image_sm.fits image.png image.fits image_sm.fits image.png
Level-0 Processing VC02*.tlm VC02*.qac VC05*.tlm VC05*.qac NFS from dcs1 VC01*.tlm VC01*.qac VC04*.tlm VC04*.qac NFS from dcs0 ingest_lev0 VC02 cl1n001 HMI VC05 hmi.tlm filename | SUDIR *.tlm *.qac SUMS hmi.lev0 fsn | lev0 keys | SUDIR image.fits image_sm.fits image.png VC01 cl1n001 AIA VC04 aia.tlm *.tlm *.qac aia.lev0 image.fits image_sm.fits image.png AIA ~24 images per VC01 or VC04 .tlm file. HMI ~16 images per VC02 or VC05 .tlm file. Reconstructs images from tlm; no modification of CCD pixels [COMPLETE as of August 2008]

15 Level-1 Processing COMPLETE as of September 2009 fetch
level-0 keywords & segments get readout mode corrections read flatfield arguments interpolate predicted orbit vectors interpolate spacecraft pointing vectors hmi.flatfield fetch level-0 image remove overscan rows & cols correct for gain & offset sdo.fds_orbit_vectors sdo.lev0_asd_003 ID bad pixels calculate image center set quality hmi.lev0 ancillary-data input ancillary-data processing image-data input image-data processing hmi.lev1 COMPLETE as of September 2009

16 HMI Observables – “Level-1” Products

17 HMI Higher-Level Processing Status
Higher-Level Science Products Internal Rotation Ω(r,Θ) [estimated complete as of August 2009] Internal Sound Speed cs(r,Θ) [estimated complete as of August 2009] Full-Disk Velocity v(r,Θ,Φ) [estimated complete as of December 2009] Sound Speed cs(r,Θ,Φ) [estimated complete as of December 2009] Carrington Synoptic Velocity Maps [estimated complete as of December 2009] Carrington Synoptic Speed Maps [estimated complete as of December 2009] High-Resolution Velocity Maps [estimated complete as of December 2009] High-Resolution Speed Maps [estimated complete as of December 2009] Deep Focus Maps [estimated complete as of July 2010] Far-Side Activity Maps [estimated complete as of December 2009] Line-of-Sight Magnetic Field Maps [COMPLETE as of July 2009] Vector field inversion and direction disambiguation [estimated complete as of March 2010] Vector Magnetic Field Maps [estimated complete as of April 2010] Coronal Magnetic Field Extrapolations [COMPLETE as of July 2009] Coronal and Solar Wind Models [estimated complete as of April 2010] Brightness Images [estimated complete as of August 2009]

18 Data Distribution and Export
Scope AIA – Level-0 and Level-1 data HMI – Level-0 data through Level-2 data Web Export Query for desired data, then download via web Supports several data formats (internal files, FITS files, tar files, compressed files) Provides support for special processing (such as extracting regions) Other developers can expand on this export method by writing javascript that is allowed to access our web cgi programs Functional now; enhancements estimated complete as of August 2009 NetDRMS Network of DRMS sites Can share DRMS data (not just data files) among sites using RemoteDRMS and RemoteSUMS Scientists can request the same data from one of many sites Virtual Solar Observatory (VSO) Integration Provides UI that allows uniform search of disparate types of data Obtains metadata and data files from NetDRMS sites experiencing the least congestion Estimated complete as of December 2009

19 Maintenance and Expansion During Mission
Hardware Each hardware component is covered under a vendor service plan; as plans expire, they are renewed Planned phased replacement/upgrades throughout Phase-E Software Lead software developers are part of continuing team for Phase-E Storage File Server – 150 TB per year disk Tape Library – filled tapes stored in Data Center, replaced with new tapes as needed; library expansion entails new 1300-slot cabinet when needed Functionality Anticipate continued development of science processing and distribution tools during Phase-E

20 Documentation Wiki - http://jsoc.stanford.edu/jsocwiki
Overview of JSOC Data Series, DRMS, and SUMS JSOC Series Definition Files DRMS Names JSOC User’s Guide JSOC Developer’s Guide Doxygen - Manual describing DRMS API functions and modules Provides synopsis and describes input parameters, output, return values To date, ~ 1/2 of functions/modules have documentation Flow Diagrams Tree diagrams illustrating connections between various programs, data series, data tables, etc. Diagrammatic view of pipeline processes links to documentation Note stages of development (A – E) and estimated completion date Procedures documented Database maintenance DCS operations Level-0 processing management RemoteDRMS/SUMS installation and maintenance Procedure documentation in progress Calibration Processing (filters, flat fields, etc.) Pipeline Dataflow Management System Export management Weekly data product report generation

21 Summary The JSOC-SDP can support:
Archive and distribution functions now Analysis for instrument commissioning now Initial science data processing by launch

22 Backup Slides

23 JSOC-SDP Stages of Software Development
Stage A – Code specification exists, but working code does not exist Stage B – Prototype code exists, but not necessarily on HMI data and not necessarily in the correct language Stage C – Working code exists but cannot run inside JSOC pipeline Stage D – Working code capable of running in JSOC pipeline, but undergoing final testing and not released for general use Stage E – Working code complete and integrated into JSOC pipeline Following dataflow charts show status as of FORR with estimated months to complete to stage E shown

24 JSOC-SDP Dataflow – Data Capture

25 JSOC-SDP Dataflow HMI Level-0 and Level-1

26 JSOC-SDP Dataflow HMI Level-1 – Detail

27 JSOC-SDP Dataflow HMI Level-2 Helioseismology

28 JSOC-SDP Dataflow HMI Level-2 – LOS Mag

29 JSOC-SDP Dataflow HMI Level-2 – Vector Mag

30 AVC Dataflow – Data Distribution
Minutes Days Months 30


Download ppt "HMI JSOC Science Data Processing Art Amezcua"

Similar presentations


Ads by Google