Page 1LWS Teams Day JSOC Overview HMI-AIA Joint Science Operations Center Science Data Processing a.k.a. JSOC-SDP Overview.

Slides:



Advertisements
Similar presentations
MAP REDUCE PROGRAMMING Dr G Sudha Sadasivam. Map - reduce sort/merge based distributed processing Best for batch- oriented processing Sort/merge is primitive.
Advertisements

HMI Data Analysis Software Plan for Phase-D. JSOC - HMI Pipeline HMI Data Analysis Pipeline Doppler Velocity Heliographic Doppler velocity maps Tracked.
ESDO Algorithms, Visualization and Data Access Elizabeth Auden 27 September 2006 HELAS Workshop Nice.
Page 1JSOC for SDO MOR October 2007 JSOC Status Review.
Rasmus Munk Larsen / Pipeline Processing 1 JSOC Pipeline Processing Overview Rasmus Munk Larsen, Stanford University
HMI - Page 1HMI Team Meeting – Jan 26, 2005 JSOC Implementation HMI Team Meeting 26 Jan 2005 Jim Aloise System Programmer
Page 1JSOC Peer Review – AIA Science Center – March 17, 2005 AIA Science Data Processing Infrastructure Neal Hurlburt AIA Data Scientist
Page 1HMI/AIA Joint Sciemce Meeting – AIA Science Center – Feb 2006 AIA Visualization Center N. Hurlburt.
200Mm Time-Distance Pipeline Status 11/5/ Mm Time-Distance Lead – Tom Duvall Lead – Tom Duvall Task – From Dopplergrams generate deep-focus synoptic.
JSOC Overview - 1HMI/AIA Team Meeting – Feb 2006 JSOC Summary- Phil Scherrer Pipeline Processing - Rasmus Larsen Data Access - Rick Bogart Data Visualization.
Rasmus Munk Larsen / Pipeline Processing 1HMI Science Team Meeting – January, 2005 JSOC Pipeline Processing Environment Rasmus Munk Larsen, Stanford University.
Page 1LWS Teams Day JSOC Overview HMI Data Products: Plan and Status.
HMI – Synoptic Data Sets HMI Team Meeting Jan. 26, 2005 Stanford, CA.
HMI/AIA Science Team Meeting, HMI Science Goals Alexander Kosovichev & HMI Team.
SUMS Storage Requirement 250 TB fixed disk cache 130 TB annual increment for permanently on- line data 100 TB work area (not controlled by SUMS) 2 PB near-line.
1 JSOC-SDP Web access to SDO HMI and AIA data The following pages show an excursion through the SDO JSOC-SDP web pages to provide an example of how to.
Building a Framework for Data Preservation of Large-Scale Astronomical Data ADASS London, UK September 23-26, 2007 Jeffrey Kantor (LSST Corporation), Ray.
Time-Distance Pipeline for the Upper 30Mm Convection Zone Status 11/5/2007.
HMI Science Analysis. Primary Goal: Origins of Solar Variability The primary goal of the Helioseismic and Magnetic Imager (HMI) investigation is to study.
Investigation Overview, Scherrer, Page 1SDO Science Writers Workshop – 16 Dec 2009 HMI Investigation Overview Philip Scherrer HMI Principal Investigator.
Page 1JSOC Review – 17 March 2005 DRMS Core System Karen Tian
25 September 2007eSDO and the VO, ADASS 2007Elizabeth Auden Accessing eSDO Solar Image Processing and Visualisation through AstroGrid Elizabeth Auden ADASS.
HMI Magnetic Products & Pipeline
Advanced Technology Center 1 HMI Rasmus Larsen / Processing Modules Stanford University HMI Team Meeting – May 2003 Processing Module Development Rasmus.
HMI Science Objectives Convection-zone dynamics and the solar dynamo  Structure and dynamics of the tachocline  Variations in differential rotation 
Copyright © 2011 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Chapter 2 Overview of Database Languages and Architectures.
Science Data Products – HMI Magnetic Field Images Pipeline 45-second Magnetic line-of-sight velocity on full disk Continuum intensity on full disk Vlos.
ESDO Algorithms, Visualization and Data Access Elizabeth Auden 21 September 2006 AHM 2006 Nottingham.
F. I. Suárez-Sol á 1, E. González-Suárez 1, I. González-Hernández 1, A.R. Davey 2,J. Hourcl é 3, VSO Team 1 National Solar Observatory, Tucson AZ – 2 Harvard-Smithsonian.
Input/OUTPUT [I/O Module structure].
MASSACHUSETTS INSTITUTE OF TECHNOLOGY NASA GODDARD SPACE FLIGHT CENTER ORBITAL SCIENCES CORPORATION NASA AMES RESEARCH CENTER SPACE TELESCOPE SCIENCE INSTITUTE.
HBase A column-centered database 1. Overview An Apache project Influenced by Google’s BigTable Built on Hadoop ▫A distributed file system ▫Supports Map-Reduce.
GONG data and pipelines: Present & future. Present data products 800x800 full-disk images, one per minute, continuous (0.87 average duty cycle) Observables:
AIA Core-Team Meeting April 2009 JSOC Stuff Phil Scherrer.
Page 1GS CDR May 2005 JSOC SDP Agenda Significant Level 4 requirements SDP Architecture Decomposition –Datacapture System –Datacapture System Components.
Computer Science Lecture 19, page 1 CS677: Distributed OS Last Class: Fault tolerance Reliable communication –One-one communication –One-many communication.
Page 1JSOC Peer Review 17Mar2005 HMI & AIA JSOC Architecture Science Team Forecast Centers EPO Public Catalog Primary Archive HMI & AIA Operations House-
HMI Major Science Objectives The primary goal of the Helioseismic and Magnetic Imager (HMI) investigation is to study the origin of solar variability and.
Page 1JSOC Peer Review 17Mar2005 HMI & AIA JSOC Architecture Science Team Forecast Centers EPO Public Catalog Primary Archive HMI & AIA Operations House-
Page 1JSOC Overview August 2007 HMI Status HMI is virtually done. –Virtually  similar to but not in fact –Front window issue resolved, flight window now.
UK E-Science Initiative and its Application to SDO J.L. Culhane MSSL.
Algorithm Preparation and Data Availability 1. Mullard Space Science Laboratory, University College London. 2. Physics and Astronomy Department, University.
Event and Feature Catalogs in the Virtual Solar Observatory Joseph A. Hourclé and the VSO Team SP54A-07 : 2008 May 30.
Heliophysics Projects Division SDO Flight Operations Readiness Review (FORR) July 21-22, 2009 SOC Operations Readiness HMI-AIA JSOC Science Data.
Page 1SDO Teams Meeting, March 2008 Status and Overview of HMI–AIA Joint Science Operations Center (JSOC) Science Data Processing (SDP) P. Scherrer Science.
Page 1JSOC Review – 17 March 2005 JSOC Pipeline Processing: Data organization, Infrastructure, and Data Products Rasmus Munk Larsen, Stanford University.
8:30-9:10 AM: Philip Scherrer, What Can We Hope to Learn from SDO Overview of SDO HMI Investigation HMI Instrument.
Simulations of Solar Convection Zone Nagi N. Mansour.
Page 1JSOC Review – 17 March 2005 JSOC 17 March 2005 “Peer” Review Stanford SDP Development Plan Philip Scherrer
Simulation Production System Science Advisory Committee Meeting UW-Madison March 1 st -2 nd 2007 Juan Carlos Díaz Vélez.
ESDO: UK Access to the Solar Dynamics Observatory 1. Mullard Space Science Laboratory, University College London. Holmbury St. Mary, Dorking, Surrey SM2.
Computer Science Lecture 19, page 1 CS677: Distributed OS Last Class: Fault tolerance Reliable communication –One-one communication –One-many communication.
ISC321 Database Systems I Chapter 2: Overview of Database Languages and Architectures Fall 2015 Dr. Abdullah Almutairi.
Helioseismology for HMI Science objectives and tasks* Data analysis plan* Helioseismology working groups and meetings *HMI Concept Study Report, Appendix.
Managing Multi-User Databases
JSOC Status at SDO PSR-1 January 21-22, 2009 GSFC
E.C. Auden1, J.L. Culhane1, Y. P. Elsworth2, A. Fludra3, M. Thompson4
Database System Concepts and Architecture
Existing Perl/Oracle Pipeline
HMI JSOC Science Data Processing Art Amezcua
The Helioseismic & Magnetic Imager – Magnetic Investigations
HMI Science Investigation Overview
JDAT Production Hardware
HMI Data Analysis Pipeline
Status and Overview of HMI–AIA Joint Science Operations Center (JSOC) Science Data Processing (SDP) April 22, 2008 P. Scherrer.
A Web-Based Data Grid Chip Watson, Ian Bird, Jie Chen,
JSOC Pipeline Processing System Components
Science Data Capture Jim Aloise Nov , 2004 System Programmer
HMI – AIA Joint Science Operations Center “Peer” Overview, 17 March AM HEPL Conference Room, Stanford Overview of JSOC Parts (Phil, 5 min.) JSOC.
HMI Data Analysis Pipeline
Presentation transcript:

Page 1LWS Teams Day JSOC Overview HMI-AIA Joint Science Operations Center Science Data Processing a.k.a. JSOC-SDP Overview

Page 2LWS Teams Day JSOC Overview HMI & AIA JSOC Concept Science Team Forecast Centers EPO Public Catalog Primary Archive HMI & AIA Operations House- keeping Database MOC SDOGS DDS Redundant Data Capture System 10-Day Archive Offsite Archiv e LMSA L Offline Archiv e HMI JSOC Pipeline Processing System Data Export & Web Service Stanford LMSAL High-Level Data Import AIA Analysis System Local Archive Quicklook Viewing GSFC White Sands World

Page 3LWS Teams Day JSOC Overview JSOC Dataflow Rates

Page 4LWS Teams Day JSOC Overview JSOC-SDP Major Components

Page 5LWS Teams Day JSOC Overview JSOC DRMS/SUMS Basic Concepts Each “image” is stored as a record in a data “series”. There will be many series: e.g. hmi_ground.lev0 is ground test data The image metadata is stored in a relational database – our Data Record Management System (DRMS) The image data is stored in SUMS (Storage Unit Management System) which itself has database tables to manage its millions of files. SUMS owns the disk and tape resources. Users interact with DRMS via a programming lauguage, e.g. c, FORTRAN, IDL. The “name” of a dataset is actually a query in a simplified DRMS naming language that also allows general SQL clauses. Efficient use of the system relies on direct use of DRMS Data may be exported from DRMS as FITS or other (TBD) protocols for remote users. Several Remote DRMS (RDRMS?) sites will be established which will “subscribe” to series of their choice. They will maintain RSUMS containing their local series and cached JSOC series. The JSOC may act as an RDRMS to access products made at remote sites.

Page 6LWS Teams Day JSOC Overview Pipeline Program, “module” Record Manage ment Keyword Access Data Access DRMS Library Link Manage ment Utility Libraries JSOC Science Libraries Record Cache JSOC Pipeline Processing System Components Database Server SUMS Storage Unit Management System DRMS Data Record Management System SUMS Tape Farm SUMS Disks PUI Pipeline User Interface Pipeline processing plan Processing script, “mapfile” List of pipeline modules with needed datasets for input, output Pipeline Operato r Processing History Log

Page 7LWS Teams Day JSOC Overview Simple example – find and look at an image P% P% show_keys "ds=hmi_ground.lev0[][ _23:50/1m]" key=FSN,T_OBS FSN T_OBS _23:50:01_UTC _23:50:06_UTC _23:50:11_UTC _23:50:16_UTC _23:50:21_UTC _23:50:26_UTC _23:50:31_UTC _23:50:36_UTC _23:50:41_UTC _23:50:46_UTC _23:50:51_UTC _23:50:56_UTC P% P% ds9 `show_keys "ds=hmi_ground.lev0[566686]" seg=file -p -q` P% Example of simple utility “module” called “show_”keys” First find images in the minute starting 9 Sept at 23:50 then look at one with ds9

Page 8LWS Teams Day JSOC Overview JSOC Export JSOC will support VSO access JSOC will also have a direct web access There will be remote DRMS/SUMS systems at key Co-I institutions ALL HMI and AIA data will be available for export at level-0 through standard products (level-1 for both and level-2 for HMI) It would be unwise to expect to export all of the data. It is simply not a reasonable thing to expect and would be a waste or resources. Our goal is to make all useful data easily accessible. This means “we” must develop browse and search tools to help generate efficient data export requests.

Page 9LWS Teams Day JSOC Overview DRMS/SUMS Configuration DRMS and SUMS use the open-source PostgreSQL database engine. DRMS will run on a pair of dedicated servers likely 4-quad-core processors with up to 10TB fast disk. SUMS will consist of file servers with attached tape systems. SUMS will manage 200TB cache disk with 150TB/year of permanent archive for level-1 and higher level products. SUMS archive will use LTO-4 tapes in a robotic system with at least 10 drives and 2000 tapes near-line. The pipeline processing system will have about 50 processor cores dedicated to level-0 to level-1 processing and about 450 cores for higher level processing in the pipeline. HMI standard products will need about half of these.

Page 10LWS Teams Day JSOC Overview Extra Info

Page 11LWS Teams Day JSOC Overview Data Record Management Service (DRMS) Data Record Management Service (DRMS) Pipeline client-server architecture JSOC Disks Analysis code C/Fortran/IDL/Matlab JSOC Library Record Cache (Keywords+Links+Data paths) OpenRecords CloseRecords GetKeyword, SetKeyword GetLink, SetLink OpenDataSegment CloseDataSegment Pipeline client process JSOC Disks Data Record Management Service (DRMS) Storage Unit Management Service (SUMS) Tape Archive Service Series Tables Record Catalogs Storage Unit Tables Record Catalogs Record Tables PostgreSQL Database Server AllocUnit GetUnit PutUnit SQL queries Storage unit transfer Data Segment I/O Storage unit transfer Generic file I/O DRMS socket protocol

Page 12LWS Teams Day JSOC Overview Pipeline batch processing A pipeline batch is encapsulated in a single database transaction, “DRMS session”: –If no module fails all data records are commited and become visible to other clients of the JSOC catalog at the end of the session –If failure occurs all data records are deleted and the database rolled back –It is possible to commit data produced up to intermediate checkpoints during sessions DRMS Service = Session Master Input data records Output data records DRMS API Register session DRMS API Module 1 DRMS API Module 2.2 … DRMS API Module N DRMS API Commit Data & Deregister Pipeline batch = atomic transaction Record & Series Database SUMS DRMS API Module 2.1

Page 13LWS Teams Day JSOC Overview HMI module status and MDI heritage Doppler Velocity Heliographic Doppler velocity maps Tracked Tiles Of Dopplergrams Stokes I,V Continuum Brightness Tracked full-disk 1-hour averaged Continuum maps Brightness feature maps Solar limb parameters Stokes I,Q,U,V Full-disk 10-min Averaged maps Tracked Tiles Line-of-sight Magnetograms Vector Magnetograms Fast algorithm Vector Magnetograms Inversion algorithm Egression and Ingression maps Time-distance Cross-covariance function Ring diagrams Wave phase shift maps Wave travel times Local wave frequency shifts Spherical Harmonic Time series Mode frequencies And splitting Brightness Images Line-of-Sight Magnetic Field Maps Coronal magnetic Field Extrapolations Coronal and Solar wind models Far-side activity index Deep-focus v and c s maps (0-200Mm) High-resolution v and c s maps (0-30Mm) Carrington synoptic v and c s maps (0-30Mm) Full-disk velocity, sound speed, Maps (0-30Mm) Internal sound speed Internal rotation Vector Magnetic Field Maps MDI pipeline modules exist Standalone production codes in use at Stanford Research codes in use by team Codes to be developed at HAO Codes being developed in the community Codes to be developed at Stanford Primary observables Intermediate and high level data products

Page 14LWS Teams Day JSOC Overview AIA Level-2

Page 15LWS Teams Day JSOC Overview JSOC Data Volumes from Proposal

Page 16LWS Teams Day JSOC Overview Sample of DRMS/SUMS Resource Assessment

Page 17LWS Teams Day JSOC Overview JSOC Level-0 processing flow

Page 18LWS Teams Day JSOC Overview JSOC-SDP Status Capture system complete, waiting testing with DDS at White Sands DRMS and SUMS virtually done Level-0 work progressing, to be done by instrument deliveries Level-1 work to start after delivery Getting serious about work on basic pipeline modules Review of status and computer hardware plans in early November

Page 19LWS Teams Day JSOC Overview Stanford JSOC effort plan