Tim CornwellEVLA System PDR Dec 5-6, 2001 1 EVLA: Data Management EVLA sub-contracts data management to NRAO Data Management group –End-to-end processing.

Slides:



Advertisements
Similar presentations
National Radio Astronomy Observatory June 13/14, 2005 EVLA Phase II Proposal Review EVLA Phase II Computing Development Bryan Butler (EVLA System Engineer.
Advertisements

Distributed Processing of Future Radio Astronomical Observations Ger van Diepen ASTRON, Dwingeloo ATNF, Sydney.
Jim UlvestadEVLA Mid-Project Review May 11-12, EVLA Operations Jim Ulvestad Array Science Center Concept (Requirements) Staffing and Cost Plans.
C. ChandlerEVLA Advisory Committee Meeting September 6-7, Scientific Commissioning Plan Claire Chandler.
Hunt for Molecules, Paris, 2005-Sep-20 Software Development for ALMA Robert LUCAS IRAM Grenoble France.
ADASS XI Sept30-Oct3, 2001 The ALMA Common Software (ACS) as a basis for a distributed software development G.Raffi, G.Chiozzi (ESO), B.Glendenning (NRAO)
National Radio Astronomy Observatory May 17, 2006 – Legacy Projects Workshop VLA/VLBA Large Projects Jim Ulvestad Assistant Director, NRAO.
EVLA Computing Overview Gareth Hunt EVLA Advisory Committee 2002 June
EVLA Early Science: Shared Risk Observing EVLA Advisory Committee Meeting, March 19-20, 2009 Claire Chandler Deputy AD for Science, NM Ops.
McMullinEVLA Advisory Committee Meeting December 14-15, 2004 EVLA Data Post-processing: SSG (AIPS++/CASA) Development J. McMullin.
S.T. MyersEVLA Advisory Committee Meeting September 6-7, 2007 EVLA Algorithm Research & Development Steven T. Myers (NRAO) CASA Project Scientist with.
ALMA Software B.E. Glendenning (NRAO). 2 ALMA “High Frequency VLA” in Chile Presently a European/North American Project –Japan is almost certainly joining.
Mark McKinnon EVLA Advisory Committee Meeting May 8-9, Project Overview Mark McKinnon Project Manager.
The Expanded Very Large Array: Phase I Science and Technical Requirements Rick Perley NRAO - Socorro.
Atacama Large Millimeter/submillimeter Array Expanded Very Large Array Robert C. Byrd Green Bank Telescope Very Long Baseline Array Meeting Summary Richard.
Doug Tody E2E Perspective EVLA Advisory Committee Meeting December 14-15, 2004 EVLA Software E2E Perspective.
Overall Data Processing Architecture Review EVLA Monitor and Control Interfaces July , 2002EVLA Data Processing PDR Bill Sahr.
N. RadziwillEVLA NSF Mid-Project Report May 11-12, 2006 NRAO End to End (e2e) Operations Division Nicole M. Radziwill.
EVLA Transition to Science Operations: An Overview EVLA Advisory Committee Meeting, March 19-20, 2009 Bob Dickman AD, NM Operations.
Gustaaf van MoorselEVLA Advisory Committee Meeting May 8-9, 2006 EVLA Computing Software Overview.
Peter NapierEVLA Correlator CoDR Nov 2, EVLA Correlator CoDR P. Napier Overall EVLA Project Plan.
EVLA Software Bryan Butler. 2007May22EVLA SAGE Meeting2 Requirements and Goals of EVLA Software Maximize scientific throughput of the instrument At a.
2007Sep06 EAC Butler - Software Overview 1 Software Overview Bryan Butler.
Sanjay BhatnagarEVLA Advisory Committee Meeting May 8-9, EVLA Algorithm Research & Development Progress & Plans Sanjay Bhatnagar CASA/EVLA.
1 KFPA Critical Design Review – Fri., Jan. 30, 2009 KFPA Data Pipeline Bob Garwood- NRAO-CV.
SAGE meeting Socorro, May 22-23, 2007 EVLA Science Operations: the Array Science Center Claire Chandler NRAO/Socorro.
P. NapierEVLA Advisory Committee Meeting September 8-9, EVLA Advisory Committee Project Overview P. Napier, Project Manager Management, Schedule,
EVLA Data Processing PDR Scale of processing needs Tim Cornwell, NRAO.
December2002, Garching ALMA Computing IDR ALMA AIPS++ Audit Steven T. Myers (NRAO)
Large Area Surveys - I Large area surveys can answer fundamental questions about the distribution of gas in galaxy clusters, how gas cycles in and out.
P.NapierEVLA Advisory Comm, 14 Dec 2004 Project Overview Peter Napier, EVLA Project Manager Status 2003 Committee Response.
Introduction to EVLA Software Bryan Butler. 2006Dec05/06EVLA M&C Transition Software CDR2 EVLA Computing (Terse) History The original EVLA Phase I proposal.
Gustaaf van MoorselEVLA Advisory Committee Meeting December 14-15, 2004 EVLA Computing Organization/Development.
EVLA Computing Software Overview. Gustaaf van MoorselEVLA Advisory Committee Meeting May 8-9, Contents History Organization and staffing Staffing.
Tim CornwellEVLA Advisory May 10-11, EVLA: Data Management EVLA has sub-contracted data management to NRAO Data Management group End-to-end processing.
Mark McKinnon NSF Mid-Project Review May 11-12, Baseline Project Definition Mark McKinnon Project Manager.
Recent progress in EVLA-specific algorithms EVLA Advisory Committee Meeting, March 19-20, 2009 S. Bhatnagar and U. Rau.
EVLA Monitor & Control Software PDR E2E Interfaces: Observation Scheduling Complaints to: Boyd Waters John Benson, Barry Clark, Tim Cornwell, Rich Moeser,
N. RadziwillEVLA Advisory Committee Meeting May 8-9, 2006 NRAO End to End (e2e) Operations Division Nicole M. Radziwill.
UlvestadEVLA Advisory Committee Meeting September 6-7, Future EVLA Operations Jim Ulvestad.
EVLA Data Processing PDR E2E Data Archive System John Benson, NRAO July 18, 2002.
M.P. RupenEVLA Advisory Committee Meeting September 6-7, Correlator Test Plan Michael P. Rupen.
Mark McKinnon NSF Mid-Project Review May 11-12, EVLA Advisory Committee Reports Mark McKinnon Project Manager.
Gustaaf van MoorselEVLA Advisory Committee Meeting December 14-15, 2004 EVLA Computing End-to-end (E2e) software.
10 January 2006AAS EVLA Town Hall Meeting1 The EVLA: A North American Partnership The EVLA Project on the Web
Computing Introduction Gareth Hunt EVLA System PDR 2001 December 04.
P.NapierEVLA Advisory Committee, 10 June, EVLA ADVISORY COMMITTEE PROJECT OVERVIEW Scope Organization Schedule Budget, personnel.
Overall Data Processing Architecture EVLA timeline and priorities.
EVLA Data Processing PDR Pipeline design Tim Cornwell, NRAO.
Jim UlvestadEVLA Advisory Committee Meeting May 8-9, EVLA Operations Jim Ulvestad Array Science Center Concept Requirements Staffing and Cost Plans.
S.T.MyersEVLA Advisory Committee Meeting December 14-15, 2004 EVLA Data Post-processing Overview Steven T. Myers AIPS++ Project Scientist.
P.NapierEVLA Advisory Comm, 14 Dec 2004 Project Management Plan Peter Napier Project Organization Schedule Budget.
Bryan Butler EVLA Computing Division Head
NRAO VLA Archive Survey
EVLA Overall Software Design
EVLA Computing Software Overview.
Budget, Schedule, Contingency
EVLA Data Processing PDR
Scheduling Toolkit Observation Scheduling Boyd Waters, NRAO
VLA to EVLA Transition Plan
Gustaaf van Moorsel September 9, 2003
Software Requirements
Mark McKinnon EVLA Project Manager
Background Information P.Napier
Observatory Science Operations
EVLA Algorithm Research & Development
EVLA Operations Jim Ulvestad
Observatory Science Operations
EVLA Construction Status
Rick Perley NRAO - Socorro
Presentation transcript:

Tim CornwellEVLA System PDR Dec 5-6, EVLA: Data Management EVLA sub-contracts data management to NRAO Data Management group –End-to-end processing needs being addressed by e2e project –Data reduction needs being addressed by AIPS++ project Large data volumes, parallel processing New processing needs e.g. wide-field, high dynamic range Sub-band combination

Tim CornwellEVLA System PDR Dec 5-6, EVLA management chart EVLA Project Scientist EVLA Sys Engineer, Electronics EVLA Sys Engineer, Software EVLA Budget/ Schedule Asst Director Socorro Operations Contracts EVLA Project Manager Data Mngmnt Group Canadian Partner Mexican Partner Electronics Engineering Services Computing Scientific Staff Business Socorro Divisions P. Napier J. JacksonG. HuntE. Cole Correlator B. Carlson C. Janes G. van Moorsel F.E. - D. Mertely LO/IF - T. Cotter M/C - G. Peck FO - S. Durand - T. Baldwin M/C - B. Sahr R. Moeser K. Ryan B. Rowen e2e T. Cornwell M. Goss / J. Ulvestad R. Perley

Tim CornwellEVLA System PDR Dec 5-6, End-to-End Processing Is software development for end-to-end management proceeding satisfactorily?

1 Proposal Preparation and Submission 2 Observation Preparation Software 3 Observation Scheduling Software 4 Control and Monitor System 5 Antenna 6 Feed 7 Receiver 8 IF System 9 Local Oscillator 10 Fiber Optics Transmission System 11 Correlator 12 Image Pipeline 13 Data Archive 14 Data Post-Processing NRAO Data Management Group Canadian Partner EVLA Project Primary Data Flow Monitor and Control data EVLA data flow

Tim CornwellEVLA System PDR Dec 5-6, End-to-End project (e2e) End-to-End processing for all NRAO telescopes –Improve accessibility and usability of NRAO telescopes (VLA/VLBA, GBT, EVLA) –Build on and consolidate existing resources as much as possible e.g. AIPS++ Development costs shared across NRAO –DM: project manager, project architect –Basic research: project scientist –Active construction projects: EVLA and ALMA –Sites (VLA/VLBA, GBT) and projects (AIPS++) Funding –Use internal contracts with EVLA, ALMA, GBT, VLA/VLBA –New collaborations: NVO, mini-COBRA –Have ~ 65 FTE-years Progress –Officially started July 1 –Project book ( –Start slowly: entering phase 1 development: interim VLA archive and pipeline

Tim CornwellEVLA System PDR Dec 5-6, Development Current staff –Tim Cornwell, Boyd Waters, John Benson –Job hiring in progress for C++/Java software engineer –2 Pipeline developers soon (funded by ALMA) –Expect 6 – 8 developers by middle of 2002 Use spiral development model –Five year development plan –Develop in 9 month cycles –Get requirements, plan, design, implement, test –Review requirements, plan, design, implement, test….. –Add new staff incrementally First iteration: work on core of e2e –Interim VLA archive: get all VLA export tapes on line, investigate various archiving issues –Interim VLA pipeline: process some data from archive –Start initial development of scripting for observing and pipeline setup –Calibration source unification for VLA and VLBA

Tim CornwellEVLA System PDR Dec 5-6, Development Extensive discussion of scientific requirements with Scientific Working Group –Captured in e2e project book Description of workflow from proposal to observing script –Converted to high level architecture and data flow Proceeding on basis of current requirements Revisit after ~ 9 months development of prototypes

e2e Architectural Diagrams

Tim CornwellEVLA System PDR Dec 5-6, Overall e2e architecture

Tim CornwellEVLA System PDR Dec 5-6, Operational model Describes/prescribes operation of NRAO telescopes –Currently based on VLA/VLBA operational model –Will extend and make consistent with GBT –Yet to be agreed with telescope directors Covers –Proposal submission and management –Observing scripts –Scheduling of observations –Calibration and imaging –Interactive observing –Pipeline processing –Archive use –Quality assessment –Final products

Tim CornwellEVLA System PDR Dec 5-6, Interfaces to EVLA M&C Observing scripts: –Observing blocks ~ 20min duration Observed data: –Data in ~ AIPS++ MeasurementSets, one per observing block –Sent to archive by M&C –Evaluation by pipeline Calibration information: –e.g. antenna gains, baselines, etc.

Tim CornwellEVLA System PDR Dec 5-6, Resources ALMA numbers estimated by ALMA computing management Seem to be in line with other ground based projects e2e numbers based upon straw man designs, reuse e2e scope will be adjusted to fit resources (~ 65 FTE-years) Neither constitute a detailed bottom-up derivation of resources from requirements

From NRAO to the National Virtual Observatory e2e Data Services ReferralsImages, catalogs Q A Produce images and catalogs from well-documented pipeline processing Images and catalogs “sent” to NVO All radio data stays within NRAO Other wavebands have similar relationships to NVO

Tim CornwellEVLA System PDR Dec 5-6, Data processing Scale of processing: can it be handled by 2009-era hardware?

Tim CornwellEVLA System PDR Dec 5-6, The numbers Peak data rate ~ 25 MB/s Data for Peak 8-hr observation ~ 700GB Floating point operations per float ~ Peak compute rate ~ 5Tflop Average/Peak computing load ~ 0.1 Average compute rate ~ 0.5Tflop Turnaround for 8-hr peak observation ~ 40 minutes Average/Peak data volume ~ 0.1 Data for Average 8-hr observation ~ 70GB Data for Average 1-yr ~ 80TB

Tim CornwellEVLA System PDR Dec 5-6, Detailed analysis Analyze processing in terms of FFT and Gridding costs Find scaling laws for various types of processing Express in terms of 450MHz Pentium III with Ultra- SCSI disk Use Moore’s Law to scale to e.g –Performance/cost doubles every 18 months Many more details in EVLA Memo 24

Tim CornwellEVLA System PDR Dec 5-6, Detailed analysis

Tim CornwellEVLA System PDR Dec 5-6, Scale of processing Assume Moore’s Law holds to 2009 –Moore himself believes this…….. Cost of computing for EVLA –~ 10 – 20 processor parallel machine ~ $100K - $200K (2009) –Archive ~ 50TB per year ~ $50K - $100K (2009) Comparable to computing cost for ALMA Software costs –AIPS++ as-is can do much of the processing –Development needed for high-end, pipelined processing –Some scientific/algorithmic work e.g. achieving full sensitivity, high dynamic range

Tim CornwellEVLA System PDR Dec 5-6, Desktops vs. servers Moore’s Law gives ~ 64 fold increase for a desktop –I.e. $nK where n ~ 1-3 Many projects do-able on (2009-era) desktop –e.g km/s velocity range of HI for galaxy –e.g. Mosaic of SGRA West in all H recombination lines between 28 and 41 GHz Larger projects may require parallel machine or many days on a desktop –e.g. Full sensitivity continuum image of full resolution 20cm field –NRAO would provide access over the net

Tim CornwellEVLA System PDR Dec 5-6, Data processing Is software development for data processing proceeding satisfactorily?

Tim CornwellEVLA System PDR Dec 5-6, General AIPS++ performance –Performance standards for AIPS++: Must be comparable to other disk-based packages If not, filed and handled as a high-severity defect –Analysis of existing performance defects: No inherent design-related problems found so far Cases of poor performance have been invariably due to drift as part of regular code evolution –Current approach to performance issues: Have existing correctness tests which are run regularly Building separate performance benchmark suite Will run routinely to inter-compare AIPS++ and other packages, and catch performance drift early Performance benchmarks will cover a wide range of problem sizes and types Have a separate high-performance computing group within AIPS++

Tim CornwellEVLA System PDR Dec 5-6, AIPS++ high-performance computing group –Joint initiative with the National Center for Supercomputing Applications (NCSA) in Urbana-Champaign, as part of the broader NCSA Alliance program –Separately funded by an NSF grant –Objectives: Address computationally challenging problems in radio astronomy which require supercomputer resources Provide an AIPS++ infrastructure to integrate support for HPC applications Provide portable solutions on common supercomputer architectures and Linux clusters Build expertise in HPC issues such as parallel I/O, profiling and algorithm optimization

Tim CornwellEVLA System PDR Dec 5-6, 2001 Example: parallelized wide-field VLA imaging VLA observations of the Coma cluster (test data courtesy Perley et al.) 225 imaging facets, 32 processors, speed-up factor ~20 to a net 10 hours elapsed time

Tim CornwellEVLA System PDR Dec 5-6, AIPS++ pipeline development Pipelines in AIPS++: A key requirement across the consortium and affiliates Prototypes (ATCA, BIMA) or full systems (ACSIS, Parkes multi-beam) underway Design effort within AIPS++ and with other projects (e.g. ALMA) VLA prototype pipeline: Under development as part of the first e2e prototype Based on the 2 TB VLA disk archive to be deployed soon Have purchased a pipeline server (4-processor Linux IBM x370 system) for the prototype pipeline system Early version will be confined to very restricted VLA observing modes (likely continuum) Prototype will test prototype pipeline design, implementation and performance issues on a short time-scale (Spring 2002) Vital feedback for more complete pipeline design and development work for the VLA/EVLA

Tim CornwellEVLA System PDR Dec 5-6, Post processing Mostly well-understood and in place –AIPS++ package: can reduce VLA data end-to-end EVLA-specific areas requiring more development –Very high dynamic range –Achieving full continuum sensitivity at 1.4 GHz and below Asymmetric primary beams –RFI mitigation ATNF post-correlation scheme Masking, passive and active –Very large data volumes