An Astronomical Image Mosaic Service for the National Virtual Observatory / ESTO.

Slides:



Advertisements
Similar presentations
3 September 2004NVO Coordination Meeting1 Grid-Technologies NVO and the Grid Reagan W. Moore George Kremenek Leesa Brieger Ewa Deelman Roy Williams John.
Advertisements

Aug 3-4, 2003 Victoria1 Montage, Pegasus and ROME G. B. Berriman, J.C. Good, M. Kong, A. Laity IPAC/Caltech J. Jacob, A. Bergou, D. S. Katz JPL R. Williams.
A Workflow Engine with Multi-Level Parallelism Supports Qifeng Huang and Yan Huang School of Computer Science Cardiff University
Grid Astronomy with Image Federation Roy Williams Michael Feldmann California Institute of Technology.
Pegasus on the Virtual Grid: A Case Study of Workflow Planning over Captive Resources Yang-Suk Kee, Eun-Kyu Byun, Ewa Deelman, Kran Vahi, Jin-Soo Kim Oracle.
Science Archives in the 21st Century Best Practices in Ingestion and Data Access at the NASA/IPAC Infrared Science Archive
The ADAMANT Project: Linking Scientific Workflows and Networks “Adaptive Data-Aware Multi-Domain Application Network Topologies” Ilia Baldine, Charles.
Ewa Deelman, Integrating Existing Scientific Workflow Systems: The Kepler/Pegasus Example Nandita Mangal,
AME: An Any-scale many-task computing Engine Zhao Zhang, University of Chicago Daniel S. Katz, CI University of Chicago.
Grist: Grid Data Mining for Astronomy Joseph Jacob, Daniel Katz, Craig Miller, and Harshpreet Walia (JPL) Roy Williams, Matthew Graham, and Michael Aivazis.
A Web service for Distributed Covariance Computation on Astronomy Catalogs Presented by Haimonti Dutta CMSC 691D.
Science Applications of Montage: An Astronomical Image Mosaic Engine ESTO G. Bruce Berriman, John Good and Anastasia Laity.
The Montage Image Mosaic Service: Custom Mosaics on Demand ESTO John Good, Bruce Berriman Mihseh Kong and Anastasia Laity IPAC, Caltech
Lsevlet: web-based processing tool for LiDAR data J Ramon Arrowsmith December 14, 2004.
Montage: Experiences in Astronomical Image Mosaicking on the TeraGrid ESTO.
Pegasus: Mapping complex applications onto the Grid Ewa Deelman Center for Grid Technologies USC Information Sciences Institute.
An Astronomical Image Mosaic Service for the National Virtual Observatory
A Grid-Enabled Engine for Delivering Custom Science- Grade Images on Demand
Ewa Deelman Using Grid Technologies to Support Large-Scale Astronomy Applications Ewa Deelman Center for Grid Technologies USC Information.
Astronomical Image Mosaicking ESTO.
CREATING A MULTI-WAVELENGTH GALACTIC PLANE ATLAS WITH AMAZON WEB SERVICES G. Bruce Berriman, John Good IPAC, California Institute of Technolog y Ewa Deelman,
Managing Workflows with the Pegasus Workflow Management System
TeraGrid Gateway User Concept – Supporting Users V. E. Lynch, M. L. Chen, J. W. Cobb, J. A. Kohl, S. D. Miller, S. S. Vazhkudai Oak Ridge National Laboratory.
Ewa Deelman, Pegasus and DAGMan: From Concept to Execution Mapping Scientific Workflows onto the National.
CONDOR DAGMan and Pegasus Selim Kalayci Florida International University 07/28/2009 Note: Slides are compiled from various TeraGrid Documentations.
Pegasus A Framework for Workflow Planning on the Grid Ewa Deelman USC Information Sciences Institute Pegasus Acknowledgments: Carl Kesselman, Gaurang Mehta,
Why Build Image Mosaics for Wide Area Surveys? An All-Sky 2MASS Mosaic Constructed on the TeraGrid A. C. Laity, G. B. Berriman, J. C. Good (IPAC, Caltech);
The Grid is a complex, distributed and heterogeneous execution environment. Running applications requires the knowledge of many grid services: users need.
Large-Scale Science Through Workflow Management Ewa Deelman Center for Grid Technologies USC Information Sciences Institute.
1 Dryad Distributed Data-Parallel Programs from Sequential Building Blocks Michael Isard, Mihai Budiu, Yuan Yu, Andrew Birrell, Dennis Fetterly of Microsoft.
Combining the strengths of UMIST and The Victoria University of Manchester Utility Driven Adaptive Workflow Execution Kevin Lee School of Computer Science,
 The workflow description modified to output a VDS DAX.  The workflow description toolkit developed allows any concrete workflow description to be migrated.
Pegasus: Planning for Execution in Grids Ewa Deelman Information Sciences Institute University of Southern California.
Part 8: DAGMan A: Grid Workflow Management B: DAGMan C: Laboratory: DAGMan.
Dr. Ahmed Abdeen Hamed, Ph.D. University of Vermont, EPSCoR Research on Adaptation to Climate Change (RACC) Burlington Vermont USA MODELING THE IMPACTS.
Pegasus: Mapping Scientific Workflows onto the Grid Ewa Deelman Center for Grid Technologies USC Information Sciences Institute.
Condor Week 2005Optimizing Workflows on the Grid1 Optimizing workflow execution on the Grid Gaurang Mehta - Based on “Optimizing.
Ten Years of Software Sustainability at The Infrared Processing and Analysis Center G. Bruce Berriman and John Good NASA Exoplanet Science Institute, Infrared.
PROGRESS: ICCS'2003 GRID SERVICE PROVIDER: How to improve flexibility of grid user interfaces? Michał Kosiedowski.
Combining the strengths of UMIST and The Victoria University of Manchester Adaptive Workflow Processing and Execution in Pegasus Kevin Lee School of Computer.
09/02 ID099-1 September 9, 2002Grid Technology Panel Patrick Dreher Technical Panel Discussion: Progress in Developing a Web Services Data Analysis Grid.
Pegasus: Running Large-Scale Scientific Workflows on the TeraGrid Ewa Deelman USC Information Sciences Institute
What is Triana?. GAPGAP Triana Distributed Work-flow Network Action Commands Workflow, e.g. BPEL4WS Triana Engine Triana Controlling Service (TCS) Triana.
Pegasus: Mapping complex applications onto the Grid Ewa Deelman Center for Grid Technologies USC Information Sciences Institute.
GriPhyN Virtual Data System Grid Execution of Virtual Data Workflows Mike Wilde Argonne National Laboratory Mathematics and Computer Science Division.
CEOS Data Cube Open Source Software Status Brian Killough CEOS Systems Engineering Office (SEO) WGISS-40 Harwell, Oxfordshire, UK September 30, 2015 (remote.
Virtual Observatories, Astronomical Telescopes and Instrumentation (August 2002)1 An Architecture for Access To A Compute Intensive Image Mosaic Service.
Pegasus-a framework for planning for execution in grids Karan Vahi USC Information Sciences Institute May 5 th, 2004.
TeraGrid Gateway User Concept – Supporting Users V. E. Lynch, M. L. Chen, J. W. Cobb, J. A. Kohl, S. D. Miller, S. S. Vazhkudai Oak Ridge National Laboratory.
Planning Ewa Deelman USC Information Sciences Institute GriPhyN NSF Project Review January 2003 Chicago.
Pegasus: Planning for Execution in Grids Ewa Deelman, Carl Kesselman, Gaurang Mehta, Gurmeet Singh, Karan Vahi Information Sciences Institute University.
August 2003 At A Glance The IRC is a platform independent, extensible, and adaptive framework that provides robust, interactive, and distributed control.
AHM04: Sep 2004 Nottingham CCLRC e-Science Centre eMinerals: Environment from the Molecular Level Managing simulation data Lisa Blanshard e- Science Data.
Bulk Data Transfer Activities We regard data transfers as “first class citizens,” just like computational jobs. We have transferred ~3 TB of DPOSS data.
Lecture 3 With every passing hour our solar system comes forty-three thousand miles closer to globular cluster 13 in the constellation Hercules, and still.
Origami: Scientific Distributed Workflow in McIDAS-V Maciek Smuga-Otto, Bruce Flynn (also Bob Knuteson, Ray Garcia) SSEC.
1 USC Information Sciences InstituteYolanda Gil AAAI-08 Tutorial July 13, 2008 Part IV Workflow Mapping and Execution in Pegasus (Thanks.
Managing LIGO Workflows on OSG with Pegasus Karan Vahi USC Information Sciences Institute
Resource Allocation and Scheduling for Workflows Gurmeet Singh, Carl Kesselman, Ewa Deelman.
1 Performance Impact of Resource Provisioning on Workflows Gurmeet Singh, Carl Kesselman and Ewa Deelman Information Science Institute University of Southern.
Research and Service Support Resources for EO data exploitation RSS Team, ESRIN, 23/01/2013 Requirements for a Federated Infrastructure.
Pegasus WMS Extends DAGMan to the grid world
Cloudy Skies: Astronomy and Utility Computing
Montage: An On-Demand Image Mosaic Service for the NVO
Patrick Dreher Research Scientist & Associate Director
Pegasus and Condor Gaurang Mehta, Ewa Deelman, Carl Kesselman, Karan Vahi Center For Grid Technologies USC/ISI.
Pegasus Workflows on XSEDE
Overview of Workflows: Why Use Them?
High Throughput Computing for Astronomers
Frieda meets Pegasus-WMS
Presentation transcript:

An Astronomical Image Mosaic Service for the National Virtual Observatory / ESTO

Contributors Attila Bergou - JPL Bruce Berriman - IPAC Ewa Deelman - ISI John Good - IPAC Joseph C. Jacob - JPL Daniel S. Katz - JPL Carl Kesselman - ISI Anastasia Laity - IPAC Thomas Prince - Caltech Gurmeet Singh - ISI Mei-Hui Su - ISI Roy Williams - CACR

What is Montage? Delivers custom, science grade image mosaics User specifies projection, coordinates, spatial sampling, mosaic size, image rotation Preserve astrometry & flux Modular “toolbox” design Loosely-coupled Engines for Image Reprojection, Background Removal, Co-addition Control testing and maintenance costs Flexibility; e.g., custom background algorithm; use as a reprojection and co-registration engine Implemented in ANSI C for portability Public service will be deployed on the Teragrid Order mosaics through web portal

Public Release of Montage  Version and User’s Guide available for download at Emphasizes accuracy in photometry and astrometry Images processed serially Tested and validated on 2MASS 2IDR images on Linux Red Hat 8.0 (Kernel release ) on a 32-bit processor Tested on 10 WCS projections with mosaics smaller than 2 x 2 degrees and coordinate transformations Equ J2000 to Galactic and Ecliptic Extensively tested 2,595 test cases executed 119 defects reported and 116 corrected 3 remaining defects to be corrected in future Montage release

Applications of Montage Large scale processing of the sky; e.g., Atlasmaker Mosaics of the Infrared Sky This is the age of Infrared Astronomy! Infrared astronomers study regions much larger than covered by individual cameras ==> need to make mosaics to investigate star formation, redshift distribution of galaxies Mosaics of the far infrared sky a primary data product of the SIRTF mission Two SIRTF Legacy teams using Montage as a co-registration and mosaic engine to generate science mosaics, perform image simulations, mission planning & pipeline testing Public Outreach - the wow factor! Combine single color images in Photoshop Example: back-lit display in NASA booth See next image - Rho Ophiuchi

Rho Ophiuchi 324 2MASS images in each band => 972 images On a 1 GHz Sun, mosaicking takes about 15 hours

Montage: The Grid Years Re-projection is slow (100 seconds for one 1024 x 512 pixel 2MASS image on a single processor 1.4 GHz Linux box), so use parallelization inherent in design Grid is an abstraction - array of processors, grid of clusters, … Montage has modular design - run on any environment Prototype architecture for ordering a mosaic through a web portal Request processed on a computing grid Prototype uses the Distributed Terascale Facility (Teragrid) This is one instance of how Montage could run on a grid Atlasmaker is another example of Montage parallelization

Montage: The Grid Years (cont.) Prototype version of a methodology for running on any “grid environment” Many parts of the process can be parallelized Build a script to enable parallelization Called a Directed Acyclical Graph (DAG) Describes flow of data and processing Describes which data are needed by which part of the job Describes what is to be run and when Standard tools can execute a DAG

Using Montage Grid Prototype Web service at JPL creates an abstract workflow description of Montage run (in XML) Workflow description run through Pegasus to create concrete DAG Pegasus includes transfer nodes in the concrete DAG for staging in the input image files and transferring out the generated mosaic Concrete DAG submitted to Condor Region Name, Degrees Pegasus Concrete DAG Condor DAGMAN Teragrid Cluster SDSC NCSA ISI Condor Pool JPL Abstract DAG

1 2 3 mProject 1mProject 2mProject mDiff 1 2mDiff 2 3 mFitplane D 12 mFitplane D 23 ax + by + c = 0dx + ey + f = 0 a 1 x + b 1 y + c 1 = 0 a 2 x + b 2 y + c 2 = 0 a 3 x + b 3 y + c 3 = 0 mBackground 1mBackground 2mBackground mAdd Final Mosaic Described as abstract DAG - specifies: Input, output, and intermediate files Processing jobs Dependencies between them D 12 D 23 Montage Workflow mConcatFit mBgModel ax + by + c = 0 dx + ey + f = 0

Montage – Concrete DAG (single pool) Data Stage in nodes Montage compute nodes Data stage out nodes Registration nodes Example DAG for 10 input files mAdd mBackground mBgModel mProject mDiff mFitPlane mConcatFit

Montage Runs on the Teragrid Test runs were done on the 1.5 degree x 1.5 degree area including M42 (Orion Nebula) Required 113 input image files Single pool DAG for SDSC consisted of 951 jobs 117 were data transfer jobs 113 for transferring the input image files 3 for transferring other header files 1 for transferring the final output mosaic Run took 94 minutes

Montage Runs on the Teragrid (2) Using same abstract DAG for multi pool DAG at SDSC and NCSA created 1202 jobs 367 were data transfer jobs Some of these jobs transferred multiple files 249 of these 367 were inter pool data transfer jobs

Montage Computations Building a mosaic from N 1024 x 512 pixel 2MASS images on a single processor 1.4 GHz Linux box takes roughly (N x 100) seconds 98-99% of this time is in the reprojection, which can be perfectly parallelized (this doesn’t embarrass us) Dataset# of images Size of each image Sky coverage Total number of pixels (x ) Storage size (TB) Processing time for all data in 1.4 GHz IA32 processor hours (x 1,000) 2MASS~ 4 million ~ 17’ x 8.5’ at 1” ~ 100%~ 2.1~ 8~ 111 DPOSS~ 2,600 ~ 6.6° x 6.6° at 1” ~ 50%~ 1.4~ 3~ 74 SDSS (DR1) ~ 50,000 ~ 13.6’ x 9’ at 0.4” ~ 25%~ 1.2~ 2.4~ 65

Summary Montage is a custom astronomical image mosaicking service that emphasizes astrometric and photometric accuracy First public release, Montage_v1.7.1, available for download at the Montage website A prototype Montage service has been deployed on the Teragrid; ties together distributed services at JPL, Caltech IPAC, and ISI More on Montage at SC2003: See ISI at ANL booth for demo of Pegasus portal running Montage on the Teragrid See back-lit display of Montage images at NASA booth Montage website: