Download presentation
Presentation is loading. Please wait.
Published byEmily Lindsey Modified over 9 years ago
1
Pegasus: Planning for Execution in Grids Ewa Deelman, Carl Kesselman, Gaurang Mehta, Gurmeet Singh, Karan Vahi Information Sciences Institute University of Southern California
2
Pegasus Acknowledgement l Ewa Deelman, Carl Kesselman, Saurabh Khurana, Gaurang Mehta, Sonal Patil, Gurmeet Singh, Mei-Hui Su, Karan Vahi (ISI) l James Blythe, Yolanda Gil (ISI) l http://pegasus.isi.edu http://pegasus.isi.edu l Research funded as part of the NSF GriPhyN, NVO and SCEC projects.
3
Virtual Data Concept l Capture and manage information about relationships among u Data (of widely varying representations) u Programs (& their execution needs) u Computations (& execution environments) l Apply this information to, e.g. u Discovery: Data and program discovery u Workflow: Structured paradigm for organizing, locating, specifying, & requesting data u Explanation: provenance l Research part of NSF funded GriPhyN project
4
Grid Applications l Increasing in the level of complexity l Use of individual application components l Reuse of individual intermediate data products l Description of Data Products using Metadata Attributes l Execution environment is complex and very dynamic u Resources come and go u Data is replicated u Components can be found at various locations or staged in on demand l Separation between u the application description u the actual execution description
5
Abstract Workflow Generation Concrete Workflow Generation
6
Pegasus: Planning for Execution in Grids l Maps from abstract to concrete workflow u Algorithmic and AI based techniques l Automatically locates physical locations for both components (transformations) and data u Use Globus RLS and the Transformation Catalog l Finds appropriate resources to execute u via Globus MDS l Reuses existing data products where applicable l Publishes newly derived data products u Chimera virtual data catalog
7
Chimera is developed at ANL By I. Foster, M. Wilde, and J. Voeckler
9
LIGO Scientific Collaboration l Continuous gravitational waves are expected to be produced by a variety of celestial objects l Only a small fraction of potential sources are known l Need to perform blind searches, scanning the regions of the sky where we have no a priori information of the presence of a source u Wide area, wide frequency searches l Search is performed for potential sources of continuous periodic waves near the Galactic Center and the galactic core l The search is very compute and data intensive l LSC is using the occasion of SC2003 to initiate a month- long production run with science data collected during 8 weeks in the Spring of 2003
10
Additional resources used: Grid3 iVDGL resources Thanks to everyone involved in standing up the tested and contributing the resources!
11
LIGO Acknowledgements l Bruce Allen, Scott Koranda, Brian Moe, Xavier Siemens, University of Wisconsin Milwaukee, USA l Stuart Anderson, Kent Blackburn, Albert Lazzarini, Dan Kozak, Hari Pulapaka, Peter Shawhan, Caltech, USA l Steffen Grunewald, Yousuke Itoh, Maria Alessandra Papa, Albert Einstein Institute, Germany l Many Others involved in the Testbed l www.ligo.caltech.edu l www.lsc- group.phys.uwm.edu/lscdatagrid/ l http://pandora.aei.mpg.de/merlin/ l LIGO Laboratory operates under NSF cooperative agreement PHY-0107417
12
Montage l Montage (NASA and NVO) u Deliver science-grade custom mosaics on demand u Produce mosaics from a wide range of data sources (possibly in different spectra) u User-specified parameters of projection, coordinates, size, rotation and spatial sampling. Mosaic created by Pegasus based Montage from a run of the M101 galaxy images on the Teragrid.
13
Small Montage Workflow
14
Montage Acknowledgments l Bruce Berriman, John Good, Anastasia Laity, Caltech/IPAC l Joseph C. Jacob, Daniel S. Katz, JPL l http://montage.ipac. caltech.edu/ http://montage.ipac l Testbed for Montage: Condor pools at USC/ISI, UW Madison, and Teragrid resources at NCSA, PSC, and SDSC. Montage is funded by the National Aeronautics and Space Administration's Earth Science Technology Office, Computational Technologies Project, under Cooperative Agreement Number NCC5-626 between NASA and the California Institute of Technology.
15
Current System
16
Just In-time planning l Partition Abstract workflow into partial workflows
17
Meta-DAGMan
18
Other Applications Using Pegasus l Other GriPhyN applications: u High-energy physics: Atlas, CMS (many) u Astronomy: SDSS (Fermi Lab, ANL) l Astronomy: u Galaxy Morphology (NCSA, JHU, Fermi, many others, NVO-funded) l Biology u BLAST (ANL, PDQ-funded) l Neuroscience u Tomography (SDSC, NIH-funded) l http://pegasus.isi.edu l Funding by NSF GriPhyN, NSF NVO, NIH
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.