Ewa Deelman, Pegasus and DAGMan: From Concept to Execution Mapping Scientific Workflows onto the National.

Slides:



Advertisements
Similar presentations
Exploring Issues with Workflow Scheduling on the Grid Rizos Sakellariou University of Manchester, UK with thanks to: Henan Zhao and Ewa Deelman for providing.
Advertisements

Pegasus on the Virtual Grid: A Case Study of Workflow Planning over Captive Resources Yang-Suk Kee, Eun-Kyu Byun, Ewa Deelman, Kran Vahi, Jin-Soo Kim Oracle.
Managing Workflows Within HUBzero: How to Use Pegasus to Execute Computational Pipelines Ewa Deelman USC Information Sciences Institute Acknowledgement:
The ADAMANT Project: Linking Scientific Workflows and Networks “Adaptive Data-Aware Multi-Domain Application Network Topologies” Ilia Baldine, Charles.
Ewa Deelman, Integrating Existing Scientific Workflow Systems: The Kepler/Pegasus Example Nandita Mangal,
Ewa Deelman, Optimizing for Time and Space in Distributed Scientific Workflows Ewa Deelman University.
Ewa Deelman, Clouds: An Opportunity for Scientific Applications? Ewa Deelman USC Information Sciences.
Pegasus: Mapping complex applications onto the Grid Ewa Deelman Center for Grid Technologies USC Information Sciences Institute.
An Astronomical Image Mosaic Service for the National Virtual Observatory
A Grid-Enabled Engine for Delivering Custom Science- Grade Images on Demand
An Astronomical Image Mosaic Service for the National Virtual Observatory / ESTO.
Ewa Deelman Using Grid Technologies to Support Large-Scale Astronomy Applications Ewa Deelman Center for Grid Technologies USC Information.
Ewa Deelman, Clouds: An Opportunity for Scientific Applications? Ewa Deelman USC Information Sciences.
Managing Workflows with the Pegasus Workflow Management System
CONDOR DAGMan and Pegasus Selim Kalayci Florida International University 07/28/2009 Note: Slides are compiled from various TeraGrid Documentations.
Pegasus A Framework for Workflow Planning on the Grid Ewa Deelman USC Information Sciences Institute Pegasus Acknowledgments: Carl Kesselman, Gaurang Mehta,
The Grid is a complex, distributed and heterogeneous execution environment. Running applications requires the knowledge of many grid services: users need.
10/20/05 LIGO Scientific Collaboration 1 LIGO Data Grid: Making it Go Scott Koranda University of Wisconsin-Milwaukee.
Large-Scale Science Through Workflow Management Ewa Deelman Center for Grid Technologies USC Information Sciences Institute.
OSG Area Coordinators’ Meeting LIGO Applications (NEW) Kent Blackburn Caltech / LIGO October 29 th,
Managing large-scale workflows with Pegasus Karan Vahi ( Collaborative Computing Group USC Information Sciences Institute Funded.
 The workflow description modified to output a VDS DAX.  The workflow description toolkit developed allows any concrete workflow description to be migrated.
Pegasus-a framework for planning for execution in grids Ewa Deelman USC Information Sciences Institute.
Pegasus: Planning for Execution in Grids Ewa Deelman Information Sciences Institute University of Southern California.
1 USC Information Sciences Institute Yolanda Gil AAAI-08 Tutorial July 13, 2008 AAAI-08 Tutorial on Computational Workflows for Large-Scale.
Dr. Ahmed Abdeen Hamed, Ph.D. University of Vermont, EPSCoR Research on Adaptation to Climate Change (RACC) Burlington Vermont USA MODELING THE IMPACTS.
Pegasus: Mapping Scientific Workflows onto the Grid Ewa Deelman Center for Grid Technologies USC Information Sciences Institute.
Condor Week 2005Optimizing Workflows on the Grid1 Optimizing workflow execution on the Grid Gaurang Mehta - Based on “Optimizing.
CyberShake Study 2.3 Technical Readiness Review. Study re-versioning SCEC software uses year.month versioning Suggest renaming this study to 13.4.
1 USC INFORMATION SCIENCES INSTITUTE CAT: Composition Analysis Tool Interactive Composition of Computational Pathways Yolanda Gil Jihie Kim Varun Ratnakar.
Pegasus: Running Large-Scale Scientific Workflows on the TeraGrid Ewa Deelman USC Information Sciences Institute
Intermediate Condor: Workflows Rob Quick Open Science Grid Indiana University.
Pegasus: Mapping complex applications onto the Grid Ewa Deelman Center for Grid Technologies USC Information Sciences Institute.
GriPhyN Virtual Data System Grid Execution of Virtual Data Workflows Mike Wilde Argonne National Laboratory Mathematics and Computer Science Division.
CEDPS Data Services Ann Chervenak USC Information Sciences Institute.
Pegasus WMS: Leveraging Condor for Workflow Management Ewa Deelman, Gaurang Mehta, Karan Vahi, Gideon Juve, Mats Rynge, Prasanth.
Experiment Management from a Pegasus Perspective Jens-S. Vöckler Ewa Deelman
Pegasus-a framework for planning for execution in grids Karan Vahi USC Information Sciences Institute May 5 th, 2004.
Planning Ewa Deelman USC Information Sciences Institute GriPhyN NSF Project Review January 2003 Chicago.
Pegasus: Planning for Execution in Grids Ewa Deelman, Carl Kesselman, Gaurang Mehta, Gurmeet Singh, Karan Vahi Information Sciences Institute University.
LIGO Plans for OSG J. Kent Blackburn LIGO Laboratory California Institute of Technology Open Science Grid Technical Meeting UCSD December 15-17, 2004.
State of LSC Data Analysis and Software LSC Meeting LIGO Hanford Observatory November 11 th, 2003 Kent Blackburn, Stuart Anderson, Albert Lazzarini LIGO.
CyberShake Study 15.3 Science Readiness Review. Study 15.3 Scientific Goals Calculate a 1 Hz map of Southern California Produce meaningful 2 second results.
2005 GRIDS Community Workshop1 Learning From Cyberinfrastructure Initiatives Grid Research Integration Development & Support
Funded by the NSF OCI program grants OCI and OCI Mats Rynge, Gideon Juve, Karan Vahi, Gaurang Mehta, Ewa Deelman Information Sciences Institute,
April 25, 2006Parag Mhashilkar, Fermilab1 Resource Selection in OSG & SAM-On-The-Fly Parag Mhashilkar Fermi National Accelerator Laboratory Condor Week.
LIGO-G Z1 Using Condor for Large Scale Data Analysis within the LIGO Scientific Collaboration Duncan Brown California Institute of Technology.
1 Pegasus and wings WINGS/Pegasus Provenance Challenge Ewa Deelman Yolanda Gil Jihie Kim Gaurang Mehta Varun Ratnakar USC Information Sciences Institute.
1 USC Information Sciences InstituteYolanda Gil AAAI-08 Tutorial July 13, 2008 Part IV Workflow Mapping and Execution in Pegasus (Thanks.
Managing LIGO Workflows on OSG with Pegasus Karan Vahi USC Information Sciences Institute
Ewa Deelman, Managing Scientific Workflows on OSG with Pegasus Ewa Deelman USC Information Sciences.
Resource Allocation and Scheduling for Workflows Gurmeet Singh, Carl Kesselman, Ewa Deelman.
1 Performance Impact of Resource Provisioning on Workflows Gurmeet Singh, Carl Kesselman and Ewa Deelman Information Science Institute University of Southern.
VO Experiences with Open Science Grid Storage OSG Storage Forum | Wednesday September 22, 2010 (10:30am)
INTRODUCTION TO XSEDE. INTRODUCTION  Extreme Science and Engineering Discovery Environment (XSEDE)  “most advanced, powerful, and robust collection.
SCEC CyberShake on TG & OSG: Options and Experiments Allan Espinosa*°, Daniel S. Katz*, Michael Wilde*, Ian Foster*°,
Pegasus WMS Extends DAGMan to the grid world
Cloudy Skies: Astronomy and Utility Computing
Workflows and the Pegasus Workflow Management System
CyberShake Study 16.9 Discussion
Pegasus and Condor Gaurang Mehta, Ewa Deelman, Carl Kesselman, Karan Vahi Center For Grid Technologies USC/ISI.
Clouds: An Opportunity for Scientific Applications?
University of Southern California
Ewa Deelman University of Southern California
Mats Rynge USC Information Sciences Institute
rvGAHP – Push-Based Job Submission Using Reverse SSH Connections
High Throughput Computing for Astronomers
CyberShake Study 14.2 Science Readiness Review
Frieda meets Pegasus-WMS
CyberShake Study 18.8 Planning
Presentation transcript:

Ewa Deelman, Pegasus and DAGMan: From Concept to Execution Mapping Scientific Workflows onto the National Cyberinfrastructure Ewa Deelman USC Information Sciences Institute

Ewa Deelman, Acknowledgments Pegasus: Gaurang Mehta, Mei-Hui Su, Karan Vahi (developers), Nandita Mandal, Arun Ramakrishnan, Tsai-Ming Tseng (students) DAGMan: Miron Livny and the Condor team Other Collaborators: Yolanda Gil, Jihie Kim, Varun Ratnakar (Wings System) LIGO: Kent Blackburn, Duncan Brown, Stephen Fairhurst, David Meyers Montage: Bruce Berriman, John Good, Dan Katz, and Joe Jacobs SCEC: Tom Jordan, Robert Graves, Phil Maechling, David Okaya, Li Zhao

Ewa Deelman, Outline Pegasus and DAGMan system Description Illustration of features through science applications running on OSG and the TeraGrid Minimizing the workflow data footprint Results of running LIGO applications on OSG

Ewa Deelman, Scientific (Computational) Workflows Enable the assembly of community codes into large- scale analysis Montage example: Generating science-grade mosaics of the sky (Bruce Berriman, Caltech)

Ewa Deelman, Pegasus and Condor DAGMan Automatically map high-level resource-independent workflow descriptions onto distributed resources such as the Open Science Grid and the TeraGrid Improve performance of applications through: Data reuse to avoid duplicate computations and provide reliability Workflow restructuring to improve resource allocation Automated task and data transfer scheduling to improve overall runtime Provide reliability through dynamic workflow remapping and execution Pegasus and DAGMan applications include LIGO’s Binary Inspiral Analysis, NVO’s Montage, SCEC’s CyberShake simulations, Neuroscience, Artificial Intelligence, Genomics (GADU), others Workflows with thousands of tasks and TeraBytes of data Use Condor and Globus to provide the middleware for distributed environments

Ewa Deelman, Pegasus Workflow Mapping Original workflow: 15 compute nodes devoid of resource assignment Resulting workflow mapped onto 3 Grid sites: 11 compute nodes (4 reduced based on available intermediate data) 13 data stage-in nodes 8 inter-site data transfers 14 data stage-out nodes to long- term storage 14 data registration nodes (data cataloging)

Ewa Deelman, Typical Pegasus and DAGMan Deployment

Ewa Deelman, Supporting OSG Applications LIGO—Laser Interferometer Gravitational-Wave Observatory Aims to find gravitational waves emitted by objects such as binary inpirals 9.7 Years of CPU time over 6 months Work done by Kent Blackburn, David Meyers, Michael Samidi, Caltech

Ewa Deelman, SCEC workflows run each week using Pegasus and DAGMan on the TeraGrid and USC resources. Cumulatively, the workflows consisted of over half a million tasks and used over 2.5 CPU Years. Managing Large-Scale Workflow Execution from Resource Provisioning to Provenance tracking: The CyberShake Example, Ewa Deelman, Scott Callaghan, Edward Field, Hunter Francoeur, Robert Graves, Nitin Gupta, Vipin Gupta, Thomas H. Jordan, Carl Kesselman, Philip Maechling, John Mehringer, Gaurang Mehta, David Okaya, Karan Vahi, Li Zhao, e-Science 2006, Amsterdam, December 4-6, 2006, best paper award Scalability

Ewa Deelman, Montage application ~7,000 compute jobs in instance ~10,000 nodes in the executable workflow same number of clusters as processors speedup of ~15 on 32 processors Small 1,200 Montage Workflow Performance optimization through workflow restructuring Pegasus: a Framework for Mapping Complex Scientific Workflows onto Distributed Systems, Ewa Deelman, Gurmeet Singh, Mei-Hui Su, James Blythe, Yolanda Gil, Carl Kesselman, Gaurang Mehta, Karan Vahi, G. Bruce Berriman, John Good, Anastasia Laity, Joseph C. Jacob, Daniel S. Katz, Scientific Programming Journal, Volume 13, Number 3, 2005

Ewa Deelman, Data Reuse Sometimes it is cheaper to access the data than to regenerate it Keeping track of data as it is generated supports workflow-level checkpointing Mapping Complex Workflows Onto Grid Environments, E. Deelman, J. Blythe, Y. Gil, C. Kesselman, G. Mehta, K. Vahi, K. Backburn, A. Lazzarini, A. Arbee, R. Cavanaugh, S. Koranda, Journal of Grid Computing, Vol.1, No. 1, 2003., pp25-39.

Ewa Deelman, Efficient data handling (Similar order of intermediate/output files) If not enough space-failures occur Solution: Reduce Workflow Data Footprint Determine which data are no longer needed and when Add nodes to the workflow do cleanup data along the way Benefits: simulations showed up to 57% space improvements for LIGO-like workflows Workflow input data is staged dynamically, new data products are generated during execution For large workflows 10,000+ input files “Scheduling Data-Intensive Workflows onto Storage-Constrained Distributed Resources”, A. Ramakrishnan, G. Singh, H. Zhao, E. Deelman, R. Sakellariou, K. Vahi, K. Blackburn, D. Meyers, and M. Samidi, accepted to CCGrid 2007

Ewa Deelman, LIGO Inspiral Analysis Workflow Small Workflow: 164 nodes Full Scale analysis: 185,000 nodes and 466,000 edges 10 TB of input data and 1 TB of output data LIGO workflow running on OSG “Optimizing Workflow Data Footprint” G. Singh, K. Vahi, A. Ramakrishnan, G. Mehta, E. Deelman, H. Zhao, R. Sakellariou, K. Blackburn, D. Brown, S. Fairhurst, D. Meyers, G. B. Berriman, J. Good, D. S. Katz, in submission

Ewa Deelman, LIGO Workflows 26% Improvement In disk space Usage 50% slower runtime

Ewa Deelman, LIGO Workflows 56% improvement in space usage 3 times slower in runtime Looking into new DAGMan capabilities for workflow node prioritization Need automated techniques to determine priorities

Ewa Deelman, What do Pegasus & DAGMan do for an application? Provide a level of abstraction above gridftp, condor- submit, globus-job-run, etc commands Provide automated mapping and execution of workflow applications onto distributed resources Manage data files, can store and catalog intermediate and final data products Improve successful application execution Improve application performance Provide provenance tracking capabilities Provides a Grid-aware workflow management tool

Ewa Deelman, Relevant Links Pegasus: pegasus.isi.edupegasus.isi.edu Currently released as part of VDS and VDT Standalone pegasus distribution v 2.0 coming out in May 2007, will remain part of VDT DAGMan: NSF Workshop on Challenges of Scientific Workflows : E. Deelman and Y. Gil (chairs) Workflows for e-Science, Taylor, I.J.; Deelman, E.; Gannon, D.B.; Shields, M. (Eds.), Dec Open Science Grid: LIGO: SCEC: Montage: montage.ipac.caltech.edu/montage.ipac.caltech.edu/ Condor: Globus: TeraGrid: Ewa Deelman,