Managing Workflows Within HUBzero: How to Use Pegasus to Execute Computational Pipelines Ewa Deelman USC Information Sciences Institute Acknowledgement:

Slides:



Advertisements
Similar presentations
Pegasus on the Virtual Grid: A Case Study of Workflow Planning over Captive Resources Yang-Suk Kee, Eun-Kyu Byun, Ewa Deelman, Kran Vahi, Jin-Soo Kim Oracle.
Advertisements

Natasha Pavlovikj, Kevin Begcy, Sairam Behera, Malachy Campbell, Harkamal Walia, Jitender S.Deogun University of Nebraska-Lincoln Evaluating Distributed.
The ADAMANT Project: Linking Scientific Workflows and Networks “Adaptive Data-Aware Multi-Domain Application Network Topologies” Ilia Baldine, Charles.
Ewa Deelman, Integrating Existing Scientific Workflow Systems: The Kepler/Pegasus Example Nandita Mangal,
XSEDE 13 July 24, Galaxy Team: PSC Team:
RCAC Research Computing Presents: DiaGird Overview Tuesday, September 24, 2013.
ProActive Task Manager Component for SEGL Parameter Sweeping Natalia Currle-Linde and Wasseim Alzouabi High Performance Computing Center Stuttgart (HLRS),
Technical Architectures
Workload Management Massimo Sgaravatto INFN Padova.
Managing Workflows with the Pegasus Workflow Management System
CONDOR DAGMan and Pegasus Selim Kalayci Florida International University 07/28/2009 Note: Slides are compiled from various TeraGrid Documentations.
CERN - IT Department CH-1211 Genève 23 Switzerland t Monitoring the ATLAS Distributed Data Management System Ricardo Rocha (CERN) on behalf.
KARMA with ProActive Parallel Suite 12/01/2009 Air France, Sophia Antipolis Solutions and Services for Accelerating your Applications.
Connecting OurGrid & GridSAM A Short Overview. Content Goals OurGrid: architecture overview OurGrid: short overview GridSAM: short overview GridSAM: example.
The Grid is a complex, distributed and heterogeneous execution environment. Running applications requires the knowledge of many grid services: users need.
Flexibility and user-friendliness of grid portals: the PROGRESS approach Michal Kosiedowski
Introduction to Scientific Workflows and Pegasus Karan Vahi Science Automation Technologies Group USC Information Sciences Institute.
OpenSees on NEEShub Frank McKenna UC Berkeley. Bell’s Law Bell's Law of Computer Class formation was discovered about It states that technology.
Grid Resource Allocation and Management (GRAM) Execution management Execution management –Deployment, scheduling and monitoring Community Scheduler Framework.
Through the development of advanced middleware, Grid computing has evolved to a mature technology in which scientists and researchers can leverage to gain.
Managing large-scale workflows with Pegasus Karan Vahi ( Collaborative Computing Group USC Information Sciences Institute Funded.
Experiences Using Cloud Computing for A Scientific Workflow Application Jens Vöckler, Gideon Juve, Ewa Deelman, Mats Rynge, G. Bruce Berriman Funded by.
BOSCO Architecture Derek Weitzel University of Nebraska – Lincoln.
Pegasus-a framework for planning for execution in grids Ewa Deelman USC Information Sciences Institute.
SDSS Quasars Spectra Fitting N. Kuropatkin, C. Stoughton.
INFSO-RI Module 01 ETICS Overview Etics Online Tutorial Marian ŻUREK Baltic Grid II Summer School Vilnius, 2-3 July 2009.
Use of Condor on the Open Science Grid Chris Green, OSG User Group / FNAL Condor Week, April
Distributed Aircraft Maintenance Environment - DAME DAME Workflow Advisor Max Ong University of Sheffield.
Dr. Ahmed Abdeen Hamed, Ph.D. University of Vermont, EPSCoR Research on Adaptation to Climate Change (RACC) Burlington Vermont USA MODELING THE IMPACTS.
Workflow Project Status Update Luciano Piccoli - Fermilab, IIT Nov
Resource Brokering in the PROGRESS Project Juliusz Pukacki Grid Resource Management Workshop, October 2003.
Tool Integration with Data and Computation Grid GWE - “Grid Wizard Enterprise”
Tarball server (for Condor installation) Site Headnode Worker Nodes Schedd glidein - special purpose Condor pool master DB Panda Server Pilot Factory -
Pegasus: Running Large-Scale Scientific Workflows on the TeraGrid Ewa Deelman USC Information Sciences Institute
NA-MIC National Alliance for Medical Image Computing UCSD: Engineering Core 2 Portal and Grid Infrastructure.
Intermediate Condor: Workflows Rob Quick Open Science Grid Indiana University.
What is SAM-Grid? Job Handling Data Handling Monitoring and Information.
6 February 2009 ©2009 Cesare Pautasso | 1 JOpera and XtremWeb-CH in the Virtual EZ-Grid Cesare Pautasso Faculty of Informatics University.
EGEE-III INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks Services for advanced workflow programming.
CEDPS Data Services Ann Chervenak USC Information Sciences Institute.
Pegasus WMS: Leveraging Condor for Workflow Management Ewa Deelman, Gaurang Mehta, Karan Vahi, Gideon Juve, Mats Rynge, Prasanth.
Experiment Management from a Pegasus Perspective Jens-S. Vöckler Ewa Deelman
Intermediate Condor: Workflows Monday, 1:15pm Alain Roy OSG Software Coordinator University of Wisconsin-Madison.
Pegasus-a framework for planning for execution in grids Karan Vahi USC Information Sciences Institute May 5 th, 2004.
Planning Ewa Deelman USC Information Sciences Institute GriPhyN NSF Project Review January 2003 Chicago.
Development of e-Science Application Portal on GAP WeiLong Ueng Academia Sinica Grid Computing
Tool Integration with Data and Computation Grid “Grid Wizard 2”
Funded by the NSF OCI program grants OCI and OCI Mats Rynge, Gideon Juve, Karan Vahi, Gaurang Mehta, Ewa Deelman Information Sciences Institute,
ATLAS-specific functionality in Ganga - Requirements for distributed analysis - ATLAS considerations - DIAL submission from Ganga - Graphical interfaces.
© Geodise Project, University of Southampton, Workflow Support for Advanced Grid-Enabled Computing Fenglian Xu *, M.
1 USC Information Sciences InstituteYolanda Gil AAAI-08 Tutorial July 13, 2008 Part IV Workflow Mapping and Execution in Pegasus (Thanks.
Managing LIGO Workflows on OSG with Pegasus Karan Vahi USC Information Sciences Institute
Job submission overview Marco Mambelli – August OSG Summer Workshop TTU - Lubbock, TX THE UNIVERSITY OF CHICAGO.
VO Experiences with Open Science Grid Storage OSG Storage Forum | Wednesday September 22, 2010 (10:30am)
HUBzero® Platform for Scientific Collaboration Copyright © 2012 HUBzero Foundation, LLC International Workshop on Science Gateways, ETH Zürich, June 3-5,
Integrating Scientific Tools and Web Portals
Pegasus WMS Extends DAGMan to the grid world
Using simulation workspaces to “submit” jobs and workflows
GWE Core Grid Wizard Enterprise (
Workflows and the Pegasus Workflow Management System
MIK 2.1 DBNS - introduction to WS-PGRADE, 2013
Pegasus and Condor Gaurang Mehta, Ewa Deelman, Carl Kesselman, Karan Vahi Center For Grid Technologies USC/ISI.
University of Southern California
SDM workshop Strawman report History and Progress and Goal.
Module 01 ETICS Overview ETICS Online Tutorials
Overview of Workflows: Why Use Them?
Mats Rynge USC Information Sciences Institute
A General Approach to Real-time Workflow Monitoring
Frieda meets Pegasus-WMS
Integrity Checking in Pegasus
Presentation transcript:

Managing Workflows Within HUBzero: How to Use Pegasus to Execute Computational Pipelines Ewa Deelman USC Information Sciences Institute Acknowledgement: Steven Clark, Derrick Kearney, Michael McLennan (HUBzero) Frank McKenna (OpenSees) Gideon Juve, Gaurang Mehta, Mats Rynge, Karan Vahi (Pegasus)

Outline Introduction to Pegasus and workflows HUB Integration –Rappture and Pegasus –Submit command and Pegasus Example: OpenSEES / NEESHub Future directions

Computational workflows Help express multi-step computations in a declarative way Can support automation, minimize human involvement –Makes analyses easier to run Can be high-level and portable across execution platforms Keep track of provenance to support reproducibility Foster collaboration—code and data sharing

Workflow Management You may want to use different resources within a workflow or over time Need a high-level workflow specification Need a planning capability to map from high-level to executable workflow Need to manage the task dependencies Need to manage the execution of tasks on the remote resources Need to provide scalability, performance, reliability

Our Approach Analysis Representation Support a declarative representation for the workflow (dataflow) Represent the workflow structure as a Directed Acyclic Graph (DAG) in a resource-independent way Use recursion to achieve scalability System (Plan for the resources, Execute the Plan, Manage tasks) Layered architecture, each layer is responsible for a particular function (Pegasus Planner, DAGMan, Condor schedd) Mask errors at different levels of the system Modular, composed of well-defined components, where different components can be swapped in Use and adapt existing graph and other relevant algorithms Can be embedded into

Pegasus Workflow Management System (est. 2001) A collaboration with University of Wisconsin Madison Used by a number of applications in a variety of domains Provides reliability—can retry computations from the point of failure Provides scalability—can handle large data and many computations (kbytes-TB of data, tasks) Optimizes workflows for performance Automatically captures provenance information Runs workflows on distributed resources: laptop, campus cluster, Grids (DiaGrid, OSG, XSEDE), Clouds (FutureGrid, EC2, etc..)

Planning Process Assume data may be distributed in the Environment Assume you may want to use local and/or remote resources Pegasus needs information about the environment – data, executables, execution and data storage sites Pegasus generates an executable workflow Data transfer protocols – Gridftp, Condor I/O, HTTP, scp, S3, iRods, SRM, FDT (partial) Scheduling to interfaces – Local, Gram, Condor, Condor-C (for remote Condor pools), via Condor Glideins – PBS, LSF, SGE

Generating executable workflows 8 (DAX) APIs for workflow specification (DAX--- DAG in XML) Java, Perl, Python

Advanced features Performs data reuse Registers data in data catalogs Manages storage—deletes data no longer needed Can cluster tasks together for performance Can manage complex data architectures (shared and non-shared filesystem, distributed data sources) Different execution modes which leverage different computing architectures (Condor pools, HPC resources, etc..)

HUBzero Integration Pegasus with

Benefits of Pegasus for HUB Users Provides Support for Complex Computations –Can connect the existing HUB models into larger computations Portability / Reuse –User created workflows can easily be run in different environments without alteration (today DiaGrid, OSG) Performance –The Pegasus mapper can reorder, group, and prioritize tasks in order to increase the overall workflow performance. Scalability –Pegasus can easily scale both the size of the workflow, and the resources that the workflow is distributed over. 12

Benefits of Pegasus for HUB Users Provenance –Performance and provenance data is collected in a database, and the data can be summaries with tools such as pegasus- statistics, pegasus-plots, or directly with SQL queries. Reliability –Jobs and data transfers are automatically retried in case of failures. Debugging tools such as pegasus-analyzer helps the user to debug the workflow in case of non-recoverable failures. 13

Pegasus in HUBzero Pegasus as a backend to the submit command Pegasus workflows composed in Rappture – Build workflow within Rappture – Have Rappture collect inputs, call a workflow generator, and collect outputs Pegasus Tutorial tool now available in HUBzero Session that includes Pegasus on Tuesday 1:30 – 5:30 Room 206 #2 Creating and Deploying Scientific Tools (part 2) “… Scientific Workflows with Pegasus” by George Howlett & Derrick Kearney, Purdue University Acknowledgements: Steven Clark and Derrick Kearney, Purdue University

Submit host Pegasus Workflow Management System Performance Database Laptop Campus Clusters Grid Clusters Clouds Site info Abstract Workflow (DAX) Data and transformation info Execution Info

Use of Pegasus with Submit Command Used by Rappture interface to submit the workflow Submits the workflow through Pegasus to – OSG – DIAGRID Prepares the site catalog and other configuration files for Pegasus Uses pegasus-status to track the workflow Generates statistics and report about job failures using pegasus tools. Acknowledgements: Steven Clark and Derrick Kearney, Purdue University

Submit host Pegasus Workflow Management System Performance Database Laptop Campus Clusters Grid Clusters Clouds Site info Abstract Workflow (DAX) Data and transformation info Hub Execution Info

Inputs Tool description Outputs Acknowledgements: Steven Clark and Derrick Kearney, Purdue University Rappture (data definitions) Calls an external DAX generator Pegasus Workflows in the HUB

wrapper.py Python script Collects the data from the Rappture interface Generates the DAX Runs the workflow Presents the outputs to Rappture Acknowledgements: Steven Clark and Derrick Kearney, Purdue University

Workflow generation Acknowledgements: Steven Clark and Derrick Kearney, Purdue University

User provides inputs to the workflow and clicks the “Submit” button Acknowledgements: Steven Clark and Derrick Kearney, Purdue University

Workflow has completed. Outputs are available for browsing/downloading Acknowledgements: Steven Clark and Derrick Kearney, Purdue University

OpenSEES / NEEShub

The OpenSeesLab tool: Is a suite of Simulation Tools powered by OpenSees for: 1.Submitting OpenSees scripts to NEEShub resources 2.Educating students and practicing engineers Acknowledgements: Frank McKenna from UC Berkeley

Rappture MatlabOpenSeesMatlab OpenSees Matlab is used to generate random material properties 10’s to 1000’s of OpenSees Simulations Matlab is used to process the results and generate figures Pegasus is Responsible for moving the data from the NEEShub to the OSG, orchestrating the workflow and returning the results to NEEShub. Acknowledgements: Frank McKenna from UC Berkeley Rappture implementation in TCL calls out to an external Python DAX generator. OpenSees uses Pegasus to run on Open Science Grid

Future Directions Submit to manage parameter sweep computations (now only on HUBzer0) Web-based monitoring

Benefits of workflows in the HUB Support for complex applications/ builds on existing domain tools Clean separations for users/developers/operator –User: Nice high level interface via Rappture –Tool developer: Only has to build/provide a description of the workflow (DAX) –Hub operator: Ties the Hub to an existing distributed computing infrastructure (DiaGrid, OSG, …) The Hub and Pegasus handle low level details –Job scheduling to various execution environments –Data staging in a distributed environment –Job retries –Workflow analysis –Support for large workflows

Benefits of the HUB to Pegasus Provides a nice, easy to use interface to Pegasus workflows Broadens the user base Improves the software based on user’s feedback Drives innovation—new deployment scenarios, use cases I look forward to a continued collaboration

Further Information Session that includes Pegasus on Tuesday 1:30 – 5:30 – Room 206 #2 Creating and Deploying Scientific Tools (part 2) – “… Scientific Workflows with Pegasus” by George Howlett & Derrick Kearney, Purdue University Pegasus Tutorial on the HUB – General Pegasus Information Pegasus in a VM—allows you to develop DAXes – – We are happy to help! Support mailing lists pegasus- Contact me Big Thank You to the HUBzero and OpenSees teams!