Pegasus Workflows on XSEDE

Slides:



Advertisements
Similar presentations
MapReduce Online Created by: Rajesh Gadipuuri Modified by: Ying Lu.
Advertisements

EHarmony in Cloud Subtitle Brian Ko. eHarmony Online subscription-based matchmaking service Available in United States, Canada, Australia and United Kingdom.
A Dynamic World, what can Grids do for Multi-Core computing? Daniel Goodman, Anne Trefethen and Douglas Creager
Natasha Pavlovikj, Kevin Begcy, Sairam Behera, Malachy Campbell, Harkamal Walia, Jitender S.Deogun University of Nebraska-Lincoln Evaluating Distributed.
The ADAMANT Project: Linking Scientific Workflows and Networks “Adaptive Data-Aware Multi-Domain Application Network Topologies” Ilia Baldine, Charles.
1 High Performance Computing at SCEC Scott Callaghan Southern California Earthquake Center University of Southern California.
6th Biennial Ptolemy Miniconference Berkeley, CA May 12, 2005 Distributed Computing in Kepler Ilkay Altintas Lead, Scientific Workflow Automation Technologies.
Homework 2 In the docs folder of your Berkeley DB, have a careful look at documentation on how to configure BDB in main memory. In the docs folder of your.
Intermediate HTCondor: Workflows Monday pm Greg Thain Center For High Throughput Computing University of Wisconsin-Madison.
CREATING A MULTI-WAVELENGTH GALACTIC PLANE ATLAS WITH AMAZON WEB SERVICES G. Bruce Berriman, John Good IPAC, California Institute of Technolog y Ewa Deelman,
Managing Workflows with the Pegasus Workflow Management System
Communicating with Users about HTCondor and High Throughput Computing Lauren Michael, Research Computing Facilitator HTCondor Week 2015.
CONDOR DAGMan and Pegasus Selim Kalayci Florida International University 07/28/2009 Note: Slides are compiled from various TeraGrid Documentations.
Apache Airavata GSOC Knowledge and Expertise Computational Resources Scientific Instruments Algorithms and Models Archived Data and Metadata Advanced.
Authors: Weiwei Chen, Ewa Deelman 9th International Conference on Parallel Processing and Applied Mathmatics 1.
IPlant Collaborative Tools and Services Workshop iPlant Collaborative Tools and Services Workshop Collaborating with iPlant.
ISG We build general capability Introduction to Olympus Shawn T. Brown, PhD ISG MISSION 2.0 Lead Director of Public Health Applications Pittsburgh Supercomputing.
CS525: Special Topics in DBs Large-Scale Data Management Hadoop/MapReduce Computing Paradigm Spring 2013 WPI, Mohamed Eltabakh 1.
Software for Science Gateways: Open Grid Computing Environments Marlon Pierce, Suresh Marru Pervasive Technology Institute Indiana University
NSF Geoinformatics Project (Sept 2012 – August 2014) Geoinformatics: Community Computational Platforms for Developing Three-Dimensional Models of Earth.
Introduction to Scientific Workflows and Pegasus Karan Vahi Science Automation Technologies Group USC Information Sciences Institute.
IPlant Collaborative Tools and Services Workshop iPlant Collaborative Tools and Services Workshop Collaborating with iPlant.
Managing large-scale workflows with Pegasus Karan Vahi ( Collaborative Computing Group USC Information Sciences Institute Funded.
Experiences Using Cloud Computing for A Scientific Workflow Application Jens Vöckler, Gideon Juve, Ewa Deelman, Mats Rynge, G. Bruce Berriman Funded by.
IPlant Collaborative Tools and Services Workshop iPlant Collaborative Tools and Services Workshop Collaborating with iPlant.
Dr. Ahmed Abdeen Hamed, Ph.D. University of Vermont, EPSCoR Research on Adaptation to Climate Change (RACC) Burlington Vermont USA MODELING THE IMPACTS.
OPENQUAKE Mission and Vision It is GEM’s mission to engage a global community in the design, development and deployment of state-of-the-art models and.
San Diego Supercomputer Center Grid Physics Network (GriPhyN) University of Florida DGL: The Assembly Language for Grid Computing Arun swaran Jagatheesan.
Pegasus: Running Large-Scale Scientific Workflows on the TeraGrid Ewa Deelman USC Information Sciences Institute
Intermediate Condor: Workflows Rob Quick Open Science Grid Indiana University.
Experiences Running Seismic Hazard Workflows Scott Callaghan Southern California Earthquake Center University of Southern California SC13 Workflow BoF.
Hadoop/MapReduce Computing Paradigm 1 CS525: Special Topics in DBs Large-Scale Data Management Presented By Kelly Technologies
Funded by the NSF OCI program grants OCI and OCI Mats Rynge, Gideon Juve, Karan Vahi, Gaurang Mehta, Ewa Deelman Information Sciences Institute,
1 USC Information Sciences InstituteYolanda Gil AAAI-08 Tutorial July 13, 2008 Part IV Workflow Mapping and Execution in Pegasus (Thanks.
PEER 2003 Meeting 03/08/031 Interdisciplinary Framework Major focus areas Structural Representation Fault Systems Earthquake Source Physics Ground Motions.
Managing LIGO Workflows on OSG with Pegasus Karan Vahi USC Information Sciences Institute
INTRODUCTION TO XSEDE. INTRODUCTION  Extreme Science and Engineering Discovery Environment (XSEDE)  “most advanced, powerful, and robust collection.
SCEC CyberShake on TG & OSG: Options and Experiments Allan Espinosa*°, Daniel S. Katz*, Michael Wilde*, Ian Foster*°,
Scientific Data Processing Portal and Heterogeneous Computing Resources at NRC “Kurchatov Institute” V. Aulov, D. Drizhuk, A. Klimentov, R. Mashinistov,
HUBzero® Platform for Scientific Collaboration Copyright © 2012 HUBzero Foundation, LLC International Workshop on Science Gateways, ETH Zürich, June 3-5,
Overview of Scientific Workflows: Why Use Them?
HPC In The Cloud Case Study: Proteomics Workflow
Integrating Scientific Tools and Web Portals
Pegasus WMS Extends DAGMan to the grid world
Volunteer Computing for Science Gateways
Tools and Services Workshop
Cloudy Skies: Astronomy and Utility Computing
Joslynn Lee – Data Science Educator
Simplify Your Science with Workflow Tools
Seismic Hazard Analysis Using Distributed Workflows
HTCondor and LSST Stephen Pietrowicz Senior Research Programmer National Center for Supercomputing Applications HTCondor Week May 2-5, 2017.
Scott Callaghan Southern California Earthquake Center
Workflows and the Pegasus Workflow Management System
CyberShake Study 16.9 Discussion
Philip J. Maechling (SCEC) September 13, 2015
MapReduce Computing Paradigm Basics Fall 2013 Elke A. Rundensteiner
Cloud Distributed Computing Environment Hadoop
湖南大学-信息科学与工程学院-计算机与科学系
Development of the Nanoconfinement Science Gateway
Pegasus and Condor Gaurang Mehta, Ewa Deelman, Carl Kesselman, Karan Vahi Center For Grid Technologies USC/ISI.
University of Southern California
Overview of Workflows: Why Use Them?
Mats Rynge USC Information Sciences Institute
rvGAHP – Push-Based Job Submission Using Reverse SSH Connections
High Throughput Computing for Astronomers
A General Approach to Real-time Workflow Monitoring
Frieda meets Pegasus-WMS
Southern California Earthquake Center
CyberShake Study 18.8 Technical Readiness Review
CyberShake Study 18.8 Planning
Presentation transcript:

Pegasus Workflows on XSEDE Mats Rynge <rynge@isi.edu> USC Information Sciences Institute

XSEDE ECSS Workflow Community Applications Team https://www.xsede.org/web/guest/ecss-workflows The XSEDE Workflow Community Applications Team’s charter is to assist researchers to use scientific workflow technologies on XSEDE to solve challenging scientific problems involving parameter sweeps, multiple applications combined in dependency chains, tightly coupled applications, and similar execution patterns that require multiple applications and multiple XSEDE resources. The workflow team accomplishes its mission through the use of third party workflow software in collaboration with the workflow developers, service providers and XSEDE Extended Collaborative Support Services. Community mailing list: workflows@xsede.org To subscribe, email majordomo@xsede.org with “subscribe workflows” in the body of the message More Information: Marlon Pierce (mpierce@iu.edu) Mats Rynge (rynge@isi.edu) Suresh Marru (smarru@iu.edu)

Pegasus Workflow Management System NSF funded project and developed since 2001 as a collaboration between USC Information Sciences Institute and the HTCondor Team at UW Madison Builds on top of HTCondor DAGMan. Abstract Workflows - Pegasus input workflow description Workflow “high-level language” Only identifies the computation, devoid of resource descriptions, devoid of data locations Pegasus is a workflow “compiler” (plan/map) Target is DAGMan DAGs and HTCondor submit files Transforms the workflow for performance and reliability Automatically locates physical locations for both workflow components and data Collects runtime provenance

Southern California Earthquake Center’s “CyberShake” workflow Description Builders ask seismologists: “What will the peak ground motion be at my new building in the next 50 years?” Seismologists answer this question using Probabilistic Seismic Hazard Analysis (PSHA) Hierarchal workflow with mix of large MPI jobs (top) and large number of serial tasks (post processing) The size of the workflow depends on the number of sites and sampling frequency For each of the 600 sites in the input map, generate a hazard curve Each site has a sub-workflow with 820,000 tasks SGT (Strain Green Tensor - MPI jobs) output: 15.6 TB (40 * 400 GB ) Final outputs: 500 million files (820,000/site x 600 sites ) 5.8 TB (600 * 10 GB ) Probability of exceeding 0.1g in 50 yrs

Master/worker paradigm pegasus-mpi-cluster Master/worker paradigm Master manages the subgraph tasks, handing out work to the workers Efficient scheduling / handling of input/outputs Subgraph described in a DAG-similar format Failure management / rescue DAG

GALACTIC PLANE WORKFLOW Galactic Plane for generating mosaics from the NASA Telescope Missions like Spitzer etc. Used to generate tiles 360 x 40 degrees around the galactic equator A tile 5 x 5 with 1 overlap with neighbors Output datasets to be potentially used in NASA Sky and Google Sky Each per band workflow 1.6 million input files 10.5 million tasks Consumes 34,000 CPU hours Generates 1,001 tiles in FITS format Ran on XSEDE, moved to Amazon EC2 Run workflows corresponding to each of the 16 bands ( wavelengths ) Total Number of Data Files – 18 million Potential Size of Data Output – 45 TB ×16 1 .. 1,001 Acknowledgements: Bruce Berriman, John Good at IPAC - http://montage.ipac.caltech.edu/gallery.html

SoyKB – Soybean Knowledge Base Bioinformatics analysis of 1000+ resequenced soybean germplasm lines selected for major traits including oil, protein, soybean cyst nematode resistance (SCN), abiotic stress resistance (drought, heat and salt) and root system architecture. Sequencing pipeline using standard tools: Picard, BWA, GATK Environment iPlant Data Store for inputs and outputs iPlant Atmosphere for VMs for workflow management TACC Stampede as execution environment

The green partition boxes shows user defined partitions of the workflow which are used when mapping the workflow to TACC Stampede. The walltime of a partition will not exceed the 48 hour wall clock limit, given a certain number of compute nodes.

The green jobs are MPI container jobs of the partitions shown in the earlier slide. These are run on the remote supercomputer. The blue auxiliary tasks are run on the workflow submit host, and handle things like data staging and cleanup.

SoyKB - Highlights Workflow can be mapped to a distributed or a supercomputer execution environment Distributed environment is used for testing the pipeline The supercomputer environment is used for production runs, and tasks are packed up in MPI container jobs (pegasus-mpi-cluster) Input data is fetched from iPlant Data Store, or if already replicated, from the TACC iPlant Data Store node for close to computation access Outputs are automatically put back into the Data Store for easy access and further analysis in the iPlant Discovery Environment

Relevant Links http://pegasus.isi.edu Tutorial and documentation: http://pegasus.isi.edu/wms/docs/latest/ Mats Rynge rynge@isi.edu