Experiment Management from a Pegasus Perspective Jens-S. Vöckler Ewa Deelman 2011-11-181.

Slides:



Advertisements
Similar presentations
Sections 2 and 3 Chapter 1. Review of the Scientific Method The scientific method is not a list of rules that must be followed but a general guideline.
Advertisements

3rd Campus Grid SIG Meeting. Agenda Welcome OMII Requirements document Grid Data Group HTC Workshop Research Computing SIG? AOB Next meeting (AG)
What is Science?.
Pegasus on the Virtual Grid: A Case Study of Workflow Planning over Captive Resources Yang-Suk Kee, Eun-Kyu Byun, Ewa Deelman, Kran Vahi, Jin-Soo Kim Oracle.
Phillip Dickens, Department of Computer Science, University of Maine. In collaboration with Jeremy Logan, Postdoctoral Research Associate, ORNL. Improving.
Natasha Pavlovikj, Kevin Begcy, Sairam Behera, Malachy Campbell, Harkamal Walia, Jitender S.Deogun University of Nebraska-Lincoln Evaluating Distributed.
Managing Workflows Within HUBzero: How to Use Pegasus to Execute Computational Pipelines Ewa Deelman USC Information Sciences Institute Acknowledgement:
The ADAMANT Project: Linking Scientific Workflows and Networks “Adaptive Data-Aware Multi-Domain Application Network Topologies” Ilia Baldine, Charles.
Ewa Deelman, Integrating Existing Scientific Workflow Systems: The Kepler/Pegasus Example Nandita Mangal,
Nadia Ranaldo - Eugenio Zimeo Department of Engineering University of Sannio – Benevento – Italy 2008 ProActive and GCM User Group Orchestrating.
Business Continuity and DR, A Practical Implementation Mich Talebzadeh, Consultant, Deutsche Bank
Workload Management Workpackage Massimo Sgaravatto INFN Padova.
6th Biennial Ptolemy Miniconference Berkeley, CA May 12, 2005 Distributed Computing in Kepler Ilkay Altintas Lead, Scientific Workflow Automation Technologies.
Workload Management Massimo Sgaravatto INFN Padova.
Functional Modeling Chapter 6.
Managing Workflows with the Pegasus Workflow Management System
CSE 486/586 CSE 486/586 Distributed Systems PA Best Practices Steve Ko Computer Sciences and Engineering University at Buffalo.
CONDOR DAGMan and Pegasus Selim Kalayci Florida International University 07/28/2009 Note: Slides are compiled from various TeraGrid Documentations.
Computer System Architectures Computer System Software
Scientific Workflow reusing and long term big data preservation Salima Benbernou Université Paris Descartes Project.
The Grid is a complex, distributed and heterogeneous execution environment. Running applications requires the knowledge of many grid services: users need.
Sarah Drummond Dept. Computer Science University of Durham, UK MSc Research An Investigation into Computer Support for Cooperative Work in Software Engineering.
10/20/05 LIGO Scientific Collaboration 1 LIGO Data Grid: Making it Go Scott Koranda University of Wisconsin-Milwaukee.
Integrating HPC into the ATLAS Distributed Computing environment Doug Benjamin Duke University.
What are the main differences and commonalities between the IS and DA systems? How information is transferred between tasks: (i) IS it may be often achieved.
Managing large-scale workflows with Pegasus Karan Vahi ( Collaborative Computing Group USC Information Sciences Institute Funded.
Experiences Using Cloud Computing for A Scientific Workflow Application Jens Vöckler, Gideon Juve, Ewa Deelman, Mats Rynge, G. Bruce Berriman Funded by.
Building Scalable Scientific Applications using Makeflow Dinesh Rajan and Douglas Thain University of Notre Dame.
Distributed Aircraft Maintenance Environment - DAME DAME Workflow Advisor Max Ong University of Sheffield.
Scientific Workflow Scheduling in Computational Grids Report: Wei-Cheng Lee 8th Grid Computing Conference IEEE 2007 – Planning, Reservation,
Dr. Ahmed Abdeen Hamed, Ph.D. University of Vermont, EPSCoR Research on Adaptation to Climate Change (RACC) Burlington Vermont USA MODELING THE IMPACTS.
Workflow Project Status Update Luciano Piccoli - Fermilab, IIT Nov
Combining the strengths of UMIST and The Victoria University of Manchester Adaptive Workflow Processing and Execution in Pegasus Kevin Lee School of Computer.
 Apache Airavata Architecture Overview Shameera Rathnayaka Graduate Assistant Science Gateways Group Indiana University 07/27/2015.
Pegasus: Running Large-Scale Scientific Workflows on the TeraGrid Ewa Deelman USC Information Sciences Institute
Intermediate Condor: Workflows Rob Quick Open Science Grid Indiana University.
What is SAM-Grid? Job Handling Data Handling Monitoring and Information.
1 P-GRADE Portal: a workflow-oriented generic application development portal Peter Kacsuk MTA SZTAKI, Hungary Univ. of Westminster, UK.
Review of Condor,SGE,LSF,PBS
Theme 2: Data & Models One of the central processes of science is the interplay between models and data Data informs model generation and selection Models.
Scientific Workflow systems: Summary and Opportunities for SEEK and e-Science.
Pegasus WMS: Leveraging Condor for Workflow Management Ewa Deelman, Gaurang Mehta, Karan Vahi, Gideon Juve, Mats Rynge, Prasanth.
Intermediate Condor: Workflows Monday, 1:15pm Alain Roy OSG Software Coordinator University of Wisconsin-Madison.
LIGO Plans for OSG J. Kent Blackburn LIGO Laboratory California Institute of Technology Open Science Grid Technical Meeting UCSD December 15-17, 2004.
What is Science? SECTION 1.1. What Is Science and Is Not  Scientific ideas are open to testing, discussion, and revision  Science is an organize way.
Funded by the NSF OCI program grants OCI and OCI Mats Rynge, Gideon Juve, Karan Vahi, Gaurang Mehta, Ewa Deelman Information Sciences Institute,
Lecture On Introduction (DBMS) By- Jesmin Akhter Assistant Professor, IIT, Jahangirnagar University.
1 Pegasus and wings WINGS/Pegasus Provenance Challenge Ewa Deelman Yolanda Gil Jihie Kim Gaurang Mehta Varun Ratnakar USC Information Sciences Institute.
PARALLEL AND DISTRIBUTED PROGRAMMING MODELS U. Jhashuva 1 Asst. Prof Dept. of CSE om.
1 USC Information Sciences InstituteYolanda Gil AAAI-08 Tutorial July 13, 2008 Part IV Workflow Mapping and Execution in Pegasus (Thanks.
Meeting with University of Malta| CERN, May 18, 2015 | Predrag Buncic ALICE Computing in Run 2+ P. Buncic 1.
Managing LIGO Workflows on OSG with Pegasus Karan Vahi USC Information Sciences Institute
EGI-InSPIRE RI EGI Compute and Data Services for Open Access in H2020 Tiziana Ferrari Technical Director, EGI.eu
HUBzero® Platform for Scientific Collaboration Copyright © 2012 HUBzero Foundation, LLC International Workshop on Science Gateways, ETH Zürich, June 3-5,
Accessing the VI-SEEM infrastructure
Workload Management Workpackage
Component and Deployment Diagrams
Pegasus WMS Extends DAGMan to the grid world
Cloudy Skies: Astronomy and Utility Computing
Workflows and the Pegasus Workflow Management System
US CMS Testbed.
University of Southern California
Clouds from FutureGrid’s Perspective
Laura Bright David Maier Portland State University
Overview of Workflows: Why Use Them?
Mats Rynge USC Information Sciences Institute
A General Approach to Real-time Workflow Monitoring
Frieda meets Pegasus-WMS
Scientific Workflows Lecture 15
Presentation transcript:

Experiment Management from a Pegasus Perspective Jens-S. Vöckler Ewa Deelman

Outline You know what FutureGrid is, but… What is an Experiment? What is Pegasus? How do the two connect? What extra are we building?

What Is a Scientific Experiment? 1.Create a hypothesis 2.Design an experiment to prove or disprove – Document your setup (apparatus) 3.Run and observe (be ready to be surprised) – Ensure sufficient sensors (placement, granularity) – Document all observations (report) 4.Draw conclusions (paper, publication) – Others should be able to repeat the experiment in a nutshell

Experiments Using Computer Science The apparatus is often a (set of) program(s) and execution environment from the domain science The experiment often involves: 1.Processing massive data with same code (proudly parallel) 2.Complex processing in dependent steps (workflow) Sensors often constitute log files and monitoring Pegasus is set to deal well with all of the above

Pegasus Workflow Management System Developed since 2001 A collaboration between USC and the Condor Team at UW Madison (includes DAGMan) Used by a number of applications in a variety of domains Provides reliability – can retry computations from the point of failure Provides scalability – can handle large data (kByte…TB of data), – and many computations (1…10 6 tasks)

Pegasus Workflow Management System Automatically captures provenance information Can run on resources distributed among institutions, laptop, campus cluster (HPC), Grid, Cloud Enables the construction of complex workflows based on computational blocks Infers data transfers Infers data registrations ☞ apparatus, sensors

Pegasus WMS Provides a portable and re-usable workflow description Lives in user-space Provides correct, scalable, and reliable execution – Enforces dependencies between tasks – Progresses as far as possible in the face of failures Pegasus makes use of available resources, but cannot control them ☞ experiment, repeatability

Pegasus WMS runs Experiments Workflows capture hypotheses – Abstract description independent of apparatus – Can be shared to repeat experiments Multiple concurrent experiments – Automatic batch-style execution Provenance captures apparatus – Provenance helper kickstart to aide sensors

Room for Improvement Support for interactive steps – Need to separate into multiple workflows Formal apparatus description – Provenance is necessary but not sufficient Standardized sensor classes – More sensors, better resolution – Tie-ins with monitoring systems, etc. Repository of Experiments – Sharing of DAX is ad-hoc for now

Next Steps Design repository for experiments – Attempt to make it useful to all FG EM efforts Improve capture of apparatus description, Improve capture of sensor data – Work with FG-Performance group – Might be useful beyond FG EM – Big disk AMQP sink for multiple sensor streams Can be used to create repeatable experiments differently Is there something you, the audience, would like to see for FG Experiment Management?