- DAG Scheduling with Reliability - - GridSolve - - Fault Tolerance In Open MPI - Asim YarKhan, Zhiao Shi, Jack Dongarra VGrADS Workshop April 2007.

Slides:



Advertisements
Similar presentations
Multiple Processor Systems
Advertisements

A Workflow Engine with Multi-Level Parallelism Supports Qifeng Huang and Yan Huang School of Computer Science Cardiff University
Energy-efficient Task Scheduling in Heterogeneous Environment 2013/10/25.
Distributed Processing, Client/Server and Clusters
Distributed Systems Major Design Issues Presented by: Christopher Hector CS8320 – Advanced Operating Systems Spring 2007 – Section 2.6 Presentation Dr.
Scheduling in Distributed Systems Gurmeet Singh CS 599 Lecture.
Resource Management §A resource can be a logical, such as a shared file, or physical, such as a CPU (a node of the distributed system). One of the functions.
Master/Slave Architecture Pattern Source: Pattern-Oriented Software Architecture, Vol. 1, Buschmann, et al.

A system Performance Model Instructor: Dr. Yanqing Zhang Presented by: Rajapaksage Jayampthi S.
GridRPC Sources / Credits: IRISA/IFSIC IRISA/INRIA Thierry Priol et. al papers.
All Hands Meeting, 2006 Title: Grid Workflow Scheduling in WOSE (Workflow Optimisation Services for e- Science Applications) Authors: Yash Patel, Andrew.
Distributed Process Scheduling Summery Distributed Process Scheduling Summery BY:-Yonatan Negash.
Objektorienteret Middleware Presentation 2: Distributed Systems – A brush up, and relations to Middleware, Heterogeneity & Transparency.
Approaches to EJB Replication. Overview J2EE architecture –EJB, components, services Replication –Clustering, container, application Conclusions –Advantages.
Distributed Processing, Client/Server, and Clusters
Distributed components
Network Management Overview IACT 918 July 2004 Gene Awyzio SITACS University of Wollongong.
Experiments with SmartGridSolve: Achieving Higher Performance by Improving the GridRPC Model Thomas Brady, Michele Guidolin, Alexey Lastovetsky Heterogeneous.
A Grid Parallel Application Framework Jeremy Villalobos PhD student Department of Computer Science University of North Carolina Charlotte.
Computer Science Department 1 Load Balancing and Grid Computing David Finkel Computer Science Department Worcester Polytechnic Institute.
Smart Redundancy for Distributed Computation George Edwards Blue Cell Software, LLC Yuriy Brun University of Washington Jae young Bang University of Southern.
Copyright © The McGraw-Hill Companies, Inc. Permission required for reproduction or display. Parallel Programming in C with MPI and OpenMP Michael J. Quinn.
Present by Chen, Ting-Wei Adaptive Task Checkpointing and Replication: Toward Efficient Fault-Tolerant Grids Maria Chtepen, Filip H.A. Claeys, Bart Dhoedt,
Workload Management Massimo Sgaravatto INFN Padova.
.NET Mobile Application Development Introduction to Mobile and Distributed Applications.
SIMULATING ERRORS IN WEB SERVICES International Journal of Simulation: Systems, Sciences and Technology 2004 Nik Looker, Malcolm Munro and Jie Xu.
Speaker: Xin Zuo Heterogeneous Computing Laboratory (HCL) School of Computer Science and Informatics University College Dublin Ireland International Parallel.
Research on cloud computing application in the peer-to-peer based video-on-demand systems Speaker : 吳靖緯 MA0G rd International Workshop.
Self Adaptivity in Grid Computing Reporter : Po - Jen Lo Sathish S. Vadhiyar and Jack J. Dongarra.
Checkpoint & Restart for Distributed Components in XCAT3 Sriram Krishnan* Indiana University, San Diego Supercomputer Center & Dennis Gannon Indiana University.
 Escalonamento e Migração de Recursos e Balanceamento de carga Carlos Ferrão Lopes nº M6935 Bruno Simões nº M6082 Celina Alexandre nº M6807.
DISTRIBUTED COMPUTING
STRATEGIC NAMING: MULTI-THREADED ALGORITHM (Ch 27, Cormen et al.) Parallelization Four types of computing: –Instruction (single, multiple) per clock cycle.
Active Monitoring in GRID environments using Mobile Agent technology Orazio Tomarchio Andrea Calvagna Dipartimento di Ingegneria Informatica e delle Telecomunicazioni.
Chapter 3 Parallel Algorithm Design. Outline Task/channel model Task/channel model Algorithm design methodology Algorithm design methodology Case studies.
RELATIONAL FAULT TOLERANT INTERFACE TO HETEROGENEOUS DISTRIBUTED DATABASES Prof. Osama Abulnaja Afraa Khalifah
DORII Joint Research Activities DORII Joint Research Activities Status and Progress 6 th All-Hands-Meeting (AHM) Alexey Cheptsov on.
A Survey of Distributed Task Schedulers Kei Takahashi (M1)
The Grid System Design Liu Xiangrui Beijing Institute of Technology.
Recent Advances in SmartGridSolve Oleg Girko School of Computer Science & Informatics University College Dublin.
Stochastic DAG Scheduling using Monte Carlo Approach Heterogeneous Computing Workshop (at IPDPS) 2012 Extended version: Elsevier JPDC (accepted July 2013,
A Fault Tolerant Protocol for Massively Parallel Machines Sayantan Chakravorty Laxmikant Kale University of Illinois, Urbana-Champaign.
Tool Integration with Data and Computation Grid GWE - “Grid Wizard Enterprise”
Issues Autonomic operation (fault tolerance) Minimize interference to applications Hardware support for new operating systems Resource management (global.
NIH Resource for Biomolecular Modeling and Bioinformatics Beckman Institute, UIUC NAMD Development Goals L.V. (Sanjay) Kale Professor.
Static Process Scheduling Section 5.2 CSc 8320 Alex De Ruiter
 Apache Airavata Architecture Overview Shameera Rathnayaka Graduate Assistant Science Gateways Group Indiana University 07/27/2015.
Summary :-Distributed Process Scheduling Prepared By:- Monika Patel.
Fault Tolerant Extensions to Charm++ and AMPI presented by Sayantan Chakravorty Chao Huang, Celso Mendes, Gengbin Zheng, Lixia Shi.
6. A PPLICATION MAPPING 6.3 HW/SW partitioning 6.4 Mapping to heterogeneous multi-processors 1 6. Application mapping (part 2)
Lecture 4 TTH 03:30AM-04:45PM Dr. Jianjun Hu CSCE569 Parallel Computing University of South Carolina Department of.
Automatic Statistical Evaluation of Resources for Condor Daniel Nurmi, John Brevik, Rich Wolski University of California, Santa Barbara.
Toward interactive visualization in a distributed workflow Steven G. Parker Oscar Barney Ayla Khan Thiago Ize Steven G. Parker Oscar Barney Ayla Khan Thiago.
Static Process Scheduling
Parallelization Strategies Laxmikant Kale. Overview OpenMP Strategies Need for adaptive strategies –Object migration based dynamic load balancing –Minimal.
Tool Integration with Data and Computation Grid “Grid Wizard 2”
Pradeep Konduri Static Process Scheduling:  Proceedance process model  Communication system model  Application  Dicussion.
VGrADS and GridSolve Asim YarKhan Jack Dongarra, Zhiao Shi, Fengguang Song Innovative Computing Laboratory University of Tennessee VGrADS Workshop – September.
Robust Task Scheduling in Non-deterministic Heterogeneous Computing Systems Zhiao Shi Asim YarKhan, Jack Dongarra Followed by GridSolve, FT-MPI, Open MPI.
Towards a High Performance Extensible Grid Architecture Klaus Krauter Muthucumaru Maheswaran {krauter,
Workload Management Workpackage
Parallel processing is not easy
Experiments with Fault Tolerant Linear Algebra Algorithms
Abstract Machine Layer Research in VGrADS
Steven Whitham Jeremy Woods
Parallel Programming in C with MPI and OpenMP
Atlas: An Infrastructure for Global Computing
Gordon Erlebacher Florida State University
Parallel Programming in C with MPI and OpenMP
Presentation transcript:

- DAG Scheduling with Reliability - - GridSolve - - Fault Tolerance In Open MPI - Asim YarKhan, Zhiao Shi, Jack Dongarra VGrADS Workshop April 2007

Task Graph Scheduling with Reliability Scheduling task graphs in a heterogeneous system is proven to be NP-hard With the increasing size of the system, the reliability of the system should be addressed –Application Failure prevention –Task duplication –Checkpointing –Application Failure avoidance –Guarantee that the probability of application failure is kept as low as possible Jack Dongarra, Emannuel Jeannot, Erik Saule, Zhiao Shi, Bi-objective Scheduling Algorithms for Optimizing Makespan and Reliability on Heterogeneous Systems, Published in Proceedings of 19th Annual ACM Symposium on Parallelism in Algorithms and Architectures (SPAA '07)

Bi-objective Optimization Problem Task graph G=(V,E), a DAG –Task v i  V has o i operations and |V|= n –e i = e(i,j)  E has associated l i (communication cost from v i to v j ) Processor p j  P, |P| =m, –  j is the unit execution time on p j – j is the failure rate of p j, constant failure model –Success rate of task i (with o i ops) on processor j is Failure model follows exponential law C j max is the completion time on processor j The success rate of the application is Goal: minimizing the makespan and maximizing p succ. –conflicting objectives –optimizing one objective will adversely affect the other

Generalizing Scheduling Heuristics HEFT (Heterogeneous Earliest Finish Time) –Well known task graph scheduling heuristic –At each scheduling step, assign task to processors using Earliest Finish Time (EFT) RHEFT (Reliable HEFT) to minimize failure costs –At each step assign task i to to processor j that has the minimum product T endj * j –Easily extended to other heuristics –T end1 = 6; T end1 * 1 = 12 –T end2 = 9; T end2 * 2 = 9

Reliability/Makespan Tradeoff Choose a subset of processors (by some ordering) –Choose k out of 100 processors ordering the processors: reliable first -> high makespan for low num proc ordering the processors: fastest first -> increasing reliability as num proc increases to 15

Reliability/Makespan Tradeoff Tradeoff between HEFT and RHEFT ordering the processors: smallest  first -> makespan OK and reliability decreases from optimal at 1 Tradeoff variable  at k = 100  =0  RHEFT  =1  HEFT

Task Graphs: Conclusion Studied problem of optimizing DAGs with 2 objectives –Minimize makespan –Maximize reliability Created a simple way to generalize heuristics (e.g. HEFT to RHEFT) to allow for reliability Characterized the role of  as a way to allow user to tradeoff reliability and makespan by choosing a subset of processors

GridSolve Grid based, client-agent- server, RPC system –Resource discovery, dynamic problem solving capabilities, load balancing, fault tolerance, asynchronous calls, disconnected operation, network data storage,... –Dynamic service bindings –Client does not need to have stubs for services –API provides a variety of methods –Blocking, non-blocking, task farms, disconnected,... –Easy to add additional services by wrapping libraries Prime goal: ease-of-use –Supports proposed standard GridRPC client API –Matlab, Octave, IDL: enable desktop based Grid computing

GridSolve: Scheduling Joint work with Emmanuel Jeannot, INRIA –History based performance estimation –Uses historical performance of specific problem on a specific server with a template based linear model to make more accurate service performance model –Communication cost estimates –Client estimates communication costs for a subset of servers via a simple probe –Perturbation model for scheduling –The agent uses a interaction model of the currently executing jobs on the servers to schedule jobs (includes estimated completion times) –Accounts for the effect of one job on another Emmanuel Jeannot and Keith Seymour and Asim YarKhan and Jack J. Dongarra, Improved Runtime and Transfer Time Prediction Mechanisms in a Network Enabled Servers Middleware, Parallel Processing Letters, 2007

GridSolve: Workflow support Work in progress, with Alexey Lastovetsky (UC Dublin) et al GridSolve is being enhanced to support workflow –Described by a DAG submitted at the client –DAG parsed by client into GridSolve calls –Data transferred from server to server Scheduling is currently static –Done at DAG submission time, uses agent knowledge to create a smart mapping of the services to the resources

GridSolve: Summary Beta versions of GridSolve have been released –Currently v0.15 –New version with scheduling options to be released soon Integration Possibility –As in NetSolve/GrADS –vgES can act as a resource manager for the GridSolve agent –Software and data are migrated on demand GrADS/NetSolve integration for SC05

Open MPI Open source (BSD style) MPI-2 implementation (v 1.2) Thread safe Dynamic process spawning Uses multiple network interfaces simultaneously via a single library Tunable (e.g. collective operations can be tuned to platform) In development: Network and process fault tolerance

Open MPI: Fault Tolerance Notes: blue=not maintained; white=still functional; red=maintained and to be incorporated in Open MPI

Open MPI: Fault Tolerance Open MPI plans on supporting the following fault tolerance techniques: –LAM/MPI style automatic, coordinated process checkpoint and restart (expected SC 07) –MPICH-V style automatic, uncoordinated process checkpoint with message logging techniques (expected SC 07) –User directed, and communicator driven fault tolerance. Similar to those implemented in FT-MPI (expected 2008)

The End