A Survey of Distributed Task Schedulers Kei Takahashi (M1)

Slides:



Advertisements
Similar presentations
Energy-efficient Task Scheduling in Heterogeneous Environment 2013/10/25.
Advertisements

Load Balancing Parallel Applications on Heterogeneous Platforms.
Scheduling in Distributed Systems Gurmeet Singh CS 599 Lecture.
Dynamic Thread Assignment on Heterogeneous Multiprocessor Architectures Pree Thiengburanathum Advanced computer architecture Oct 24,
Resource Management §A resource can be a logical, such as a shared file, or physical, such as a CPU (a node of the distributed system). One of the functions.
1 An Adaptive GA for Multi Objective Flexible Manufacturing Systems A. Younes, H. Ghenniwa, S. Areibi uoguelph.ca.
Using Parallel Genetic Algorithm in a Predictive Job Scheduling
ISE480 Sequencing and Scheduling Izmir University of Economics ISE Fall Semestre.
A system Performance Model Instructor: Dr. Yanqing Zhang Presented by: Rajapaksage Jayampthi S.
GridFlow: Workflow Management for Grid Computing Kavita Shinde.
Network Coding for Large Scale Content Distribution Christos Gkantsidis Georgia Institute of Technology Pablo Rodriguez Microsoft Research IEEE INFOCOM.
Security-Driven Heuristics and A Fast Genetic Algorithm for Trusted Grid Job Scheduling Shanshan Song, Ricky Kwok, and Kai Hwang University of Southern.
Grid Load Balancing Scheduling Algorithm Based on Statistics Thinking The 9th International Conference for Young Computer Scientists Bin Lu, Hongbin Zhang.
On Fairness, Optimizing Replica Selection in Data Grids Husni Hamad E. AL-Mistarihi and Chan Huah Yong IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS,
A New Approach for Task Level Computational Resource Bi-Partitioning Gang Wang, Wenrui Gong, Ryan Kastner Express Lab, Dept. of ECE, University of California,
Radial Basis Function Networks
1 Reasons for parallelization Can we make GA faster? One of the most promising choices is to use parallel implementations. The reasons for parallelization.
Authors: Weiwei Chen, Ewa Deelman 9th International Conference on Parallel Processing and Applied Mathmatics 1.
A Budget Constrained Scheduling of Workflow Applications on Utility Grids using Genetic Algorithms Jia Yu and Rajkumar Buyya Grid Computing and Distributed.
Load Balancing and Termination Detection Load balance : - statically before the execution of any processes - dynamic during the execution of the processes.
Load Balancing in Distributed Computing Systems Using Fuzzy Expert Systems Author Dept. Comput. Eng., Alexandria Inst. of Technol. Content Type Conferences.
 Escalonamento e Migração de Recursos e Balanceamento de carga Carlos Ferrão Lopes nº M6935 Bruno Simões nº M6082 Celina Alexandre nº M6807.
Network Aware Resource Allocation in Distributed Clouds.
Generative Topographic Mapping in Life Science Jong Youl Choi School of Informatics and Computing Pervasive Technology Institute Indiana University
An Autonomic Framework in Cloud Environment Jiedan Zhu Advisor: Prof. Gagan Agrawal.
ROBUST RESOURCE ALLOCATION OF DAGS IN A HETEROGENEOUS MULTI-CORE SYSTEM Luis Diego Briceño, Jay Smith, H. J. Siegel, Anthony A. Maciejewski, Paul Maxwell,
1 Distributed Process Scheduling: A System Performance Model Vijay Jain CSc 8320, Spring 2007.
Young Suk Moon Chair: Dr. Hans-Peter Bischof Reader: Dr. Gregor von Laszewski Observer: Dr. Minseok Kwon 1.
DLS on Star (Single-level tree) Networks Background: A simple network model for DLS is the star network with a master-worker platform. It consists of a.
Optimization Problems - Optimization: In the real world, there are many problems (e.g. Traveling Salesman Problem, Playing Chess ) that have numerous possible.
1 Distributed Energy-Efficient Scheduling for Data-Intensive Applications with Deadline Constraints on Data Grids Cong Liu and Xiao Qin Auburn University.
Euro-Par, A Resource Allocation Approach for Supporting Time-Critical Applications in Grid Environments Qian Zhu and Gagan Agrawal Department of.
Stochastic DAG Scheduling using Monte Carlo Approach Heterogeneous Computing Workshop (at IPDPS) 2012 Extended version: Elsevier JPDC (accepted July 2013,
1 A distributed Task Scheduler Optimizing Data Transfer Time Taura lab. Kei Takahashi (56428) Taura lab. Kei Takahashi (56428)
1 Short Term Scheduling. 2  Planning horizon is short  Multiple unique jobs (tasks) with varying processing times and due dates  Multiple unique jobs.
1 ACTIVE FAULT TOLERANT SYSTEM for OPEN DISTRIBUTED COMPUTING (Autonomic and Trusted Computing 2006) Giray Kömürcü.
Resource Mapping and Scheduling for Heterogeneous Network Processor Systems Liang Yang, Tushar Gohad, Pavel Ghosh, Devesh Sinha, Arunabha Sen and Andrea.
1 A distributed Task Scheduler Optimizing Data Transfer Time (データ転送時間を最適化する分散タスクスケジュラー) Taura lab. Kei Takahashi (56428) Taura lab. Kei Takahashi (56428)
O PTIMAL SERVICE TASK PARTITION AND DISTRIBUTION IN GRID SYSTEM WITH STAR TOPOLOGY G REGORY L EVITIN, Y UAN -S HUN D AI Adviser: Frank, Yeong-Sung Lin.
MROrder: Flexible Job Ordering Optimization for Online MapReduce Workloads School of Computer Engineering Nanyang Technological University 30 th Aug 2013.
A Hyper-heuristic for scheduling independent jobs in Computational Grids Author: Juan Antonio Gonzalez Sanchez Coauthors: Maria Serna and Fatos Xhafa.
CS 484 Load Balancing. Goal: All processors working all the time Efficiency of 1 Distribute the load (work) to meet the goal Two types of load balancing.
Data Structures and Algorithms in Parallel Computing Lecture 7.
Static Process Scheduling
Efficient Load Balancing Algorithm for Cloud Computing Network Che-Lun Hung 1, Hsiao-hsi Wang 2 and Yu-Chen Hu 2 1 Dept. of Computer Science & Communication.
Scalable and Topology-Aware Load Balancers in Charm++ Amit Sharma Parallel Programming Lab, UIUC.
Scheduling MPI Workflow Applications on Computing Grids Juemin Zhang, Waleed Meleis, and David Kaeli Electrical and Computer Engineering Department, Northeastern.
A Fast Genetic Algorithm Based Static Heuristic For Scheduling Independent Tasks on Heterogeneous Systems Gaurav Menghani Department of Computer Engineering,
3/12/2013Computer Engg, IIT(BHU)1 INTRODUCTION-1.
Data Consolidation: A Task Scheduling and Data Migration Technique for Grid Networks Author: P. Kokkinos, K. Christodoulopoulos, A. Kretsis, and E. Varvarigos.
1 Low Latency Multimedia Broadcast in Multi-Rate Wireless Meshes Chun Tung Chou, Archan Misra Proc. 1st IEEE Workshop on Wireless Mesh Networks (WIMESH),
A Stable Broadcast Algorithm Kei Takahashi Hideo Saito Takeshi Shibata Kenjiro Taura (The University of Tokyo, Japan) 1 CCGrid Lyon, France.
A stochastic scheduling algorithm for precedence constrained tasks on Grid Future Generation Computer Systems (2011) Xiaoyong Tang, Kenli Li, Guiping Liao,
Euro-Par, HASTE: An Adaptive Middleware for Supporting Time-Critical Event Handling in Distributed Environments ICAC 2008 Conference June 2 nd,
Scheduling Strategies for Mapping Application Workflows Onto the Grid A. Mandal, K. Kennedy, C. Koelbel, G. Marin, J. Mellor- Crummey, B. Liu, L. Johnsson.
- 세부 1 - 이종 클라우드 플랫폼 데이터 관리 브로커 연구 및 개발 Network and Computing Lab.
1 Performance Impact of Resource Provisioning on Workflows Gurmeet Singh, Carl Kesselman and Ewa Deelman Information Science Institute University of Southern.
Jérémie Sublime Sonia Yassa Development of meta-heuristics for workflow scheduling based on quality of service requirements 1.
Scheduling Algorithms Performance Evaluation in Grid Environments R, Zhang, C. Koelbel, K. Kennedy.
Parallel processing is not easy
New Workflow Scheduling Techniques Presentation: Anirban Mandal
Load Balancing and It’s Related Works in Cloud Computing
A Dynamic Critical Path Algorithm for Scheduling Scientific Workflow Applications on Global Grids e-Science IEEE 2007 Report: Wei-Cheng Lee
Efficient Load Balancing Algorithm for Cloud
On Scheduling in Map-Reduce and Flow-Shops
湖南大学-信息科学与工程学院-计算机与科学系
Navigation In Dynamic Environment
Performance-Robust Parallel I/O
R, Zhang, A. Chien, A. Mandal, C. Koelbel,
MapReduce: Simplified Data Processing on Large Clusters
Presentation transcript:

A Survey of Distributed Task Schedulers Kei Takahashi (M1)

2 What do you want to do on a grid?  Vast computing resources  Calculation power  Memory  Data storage  Large scale computation  Numerical simulations  Statistical analyses  Data mining.. for everyone  Vast computing resources  Calculation power  Memory  Data storage  Large scale computation  Numerical simulations  Statistical analyses  Data mining.. for everyone

3 Grid Applications  For some applications, it is inevitable to develop parallel algorithms  Dedicated to parallel environment  E.g. matrix computations  However, many applications are efficiently sped up by simply running multiple serial programs in parallel  E.g. many data intensive applications  For some applications, it is inevitable to develop parallel algorithms  Dedicated to parallel environment  E.g. matrix computations  However, many applications are efficiently sped up by simply running multiple serial programs in parallel  E.g. many data intensive applications

4 Grid Schedulers  A system which distributes many serial tasks onto the grid environment  Task assignments  File transfers  A user need not rewrite serial programs to execute them in parallel  Some constraints need to be considered  Machine availability  Machine spec (CPU/Memory/HDD), load  Data location  Task priority  A system which distributes many serial tasks onto the grid environment  Task assignments  File transfers  A user need not rewrite serial programs to execute them in parallel  Some constraints need to be considered  Machine availability  Machine spec (CPU/Memory/HDD), load  Data location  Task priority

5 An Example of Scheduling  Each task is assigned to a machine A (fast) A (fast) B (slow) B (slow) Scheduler Task t0 Heavy Task t0 Heavy Task t1 Light Task t1 Light Task t2 Light Task t2 Light t0 t1 t2 A B t0 t2 A B t1 Shorter processing time

6 Efficient Scheduling  Task scheduling in heterogeneous environment is not a new problem. Some heuristics are already proposed.  However, existing algorithms could not appropriately handle some situations  Data intensive applications  Workflows  Task scheduling in heterogeneous environment is not a new problem. Some heuristics are already proposed.  However, existing algorithms could not appropriately handle some situations  Data intensive applications  Workflows

7 Data Intensive Applications  A computation using large data  Some gigabytes to petabytes  A scheduler need to consider the followings:  File transfer need to be diminished  Data replica should be effectively placed  Unused intermediate files should be cleared  A computation using large data  Some gigabytes to petabytes  A scheduler need to consider the followings:  File transfer need to be diminished  Data replica should be effectively placed  Unused intermediate files should be cleared

8 An Example of Scheduling  Each task is assigned to a machine A (fast) A (fast) B (slow) B (slow) Scheduler Task t0 Heavy Requires : f0 Task t0 Heavy Requires : f0 Task t1 Light Requires : f1 Task t1 Light Requires : f1 Task t2 Light Requires : f1 Task t2 Light Requires : f1 File f0 Large File f0 Large File f1 Small File f1 Small t0 t1 t2 A B f0 f1 t0 t2 A B t1 f1 Shorter processing time

9 Workflow  A set of tasks with dependencies  Data dependency between some tasks  Expressed by a DAG  A set of tasks with dependencies  Data dependency between some tasks  Expressed by a DAG Corpus Phrases (by words) Corpus Parsed Corpus Parsed Corpus Parsed Corpus Phrases (by words) Phrases (by words) Cooccurrence analysis Cooccurrence analysis Coocurrence analysis

10 Workflow (cont.)  Workflow is suitable for expressing some grid applications  Only necessary dependency is described by a workflow  A scheduler can adaptively map tasks to the real node environment  More factors to consider  Some tasks are important to shorten the overall makespan  Workflow is suitable for expressing some grid applications  Only necessary dependency is described by a workflow  A scheduler can adaptively map tasks to the real node environment  More factors to consider  Some tasks are important to shorten the overall makespan

11 Agenda  Introduction  Basic Scheduling Algorithms  Some heuristics  Data-intensive/Workflow Schedulers  Conclusion  Introduction  Basic Scheduling Algorithms  Some heuristics  Data-intensive/Workflow Schedulers  Conclusion

12 Basic Scheduling Heuristics  Given information :  ETC (expected completion time) for each pair of a node and a task, including data transfer cost  No congestion is assumed  Aim : minimizing the makespan (Total processing time)  Given information :  ETC (expected completion time) for each pair of a node and a task, including data transfer cost  No congestion is assumed  Aim : minimizing the makespan (Total processing time) [1] Tracy et al. A Comparison Study of Eleven Static Heuristics for Mapping a Class of Independent Tasks onto Heterogeneous Distributed Computing Systems (TR-ECE 00-04)

13 An example of ETC  ETC of (task, node) = (node available time) + (data transfer time) + (task process time) Available afterTransferProcessETC Node A200 (sec)10 (sec)100 (sec)310 (sec) Node B0 (sec) 100 (sec) Node C0 (sec)100 (sec)20 (sec)120 (sec) A A B B Data 1GB C C 1Gbps 100Mbps

14 Scheduling algorithms  An ETC matrix is given  When a task is assigned to a node, the ETC matrix is updated  An ETC matrix is consistent { if node M0 can process a task faster than M1, M0 can process every other task faster than M }  The makespan of an inconsistent ETC matrix differs more than that of a consistent ETC matrix  An ETC matrix is given  When a task is assigned to a node, the ETC matrix is updated  An ETC matrix is consistent { if node M0 can process a task faster than M1, M0 can process every other task faster than M }  The makespan of an inconsistent ETC matrix differs more than that of a consistent ETC matrix Task 0Task 1Task 2 Node A862 Node B193 Node C Assigned to A

15 Greedy approaches  Principles  Assign a task to the best node at a time  Need to decide the order of tasks  Scheduling priority  Min-min : Light task  Max-min : Heavy task  Sufferage : A task whose completion time differs most depending on the node  Principles  Assign a task to the best node at a time  Need to decide the order of tasks  Scheduling priority  Min-min : Light task  Max-min : Heavy task  Sufferage : A task whose completion time differs most depending on the node

16 Max-min / Min-min  Calculate completion times for each task and node  For each task take the minimum completion time  Take one from unscheduled tasks  Min-min : Choose a task which has “max” value  Max-min : Choose a task which has “max” value  Schedule the task to the best node  Calculate completion times for each task and node  For each task take the minimum completion time  Take one from unscheduled tasks  Min-min : Choose a task which has “max” value  Max-min : Choose a task which has “max” value  Schedule the task to the best node Task 0Task 1Task 2 node A862 node B193 node C584 Min-min Max-min

17 Sufferage  For each task, calculate Sufferage (The difference between the minimum and second minimum completion times)  Take a task which has maximum Sufferage  Schedule the task to the best node  For each task, calculate Sufferage (The difference between the minimum and second minimum completion times)  Take a task which has maximum Sufferage  Schedule the task to the best node Task 0Task 1Task 2 Node A862 Node B193 Node C584 Sufferage = 4 Sufferage = 2 Sufferage = 1

18 Comparing Scheduling Heuristics  A simulation was done to compare some scheduling tactics [1]  Greedy (Max-min / Min-min)  GA, Simulated annealing, A*, etc.  ETC matrices were randomly generated  512 tasks, 8 nodes  Consistent, inconsistent  GA performed the shortest makespan in most cases, however the calculation cost was not negligible  Min-min heuristics performed well (at most 10% worse than the best)  A simulation was done to compare some scheduling tactics [1]  Greedy (Max-min / Min-min)  GA, Simulated annealing, A*, etc.  ETC matrices were randomly generated  512 tasks, 8 nodes  Consistent, inconsistent  GA performed the shortest makespan in most cases, however the calculation cost was not negligible  Min-min heuristics performed well (at most 10% worse than the best) [1] Tracy et al. A Comparison Study of Eleven Static Heuristics for Mapping a Class of Independent Tasks onto Heterogeneous Distributed Computing Systems (TR-ECE 00-04)

19 (Agenda)  Introduction  Scheduling Algorithms  Data-intensive/Workflow Schedulers  GrADS  Phan’s approach  Conclusion  Introduction  Scheduling Algorithms  Data-intensive/Workflow Schedulers  GrADS  Phan’s approach  Conclusion

20 Scheduling Workflows  Additional Conditions to be considered  Task dependency  Every required file need to be transferred to the node before the task starts  “Non-executable” schedule exists  Data are dynamically generated  The file location is not known in advance  Intermediate files are not needed at last  Additional Conditions to be considered  Task dependency  Every required file need to be transferred to the node before the task starts  “Non-executable” schedule exists  Data are dynamically generated  The file location is not known in advance  Intermediate files are not needed at last

21 GrADS [1]  Execution time estimation  Profile the application behavior  CPU/memory usage  Data transfer cost  Greedy scheduling heuristics  Create ETC matrix for assignable tasks  After assigning a task, some tasks turn to “assignable”  Choose the best schedule from Max-min, min-min and Sufferage  Execution time estimation  Profile the application behavior  CPU/memory usage  Data transfer cost  Greedy scheduling heuristics  Create ETC matrix for assignable tasks  After assigning a task, some tasks turn to “assignable”  Choose the best schedule from Max-min, min-min and Sufferage [1] Mandal. et al. "Scheduling Strategies for Mapping Application Workflows onto the Grid“ in IEEEInternational Symposium on High Performance Distributed Computing (HPDC 2005)

22 GrADS (cont.)  An experiment was done on real tasks  The original data (2GB) was replicated to every cluster in advance  File transfer occurs in clusters  Comparing to random scheduler, it achieved 1.5 to 2.2 times better makespan  An experiment was done on real tasks  The original data (2GB) was replicated to every cluster in advance  File transfer occurs in clusters  Comparing to random scheduler, it achieved 1.5 to 2.2 times better makespan

23 Scheduling Data-intensive Applications [1]  Co-scheduling tasks and data replication  Using GA  A gene contains the followings:  Task order in the global schedule  Assignment of tasks to nodes  Assignment of replicas to nodes  Only part of the tasks are scheduled at a time  Otherwise GA takes too long time  Co-scheduling tasks and data replication  Using GA  A gene contains the followings:  Task order in the global schedule  Assignment of tasks to nodes  Assignment of replicas to nodes  Only part of the tasks are scheduled at a time  Otherwise GA takes too long time [1] Phan et al. “Evolving toward the perfect schedule: Co-scheduling task assignments and data replication in wide-area systems using a genetic algorithm.” In Proceedings of the11th Workshop on task Scheduling Strategies for Parallel Processing. Cambridge, MA. Springer-Verlag, Berlin, Germany.

24 (cont.)  An example of the gene  One schedule is expressed in the gene  An example of the gene  One schedule is expressed in the gene t0t1t4t3t2 t0:n0t1:n1t2:n0t3:n1t4:n0 d0:n0d1:n1d2:n0 t0 t1 t2 t3 t4 Replicas Task assignment Task order t0 n0 n1 t4 t1 t2 t3

25 (cont.)  A simulation was performed  Compared to min-min heuristics with randomly distributed replicas  Number of GA generations are fixed (100)  When 40 tasks are scheduled at a time, GA performs twice better makespan  However, the difference decreases when more tasks are scheduled at a time  GA has not reached the best solution  A simulation was performed  Compared to min-min heuristics with randomly distributed replicas  Number of GA generations are fixed (100)  When 40 tasks are scheduled at a time, GA performs twice better makespan  However, the difference decreases when more tasks are scheduled at a time  GA has not reached the best solution Makespan 80

26 Conclusion  Some scheduling heuristics were introduced  Greedy (Min-min, Max-min, Sufferage)  GrADS can schedule workflows by predicting node performance and using greedy heuristics  A research was done to use GA and co- schedule tasks and data replication  Some scheduling heuristics were introduced  Greedy (Min-min, Max-min, Sufferage)  GrADS can schedule workflows by predicting node performance and using greedy heuristics  A research was done to use GA and co- schedule tasks and data replication

27 Future Work  Most of the research is still on simulation  Hard to predict program/network behavior  A scheduler will be implemented  Using network topology information  Managing Intermediate files  Easy to install and execute  Most of the research is still on simulation  Hard to predict program/network behavior  A scheduler will be implemented  Using network topology information  Managing Intermediate files  Easy to install and execute