University of Texas at Arlington Scheduling and Load Balancing on the NASA Information Power Grid Sajal K. Das, Shailendra Kumar, Manish Arora Department.

Slides:



Advertisements
Similar presentations
A Workflow Engine with Multi-Level Parallelism Supports Qifeng Huang and Yan Huang School of Computer Science Cardiff University
Advertisements

Resource Management §A resource can be a logical, such as a shared file, or physical, such as a CPU (a node of the distributed system). One of the functions.
1 An Adaptive GA for Multi Objective Flexible Manufacturing Systems A. Younes, H. Ghenniwa, S. Areibi uoguelph.ca.
Bag-of-Tasks Scheduling under Budget Constraints Ana-Maria Oprescu, Thilo Kielman Presented by Bryan Rosander.
A system Performance Model Instructor: Dr. Yanqing Zhang Presented by: Rajapaksage Jayampthi S.
SHARCNET. Multicomputer Systems r A multicomputer system comprises of a number of independent machines linked by an interconnection network. r Each computer.
Lincoln University Canterbury New Zealand Evaluating the Parallel Performance of a Heterogeneous System Elizabeth Post Hendrik Goosen formerly of Department.
CISC October Goals for today: Foster’s parallel algorithm design –Partitioning –Task dependency graph Granularity Concurrency Collective communication.
Parallel Programming Models and Paradigms
Fault-tolerant Adaptive Divisible Load Scheduling Xuan Lin, Sumanth J. V. Acknowledge: a few slides of DLT are from Thomas Robertazzi ’ s presentation.
Scheduling with Optimized Communication for Time-Triggered Embedded Systems Slide 1 Scheduling with Optimized Communication for Time-Triggered Embedded.
High Performance Computing 1 Parallelization Strategies and Load Balancing Some material borrowed from lectures of J. Demmel, UC Berkeley.
GHS: A Performance Prediction and Task Scheduling System for Grid Computing Xian-He Sun Department of Computer Science Illinois Institute of Technology.
Strategies for Implementing Dynamic Load Sharing.
Mapping Techniques for Load Balancing
Multilevel Hypergraph Partitioning G. Karypis, R. Aggarwal, V. Kumar, and S. Shekhar Computer Science Department, U of MN Applications in VLSI Domain.
1 Presenter: Ming-Shiun Yang Sah, A., Balakrishnan, M., Panda, P.R. Design, Automation & Test in Europe Conference & Exhibition, DATE ‘09. A Generic.
Virtualization Technology Prof D M Dhamdhere CSE Department IIT Bombay Moving towards Virtualization… Department of Computer Science and Engineering, IIT.
Domain decomposition in parallel computing Ashok Srinivasan Florida State University COT 5410 – Spring 2004.
1 Distributed Operating Systems and Process Scheduling Brett O’Neill CSE 8343 – Group A6.
Cloud Computing 1. Outline  Introduction  Evolution  Cloud architecture  Map reduce operation  Platform 2.
Chao “Bill” Xie, Victor Bolet, Art Vandenberg Georgia State University, Atlanta, GA 30303, USA February 22/23, 2006 SURA, Washington DC Memory Efficient.
Network Aware Resource Allocation in Distributed Clouds.
1 Distributed Process Scheduling: A System Performance Model Vijay Jain CSc 8320, Spring 2007.
The Red Storm High Performance Computer March 19, 2008 Sue Kelly Sandia National Laboratories Abstract: Sandia National.
Computer Science and Engineering Parallel and Distributed Processing CSE 8380 March 01, 2005 Session 14.
DLS on Star (Single-level tree) Networks Background: A simple network model for DLS is the star network with a master-worker platform. It consists of a.
1 A Framework for Data-Intensive Computing with Cloud Bursting Tekin Bicer David ChiuGagan Agrawal Department of Compute Science and Engineering The Ohio.
Porting Irregular Reductions on Heterogeneous CPU-GPU Configurations Xin Huo, Vignesh T. Ravi, Gagan Agrawal Department of Computer Science and Engineering.
Euro-Par, A Resource Allocation Approach for Supporting Time-Critical Applications in Grid Environments Qian Zhu and Gagan Agrawal Department of.
 A System Performance Model  Static Process Scheduling  Dynamic Load Sharing and Balancing  Real-Time Scheduling.
EFFECTIVE LOAD-BALANCING VIA MIGRATION AND REPLICATION IN SPATIAL GRIDS ANIRBAN MONDAL KAZUO GODA MASARU KITSUREGAWA INSTITUTE OF INDUSTRIAL SCIENCE UNIVERSITY.
Supercomputing ‘99 Parallelization of a Dynamic Unstructured Application using Three Leading Paradigms Leonid Oliker NERSC Lawrence Berkeley National Laboratory.
Headline in Arial Bold 30pt HPC User Forum, April 2008 John Hesterberg HPC OS Directions and Requirements.
Xiaobing Wu, Guihai Chen
Computer Science and Engineering Parallelizing Defect Detection and Categorization Using FREERIDE Leonid Glimcher P. 1 ipdps’05 Scaling and Parallelizing.
Lecture 4 TTH 03:30AM-04:45PM Dr. Jianjun Hu CSCE569 Parallel Computing University of South Carolina Department of.
CS 484 Designing Parallel Algorithms Designing a parallel algorithm is not easy. There is no recipe or magical ingredient Except creativity We can benefit.
Department of Computer Science MapReduce for the Cell B. E. Architecture Marc de Kruijf University of Wisconsin−Madison Advised by Professor Sankaralingam.
PaGrid: A Mesh Partitioner for Computational Grids Virendra C. Bhavsar Professor and Dean Faculty of Computer Science UNB, Fredericton This.
CS 484 Load Balancing. Goal: All processors working all the time Efficiency of 1 Distribute the load (work) to meet the goal Two types of load balancing.
DynamicMR: A Dynamic Slot Allocation Optimization Framework for MapReduce Clusters Nanyang Technological University Shanjiang Tang, Bu-Sung Lee, Bingsheng.
An Evaluation of Partitioners for Parallel SAMR Applications Sumir Chandra & Manish Parashar ECE Dept., Rutgers University Submitted to: Euro-Par 2001.
Full and Para Virtualization
Domain decomposition in parallel computing Ashok Srinivasan Florida State University.
Globus and PlanetLab Resource Management Solutions Compared M. Ripeanu, M. Bowman, J. Chase, I. Foster, M. Milenkovic Presented by Dionysis Logothetis.
Data Structures and Algorithms in Parallel Computing Lecture 7.
Massive Semantic Web data compression with MapReduce Jacopo Urbani, Jason Maassen, Henri Bal Vrije Universiteit, Amsterdam HPDC ( High Performance Distributed.
Static Process Scheduling
Pipelined and Parallel Computing Partition for 1 Hongtao Du AICIP Research Nov 3, 2005.
Computer Organization CS224 Fall 2012 Lesson 52. Introduction  Goal: connecting multiple computers to get higher performance l Multiprocessors l Scalability,
Euro-Par, HASTE: An Adaptive Middleware for Supporting Time-Critical Event Handling in Distributed Environments ICAC 2008 Conference June 2 nd,
Load Rebalancing for Distributed File Systems in Clouds.
Background Computer System Architectures Computer System Software.
COMP7330/7336 Advanced Parallel and Distributed Computing Task Partitioning Dr. Xiao Qin Auburn University
COMP7330/7336 Advanced Parallel and Distributed Computing Task Partitioning Dynamic Mapping Dr. Xiao Qin Auburn University
MicroGrid Update & A Synthetic Grid Resource Generator Xin Liu, Yang-suk Kee, Andrew Chien Department of Computer Science and Engineering Center for Networked.
Hybrid Parallel Implementation of The DG Method Advanced Computing Department/ CAAM 03/03/2016 N. Chaabane, B. Riviere, H. Calandra, M. Sekachev, S. Hamlaoui.
CSCAPES Mission Research and development Provide load balancing and parallelization toolkits for petascale computation Develop advanced automatic differentiation.
Lecture 13 Parallel Processing. 2 What is Parallel Computing? Traditionally software has been written for serial computation. Parallel computing is the.
Department of Computer Science, Johns Hopkins University Lecture 7 Finding Concurrency EN /420 Instructor: Randal Burns 26 February 2014.
High Performance Computing Seminar
Parallel Programming By J. H. Wang May 2, 2017.
Grid Computing.
Parallel Algorithm Design
Department of Computer Science University of California, Santa Barbara
CS 584.
Department of Computer Science University of California, Santa Barbara
Parallel Feature Identification and Elimination from a CFD Dataset
Presentation transcript:

University of Texas at Arlington Scheduling and Load Balancing on the NASA Information Power Grid Sajal K. Das, Shailendra Kumar, Manish Arora Department of Computer Science and Engineering University of Texas at Arlington

Information Power Grid Distributed, heterogeneous metacomputing platform Cost effective – the future of high performance computing Orders of magnitude increase in computational capability Challenges to map varied applications on a truly heterogeneous, distributed suite of systems EuroGrid, DTF, GriPhyN, DataGrid, iVDGL, DOEScGrid

University of Texas at Arlington Abstract System Model  Distributed heterogeneous computing require a paradigm shift from conventional parallel processing  Success of grid computing lies in mapping varied applications on the grid  Development of novel scheduling and load balancing algorithms for such environments is a natural requirement Research Challenges

University of Texas at Arlington Application Classification  Type A: Parallel distributed mode of execution – large resource requirements – executed in parallel on multiple nodes of the Grid  Type B: Independent jobs – run on a SINGLE node of the Grid – will have a set of resource requirements that will determine the subset of Grid nodes on which it can run  Type C: Jobs with dependency – jobs are either Type A or B, but depend on other jobs in order to begin execution Contributions  MiniMax Heterogeneous Partitioner – Type A applications  DONAN: Decentralized overhead nullifying algorithm for N-resource grid – Type B applications

University of Texas at Arlington Type A Scheduling = Partitioning  Workload Model: – Graph representation of complex, compute intensive problems – Application Domains  Computational fluid dynamics (CFD)  Numerical methods for solving differential equations  Matrix based computations  Data Parallel computation  Extra-terrestrial and geographic information processing – Workload Graph - weighted undirected graph, G = (V, E) – Computational weight - W v – Communication weight - C (u, v)  System Model: – Weighted undirected System Graph, S = (P, L) – Processing weight - s p – Link weight - c (p, q) – Communication Cost Matrix

University of Texas at Arlington MiniMax - Multilevel Heterogeneous Partitioner  Objective – minimizing application execution time  Full heterogeneity in the system & workload graphs  3 phases : - Coarsening : Transform original graph into a sequence of smaller graphs - Initial Partition: Smallest graph is partitioned among processors. - Refinement: Uncoarsen each level up to the original graph. Optimize ET app at each level. Coarsening Refinement Initial Partition

University of Texas at Arlington Performance Measures  Partition Quality - Execution Time of the application (ET) - Standard Deviation (  ) - Imbalance Factor (Imb): ET/ ET av  Workload Graphs  NASA Test Mesh, Synthetic Uniform, Synthetic Non-Uniform  System Graphs  Homo, PHetero, LHetero, CHetero, FullHetero

University of Texas at Arlington

Type B Scheduling  Mapping a set of independent jobs onto a set of nodes.  Recent Systems (e.g. SUN E10000 and SGI 02K), made up of pool of resources. –Hardware: CPUs, Shared Memory, Large Disk farms, distinct I/O channels –Software: Licenses  Each job i has its resource requirement of  Each node i offers resources  De-centralized & scalable  Nullifies the overhead of looking for complimentary nodes  Works for a general N-resource distributed systems  Experiments done with 1, 2 and 3 resource nodes  Normalized performance as high as 0.85

University of Texas at Arlington

Conclusion & Future Work  Developed MiniMax and DONAN scheduling algorithms  Scheduler for Type C Applications  Integration with Globus  Test functionality on real Grid