Dynamic Load Balancing Experiments in a Grid Vrije Universiteit Amsterdam, The Netherlands CWI Amsterdam, The

Slides:



Advertisements
Similar presentations
A NOVEL APPROACH TO SOLVING LARGE-SCALE LINEAR SYSTEMS Ken Habgood, Itamar Arel Department of Electrical Engineering & Computer Science GABRIEL CRAMER.
Advertisements

Resource Management §A resource can be a logical, such as a shared file, or physical, such as a CPU (a node of the distributed system). One of the functions.
Building a Distributed Full-Text Index for the Web S. Melnik, S. Raghavan, B.Yang, H. Garcia-Molina.
Acoustic design by simulated annealing algorithm
A system Performance Model Instructor: Dr. Yanqing Zhang Presented by: Rajapaksage Jayampthi S.
Course Outline Introduction in algorithms and applications Parallel machines and architectures Overview of parallel machines, trends in top-500 Cluster.
Distributed Process Scheduling Summery Distributed Process Scheduling Summery BY:-Yonatan Negash.
Optimization of a FEA Beam Element Model for Computational Speed and Accuracy Brent Weight and Ryan Weight.
A Parallel Computational Model for Heterogeneous Clusters Jose Luis Bosque, Luis Pastor, IEEE TRASACTION ON PARALLEL AND DISTRIBUTED SYSTEM, VOL. 17, NO.
GridFlow: Workflow Management for Grid Computing Kavita Shinde.
Adaptive Data Collection Strategies for Lifetime-Constrained Wireless Sensor Networks Xueyan Tang Jianliang Xu Sch. of Comput. Eng., Nanyang Technol. Univ.,
Effectively Utilizing Global Cluster Memory for Large Data-Intensive Parallel Programs John Oleszkiewicz, Li Xiao, Yunhao Liu IEEE TRASACTION ON PARALLEL.
Parallel Simulation etc Roger Curry Presentation on Load Balancing.
Design and Performance Evaluation of Queue-and-Rate-Adjustment Dynamic Load Balancing Policies for Distributed Networks Zeng Zeng, Bharadwaj, IEEE TRASACTION.
Steady Aeroelastic Computations to Predict the Flying Shape of Sails Sriram Antony Jameson Dept. of Aeronautics and Astronautics Stanford University First.
Fault-tolerant Adaptive Divisible Load Scheduling Xuan Lin, Sumanth J. V. Acknowledge: a few slides of DLT are from Thomas Robertazzi ’ s presentation.
May 29, Final Presentation Sajib Barua1 Development of a Parallel Fast Fourier Transform Algorithm for Derivative Pricing Using MPI Sajib Barua.
High Performance Computing 1 Parallelization Strategies and Load Balancing Some material borrowed from lectures of J. Demmel, UC Berkeley.
Design, Implementation, and Evaluation of Differentiated Caching Services Ying Lu, Tarek F. Abdelzaher, Avneesh Saxena IEEE TRASACTION ON PARALLEL AND.
1 An Empirical Study on Large-Scale Content-Based Image Retrieval Group Meeting Presented by Wyman
12006/9/26 Load Balancing in Dynamic Structured P2P Systems Brighten Godfrey, Karthik Lakshminarayanan, Sonesh Surana, Richard Karp, Ion Stoica INFOCOM.
A Multi-Agent Learning Approach to Online Distributed Resource Allocation Chongjie Zhang Victor Lesser Prashant Shenoy Computer Science Department University.
Redundant Parallel File Transfer with Anticipative Adjustment Mechanism in Data Grids Chao-Tung Yang, Yao-Chun Chi, Chun-Pin Fu, High-Performance Computing.
Customized Dynamic Load Balancing for a Network of Workstations Taken from work done by: Mohammed Javeed Zaki, Wei Li, Srinivasan Parthasarathy Computer.
Dynamic Load Sharing and Balancing Sig Freund. Outline Introduction Distributed vs. Traditional scheduling Process Interaction models Distributed Systems.
A User Experience-based Cloud Service Redeployment Mechanism KANG Yu.
LABERIO: Dynamic load- balanced routing in OpenFlow- enabled networks Author:Hui Long, Yao Shen*, Minyi Guo, Feilong Tang IEEE 27th International.
Computer Science 320 Measuring Speedup. What Is Running Time? T(N, K) says that the running time T is a function of the problem size N and the number.
MobSched: An Optimizable Scheduler for Mobile Cloud Computing S. SindiaS. GaoB. Black A.LimV. D. AgrawalP. Agrawal Auburn University, Auburn, AL 45 th.
Self Adaptivity in Grid Computing Reporter : Po - Jen Lo Sathish S. Vadhiyar and Jack J. Dongarra.
Predictive Runtime Code Scheduling for Heterogeneous Architectures 1.
Efficient Network-Coding-Based Opportunistic Routing Through Cumulative Coded Acknowledgments Dimitrios Koutsonikolas, Chih-Chun Wang and Y. Charlie Hu.
Sensor Positioning in Wireless Ad-hoc Sensor Networks Using Multidimensional Scaling Xiang Ji and Hongyuan Zha Dept. of Computer Science and Engineering,
Wei Gao1 and Qinghua Li2 1The University of Tennessee, Knoxville
Frankfurt (Germany), 6-9 June 2011 SCHEDULING CHARGING OF ELECTRIC VEHICLES FOR OPTIMAL DISTRIBUTION SYSTEMS PLANNING AND OPERATION David STEEN*Anh Tuan.
1 On the Placement of Web Server Replicas Lili Qiu, Microsoft Research Venkata N. Padmanabhan, Microsoft Research Geoffrey M. Voelker, UCSD IEEE INFOCOM’2001,
An Efficient Approach for Content Delivery in Overlay Networks Mohammad Malli Chadi Barakat, Walid Dabbous Planete Project To appear in proceedings of.
Multi-attribute, Energy Optimal Sensor Fusion in Hurricane Model Simulations Marlon J Fuentes Bennie Lewis Spring 2008 Advance Topics in Wireless Networks.
Collective Buffering: Improving Parallel I/O Performance By Bill Nitzberg and Virginia Lo.
Intelligent Database Systems Lab 1 Advisor : Dr. Hsu Graduate : Jian-Lin Kuo Author : Silvia Nittel Kelvin T.Leung Amy Braverman 國立雲林科技大學 National Yunlin.
1 Challenge the future High accuracy machines on factory floors Anthonie Boogaard.
1 Distributed Energy-Efficient Scheduling for Data-Intensive Applications with Deadline Constraints on Data Grids Cong Liu and Xiao Qin Auburn University.
1 Exploring Custom Instruction Synthesis for Application-Specific Instruction Set Processors with Multiple Design Objectives Lin, Hai Fei, Yunsi ACM/IEEE.
Heterogeneity-Aware Peak Power Management for Accelerator-based Systems Heterogeneity-Aware Peak Power Management for Accelerator-Based Systems Gui-Bin.
Parallelization of the Classic Gram-Schmidt QR-Factorization
Prognosis of gear health using stochastic dynamical models with online parameter estimation 10th International PhD Workshop on Systems and Control a Young.
Stability of size-based scheduling in resource-sharing networks Maaike Verloop CWI & Utrecht U. Sem Borst CWI & Eindhoven U.T. & Lucent Bell Labs Sindo.
Dynamic Load Balancing and Job Replication in a Global-Scale Grid Environment: A Comparison IEEE Transactions on Parallel and Distributed Systems, Vol.
Advanced Spectrum Management in Multicell OFDMA Networks enabling Cognitive Radio Usage F. Bernardo, J. Pérez-Romero, O. Sallent, R. Agustí Radio Communications.
Adaptive Multi-Threading for Dynamic Workloads in Embedded Multiprocessors 林鼎原 Department of Electrical Engineering National Cheng Kung University Tainan,
OPTIMIZING DSP SCHEDULING VIA ADDRESS ASSIGNMENT WITH ARRAY AND LOOP TRANSFORMATION Chun Xue, Zili Shao, Ying Chen, Edwin H.-M. Sha Department of Computer.
1 1 Slide Simulation Professor Ahmadi. 2 2 Slide Simulation Chapter Outline n Computer Simulation n Simulation Modeling n Random Variables and Pseudo-Random.
OPERATING SYSTEMS CS 3530 Summer 2014 Systems and Models Chapter 03.
A System Performance Model Distributed Process Scheduling.
Computer Science 320 Load Balancing. Behavior of Parallel Program Why do 3 threads take longer than two?
Multicast Scaling Laws with Hierarchical Cooperation Chenhui Hu, Xinbing Wang, Ding Nie, Jun Zhao Shanghai Jiao Tong University, China.
V.M. Sliusar, V.I. Zhdanov Astronomical Observatory, Taras Shevchenko National University of Kyiv Observatorna str., 3, Kiev Ukraine
Data Consolidation: A Task Scheduling and Data Migration Technique for Grid Networks Author: P. Kokkinos, K. Christodoulopoulos, A. Kretsis, and E. Varvarigos.
A stochastic scheduling algorithm for precedence constrained tasks on Grid Future Generation Computer Systems (2011) Xiaoyong Tang, Kenli Li, Guiping Liao,
Network Weather Service. Introduction “NWS provides accurate forecasts of dynamically changing performance characteristics from a distributed set of metacomputing.
CSCI-455/552 Introduction to High Performance Computing Lecture 21.
WSRec: A Collaborative Filtering Based Web Service Recommender System
Mohammad Malli Chadi Barakat, Walid Dabbous Alcatel meeting
Paul Pop, Petru Eles, Zebo Peng
Parallel Density-based Hybrid Clustering
System Control based Renewable Energy Resources in Smart Grid Consumer
A Consensus-Based Clustering Method
Course Outline Introduction in algorithms and applications
Networked Real-Time Systems: Routing and Scheduling
End-to-End Internet Delay Behavior
Presentation transcript:

Dynamic Load Balancing Experiments in a Grid Vrije Universiteit Amsterdam, The Netherlands CWI Amsterdam, The Presented by 張肇烜

Outline Introduction Introduction Experimental Setup Experimental Setup Implementation of Dynamic Load Balancing Implementation of Dynamic Load Balancing Experimental results Experimental results Conclusions Conclusions

Introduction A gird environment is extremely unpredictable. A gird environment is extremely unpredictable. Exponential Smoothing (ES) was shown to be a good predictor for processing power. Exponential Smoothing (ES) was shown to be a good predictor for processing power. The experiments were performed with the classical Successive Over Relaxation (SOR) application. The experiments were performed with the classical Successive Over Relaxation (SOR) application.

Experimental Setup To carry out experiments with parallel applications in a realistic setting, the test bed must have the following key characteristics of a grid environment: To carry out experiments with parallel applications in a realistic setting, the test bed must have the following key characteristics of a grid environment: Processor capacities often differ. Processor capacities often differ. Processor loads change over time. Processor loads change over time. Processors are geographically distributed. Processors are geographically distributed. Network conditions are highly unpredictable. Network conditions are highly unpredictable.

Experimental Setup (cont.) We have chosen to conduct our experiments with four sites. We have chosen to conduct our experiments with four sites.

Experimental Setup (cont.) Our implementation of the load balancing step is as follows: Our implementation of the load balancing step is as follows: At the end of each iteration the processor predict their processing speed for the next iteration. At the end of each iteration the processor predict their processing speed for the next iteration. After every N iterations the processors send their prediction to processor 0, the DLB scheduler. After every N iterations the processors send their prediction to processor 0, the DLB scheduler. The processor calculates the “ optimal ” load distribution given those prediction and sends relevant information to each processor. The processor calculates the “ optimal ” load distribution given those prediction and sends relevant information to each processor.

Experimental Setup (cont.) Equal Load Balancing (ELB) : Equal Load Balancing (ELB) : ELB assumes no prior knowledge of processor speeds of the nodes, and consequently balances the load equally among the different nodes. ELB assumes no prior knowledge of processor speeds of the nodes, and consequently balances the load equally among the different nodes.

Implementation of Dynamic Load Balancing We use the Exponential Smoothing (ES) technique to obtain these predictions. We use the Exponential Smoothing (ES) technique to obtain these predictions. ES appears to be a simple and usable method in load balancing strategies. ES appears to be a simple and usable method in load balancing strategies.

Experimental results Stochastic behavior of the calculation times: Stochastic behavior of the calculation times:

Experimental results (cont.) Stochastic behavior of the calculation times: Stochastic behavior of the calculation times:

Experimental results (cont.) Stochastic behavior of the calculation times: Stochastic behavior of the calculation times:

Experimental results (cont.) Experiments with DLB and ELB: Experiments with DLB and ELB: Running time for SOR based on DLB compared to ELB: Running time for SOR based on DLB compared to ELB:

Experimental results (cont.) Experiments with DLB and ELB: Experiments with DLB and ELB: Cumulative running time as a function of the iteration number: Cumulative running time as a function of the iteration number:

Experimental results (cont.) Experiments with DLB and ELB: Experiments with DLB and ELB: Running times as a function of the number of rows: Running times as a function of the number of rows:

Conclusions Extensive experimentation in the testbed environment PlanetLab have led to the following conclusions: Extensive experimentation in the testbed environment PlanetLab have led to the following conclusions: A significant speedup factor of on average 1.8 can be consistently achieved by implementing DLB. A significant speedup factor of on average 1.8 can be consistently achieved by implementing DLB. The relation between the running time and the problem size is approximately linear. The relation between the running time and the problem size is approximately linear.

Conclusions (cont.) We address a number of challenges for further research: We address a number of challenges for further research: In depth-analysis of parallel applications with a non-linear structure. In depth-analysis of parallel applications with a non-linear structure. Develop optimal load balancing schemes advanced and accurate predictions of the calculation times are needed. Develop optimal load balancing schemes advanced and accurate predictions of the calculation times are needed.