Load Balancing and Data centers

Slides:



Advertisements
Similar presentations
VARUN GUPTA Carnegie Mellon University 1 With: Mor Harchol-Balter (CMU)
Advertisements

LESSONs NINE and TEN QUEUING MODELS.
Load Balancing of Elastic Traffic in Heterogeneous Wireless Networks Abdulfetah Khalid, Samuli Aalto and Pasi Lassila
Lindsay Mullen Seminar Presentation #2 November 4, 2013
Queuing Analysis Based on noted from Appendix A of Stallings Operating System text 6/10/20151.
ECS 152A Acknowledgement: slides from S. Kalyanaraman & B.Sikdar
OS Fall ’ 02 Performance Evaluation Operating Systems Fall 2002.
1 Performance Evaluation of Computer Networks Objectives  Introduction to Queuing Theory  Little’s Theorem  Standard Notation of Queuing Systems  Poisson.
1 Queuing Theory 2 Queuing theory is the study of waiting in lines or queues. Server Pool of potential customers Rear of queue Front of queue Line (or.
Data Communication and Networks Lecture 13 Performance December 9, 2004 Joseph Conron Computer Science Department New York University
1 Queueing Theory H Plan: –Introduce basics of Queueing Theory –Define notation and terminology used –Discuss properties of queuing models –Show examples.
Performance Evaluation
Join-the-Shortest-Queue (JSQ) Routing in Web Server Farms
Queueing Network Model. Single Class Model Open - Infinite stream of arriving customers Closed - Finite population eg Intranet users Indistinguishable.
Queuing Analysis Based on noted from Appendix A of Stallings Operating System text 6/28/20151.
Queuing Theory. Queuing theory is the study of waiting in lines or queues. Server Pool of potential customers Rear of queue Front of queue Line (or queue)
Introduction to Queuing Theory. 2 Queuing theory definitions  (Kleinrock) “We study the phenomena of standing, waiting, and serving, and we call this.

1 Real-Time Queueing Network Theory Presented by Akramul Azim Department of Electrical and Computer Engineering University of Waterloo, Canada John P.
Introduction to Queuing Theory
Queueing Theory I. Summary Little’s Law Queueing System Notation Stationary Analysis of Elementary Queueing Systems  M/M/1  M/M/m  M/M/1/K  …
OPTIMAL SERVER PROVISIONING AND FREQUENCY ADJUSTMENT IN SERVER CLUSTERS Presented by: Xinying Zheng 09/13/ XINYING ZHENG, YU CAI MICHIGAN TECHNOLOGICAL.
Probability Review Thinh Nguyen. Probability Theory Review Sample space Bayes’ Rule Independence Expectation Distributions.
Management of Waiting Lines McGraw-Hill/Irwin Copyright © 2012 by The McGraw-Hill Companies, Inc. All rights reserved.
CS433 Modeling and Simulation Lecture 13 Queueing Theory Dr. Anis Koubâa 03 May 2009 Al-Imam Mohammad Ibn Saud University.
Chapter 6 Scheduling. Basic concepts Goal is maximum utilization –what does this mean? –cpu pegged at 100% ?? Most programs are I/O bound Thus some other.
18 Management of Waiting Lines.
Queuing Theory Basic properties, Markovian models, Networks of queues, General service time distributions, Finite source models, Multiserver queues Chapter.
Queueing Theory What is a queue? Examples of queues: Grocery store checkout Fast food (McDonalds – vs- Wendy’s) Hospital Emergency rooms Machines waiting.
TexPoint fonts used in EMF.
1 Elements of Queuing Theory The queuing model –Core components; –Notation; –Parameters and performance measures –Characteristics; Markov Process –Discrete-time.
1 Systems Analysis Methods Dr. Jerrell T. Stracener, SAE Fellow SMU EMIS 5300/7300 NTU SY-521-N NTU SY-521-N SMU EMIS 5300/7300 Queuing Modeling and Analysis.
CS433 Modeling and Simulation Lecture 12 Queueing Theory Dr. Anis Koubâa 03 May 2008 Al-Imam Mohammad Ibn Saud University.
1 Chapters 8 Overview of Queuing Analysis. Chapter 8 Overview of Queuing Analysis 2 Projected vs. Actual Response Time.
1 The Effect of Heavy-Tailed Job Size Distributions on System Design Mor Harchol-Balter MIT Laboratory for Computer Science.
Modeling and Simulation Queuing theory
CS352 - Introduction to Queuing Theory Rutgers University.
Maciej Stasiak, Mariusz Głąbowski Arkadiusz Wiśniewski, Piotr Zwierzykowski Model of the Nodes in the Packet Network Chapter 10.
Structure of a Waiting Line System Queuing theory is the study of waiting lines Four characteristics of a queuing system: –The manner in which customers.
Modelling job allocation where service duration is unknown Nigel Thomas University of Newcastle upon Tyne.
1 Queuing Delay and Queuing Analysis. RECALL: Delays in Packet Switched (e.g. IP) Networks End-to-end delay (simplified) = End-to-end delay (simplified)
Queuing Theory.  Queuing Theory deals with systems of the following type:  Typically we are interested in how much queuing occurs or in the delays at.
1 Task Assignment with Unknown Duration Mor Harchol-Balter Carnegie Mellon.
Management of Waiting Lines Copyright © 2015 McGraw-Hill Education. All rights reserved. No reproduction or distribution without the prior written consent.
Mohammad Khalily Islamic Azad University.  Usually buffer size is finite  Interarrival time and service times are independent  State of the system.
QUEUING THEORY 1.  - means the number of arrivals per second   - service rate of a device  T - mean service time for each arrival   = ( ) Utilization,
Managerial Decision Making Chapter 13 Queuing Models.
McGraw-Hill/Irwin Copyright © 2009 by The McGraw-Hill Companies, Inc. All Rights Reserved. Chapter 18 Management of Waiting Lines.
Prepared by Lloyd R. Jaisingh
Serve Assignment Policies
Al-Imam Mohammad Ibn Saud University
Dynamic Graph Partitioning Algorithm
Queueing Theory What is a queue? Examples of queues:
Queuing Theory Non-Markov Systems
Chapter 9: Queuing Models
B.Ramamurthy Appendix A
CPSC 531: System Modeling and Simulation
Evaluation of Load Balancing Algorithms and Internet Traffic Modeling for Performance Analysis By Arthur L. Blais.
Queueing Theory Carey Williamson Department of Computer Science
Introduction Notation Little’s Law aka Little’s Result
System Performance: Queuing
Handling Routing Transport Haifa JFK TLV BGN To: Yishay From: Vered
TexPoint fonts used in EMF.
Lecture 2 Part 3 CPU Scheduling
Queueing Theory 2008.
Carey Williamson Department of Computer Science University of Calgary
CSE 550 Computer Network Design
Waiting Line Models Waiting takes place in virtually every productive process or service. Since the time spent by people and things waiting in line is.
Course Description Queuing Analysis This queuing course
VIRTUE MARYLEE MUGURACHANI QUEING THEORY BIRTH and DEATH.
Presentation transcript:

Load Balancing and Data centers ELEN6909 Network Algorithm and Dynamics —— project presentation Topic: Load Balancing and Data centers Reference Paper: "On choosing a task assignment policy for a distributed server system.” Harchol-Balter, Mor, Mark E. Crovella, and Cristina D. Murta. Lin Su UNI: ls3201 May 18, 2015

2. Task size distribution 3. Four task assignments approaches overview Topic: Task assignment policy for a distributed server system 1. Problem and model 2. Task size distribution 3. Four task assignments approaches 4. Simulation 5. Analysis and conclusion

1. Problem and model Problem: Model: In a distributed server system, requests for service arrive and must be assigned to one of the host machines for processing. —— task assignment policy. Which task assignment policy is better? (Random; Round-Robin; Size-based; Dynamic) Model: Model task size using Bounded Pareto distribution. Tasks arrive to the system according to a Poisson process. Dispatcher facility assigns arrived tasks to one of the hosts. Tasks assigned to each host are served in FCFS order.

How to determine which task assignment is better? 1. Problem and model How to determine which task assignment is better? Take task size as variable (Bounded Pareto distribution) Measure each policy’s performance by mean waiting time and mean slowdown. ——Minimize mean waiting time. ——Minimize mean slowdown. (slowdown = waiting time/service requirement)

The distribution of task size variable: Pareto distribution 2. Task size distribution The distribution of task size variable: Pareto distribution Exponential → Heavy-tailed → Pareto → Bounded Pareto Heavy-tailed distribution Bounded Pareto distribution

Heavy-tailed distribution Exponential distribution: Heavy-tailed distribution: A distribution with a “tail” that is “heavier” than an Exponential.

Properties of Heavy-tailed distribution There may be only one large size of task increases the mean waiting time a lot.

Pareto distribution (Power-tailed distribution) Power-tailed distribution is a particular case of Heavy-tailed distribution. Pareto distribution is a particular case of Power-tailed distribution.

Properties of Pareto distribution

Bounded Pareto distribution The lower α, the more variable the distribution. (Intuitively, you can think that more tasks of large size occur.)

3. Four task assignment approaches Four approaches: Random; Round-Robin; Size-based; Dynamic A new size-based task assignment: SITA-E Random: h hosts, each job gets assigned to a host with probability 1/h. Round-Robin: ith job assigned to host i mod h. Size-Based: Each host serves tasks whose service demand falls in a designated range. Dynamic: (Shortest-queue) (Least-work-remaining) job immediately dispatched to the host with the shortest queue.

Size-based task assignment: SITA-E SITA-E (Size Interval Task Assignment with Equal Load) Idea: Define the size range associated with each host such that the total work load directed to each host is the same. Motivation: Balance the load to minimizes mean waiting time. Use the task size distribution to define the cutoff points (range) so that the expected work directed to each host is the same.

4. Simulation Measurement parameters Mean waiting time. Mean slowdown. (slowdown = waiting time/service demand) Consider α in the range 1.1 to 1.9. (α value in the range 1.0 to 1.3 tend to be common in empirical measurements of computing system.) The lower α, the more variable the distribution. → the longer the mean waiting time/slowdown.

Simulation result observation Random and Round Robin policies are similar. As α declines, both of the performance metrics explode approximately exponentially. Dynamic policy performs quite well when α is larger (α→2), however when variability in task size decreases as α→1, Dynamic is unable to maintain good performance. SITA-E performs well over the entire range of α value studied. When task sizes are less variable, Dynamic task assignment exhibits better performance; When task sizes show more variability, SITA-E’s performance can be 100 times better than that of Dynamic.

5. Analysis and conclusion Notation/definitions used in Queuing Theory Pollaczek-Kinchin formula Analysis of four task assignment policies Conclusion

Notation/definitions used in Queuing Theory Little’s Law

Notation/definitions used in Queuing Theory Kendall Notation —— the standard system used to describe and classify a queuing node. M/M/1 queue represents the queue length in a system having a single server, where arrivals are determined by a Poisson process and job service times have an exponential distribution. M/D/1 queue represents the queue length in a system having a single server, where arrivals are determined by a Poisson process and job service times are fixed (deterministic). M/G/1 queue is a queue model where arrivals are Markovian (modulated by a Poisson process), service times have a General distribution and there is a single server.

Pollaczek-Kinchin formula

Analysis of four approaches

Analysis of four approaches Three reasons why SITA-E policy performs so well: By limiting the range of task sizes at each host, SITA-E greatly reduces the variance of the task size distribution, thereby improving performance at these host. Intensified by the heavy-tailed property, very few tasks are assigned to high numbered host. Small tasks observe proportionately lower waiting times.

Conclusion How the variability of the task size distribution influences which task assignment policy is best in a distributed system? When task sizes are not highly variable, the Dynamic policy is preferable. When task sizes show the degree of variability more characteristic of empirical measurements (α≈1), SITA-E is best. Random, Round-Robin and Dynamic policies show that their performance is directly proportional to the second moment of the task size distribution. SITA-E performs well for three main reasons: 1. By limiting the range of task sizes at each host, SITA-E greatly reduces the variance of the task size distribution, thereby improving performance at these host. 2. Intensified by the heavy-tailed property, very few tasks are assigned to high numbered host. 3. Small tasks observe proportionately lower waiting times.

ELEN6909 Network Algorithm and Dynamics —— project presentation Topic: Load Balancing and Data centers ——Task assignment policy for a distributed server system Reference Paper: "On choosing a task assignment policy for a distributed server system.” Harchol-Balter, Mor, Mark E. Crovella, and Cristina D. Murta. Thank you! Lin Su UNI: ls3201 May 18, 2015