Research Paper Example Exploiting Process Lifetime Distributions for Dynamic Load Balancing Mor Harchol-Balter Allen Downey SIGMETRICS 2006.

Slides:



Advertisements
Similar presentations
Dr. Kalpakis CMSC 621, Advanced Operating Systems. Distributed Scheduling.
Advertisements

1 Sizing the Streaming Media Cluster Solution for a Given Workload Lucy Cherkasova and Wenting Tang HPLabs.
Ch 11 Distributed Scheduling –Resource management component of a system which moves jobs around the processors to balance load and maximize overall performance.
Overcoming Limitations of Sampling for Agrregation Queries Surajit ChaudhuriMicrosoft Research Gautam DasMicrosoft Research Mayur DatarStanford University.
Scheduling Criteria CPU utilization – keep the CPU as busy as possible (from 0% to 100%) Throughput – # of processes that complete their execution per.
Dynamic Thread Assignment on Heterogeneous Multiprocessor Architectures Pree Thiengburanathum Advanced computer architecture Oct 24,
Resource Management §A resource can be a logical, such as a shared file, or physical, such as a CPU (a node of the distributed system). One of the functions.
Extensibility, Safety and Performance in the SPIN Operating System Presented by Allen Kerr.
Dr. Kalpakis CMSC 621, Advanced Operating Systems. Fall 2003 URL: Distributed Scheduling.
Chapter 10 Operating Systems.
Chap 5 Process Scheduling. Basic Concepts Maximum CPU utilization obtained with multiprogramming CPU–I/O Burst Cycle – Process execution consists of a.
A SYSTEM PERFORMANCE MODEL CSC 8320 Advanced Operating Systems Georgia State University Yuan Long.
1 Virtual Machine Resource Monitoring and Networking of Virtual Machines Ananth I. Sundararaj Department of Computer Science Northwestern University July.
Scheduling in Batch Systems
Effects and Implications of File Size/Service Time Correlation on Web Server Scheduling Policies Dong Lu* + Peter Dinda* Yi Qiao* Huanyuan Sheng* *Northwestern.
1 Introduction to Load Balancing: l Definition of Distributed systems. Collection of independent loosely coupled computing resources. l Load Balancing.
Present by Chen, Ting-Wei Adaptive Task Checkpointing and Replication: Toward Efficient Fault-Tolerant Grids Maria Chtepen, Filip H.A. Claeys, Bart Dhoedt,
Online Data Gathering for Maximizing Network Lifetime in Sensor Networks IEEE transactions on Mobile Computing Weifa Liang, YuZhen Liu.
Grid Load Balancing Scheduling Algorithm Based on Statistics Thinking The 9th International Conference for Young Computer Scientists Bin Lu, Hongbin Zhang.
1 Distributed Systems: Distributed Process Management – Process Migration.
Distributed Process Management1 Learning Objectives Distributed Scheduling Algorithms Coordinator Elections Orphan Processes.
1 Ekow J. Otoo Frank Olken Arie Shoshani Adaptive File Caching in Distributed Systems.
Load distribution in distributed systems
STRATEGIES INVOLVED IN REMOTE COMPUTATION
Self Adaptivity in Grid Computing Reporter : Po - Jen Lo Sathish S. Vadhiyar and Jack J. Dongarra.
CSC 360- Instructor: K. Wu CPU Scheduling. CSC 360- Instructor: K. Wu Agenda 1.What is CPU scheduling? 2.CPU burst distribution 3.CPU scheduler and dispatcher.
1 Distributed Operating Systems and Process Scheduling Brett O’Neill CSE 8343 – Group A6.
OPTIMAL SERVER PROVISIONING AND FREQUENCY ADJUSTMENT IN SERVER CLUSTERS Presented by: Xinying Zheng 09/13/ XINYING ZHENG, YU CAI MICHIGAN TECHNOLOGICAL.
Load Balancing in Distributed Computing Systems Using Fuzzy Expert Systems Author Dept. Comput. Eng., Alexandria Inst. of Technol. Content Type Conferences.
 Escalonamento e Migração de Recursos e Balanceamento de carga Carlos Ferrão Lopes nº M6935 Bruno Simões nº M6082 Celina Alexandre nº M6807.
Active Monitoring in GRID environments using Mobile Agent technology Orazio Tomarchio Andrea Calvagna Dipartimento di Ingegneria Informatica e delle Telecomunicazioni.
Chapter 5 – CPU Scheduling (Pgs 183 – 218). CPU Scheduling  Goal: To get as much done as possible  How: By never letting the CPU sit "idle" and not.
Frascati, October 9th, Accounting in DataGrid Initial Architecture Albert Werbrouck Frascati, October 9, 2001.
Overlay Network Physical LayerR : router Overlay Layer N R R R R R N.
Architectural Support for Fine-Grained Parallelism on Multi-core Architectures Sanjeev Kumar, Corporate Technology Group, Intel Corporation Christopher.
1 Distributed Process Scheduling: A System Performance Model Vijay Jain CSc 8320, Spring 2007.
Parallel and Distributed Simulation Memory Management & Other Optimistic Protocols.
1 The Effect of Heavy-Tailed Job Size Distributions on System Design Mor Harchol-Balter MIT Laboratory for Computer Science.
Process A program in execution. But, What does it mean to be “in execution”?
Page 1 Process Migration & Allocation Paul Krzyzanowski Distributed Systems Except as otherwise noted, the content of this.
Cpr E 308 Spring 2005 Process Scheduling Basic Question: Which process goes next? Personal Computers –Few processes, interactive, low response time Batch.
Chapter 5: CPU Scheduling. 5.2 Silberschatz, Galvin and Gagne ©2005 Operating System Concepts – 7 th Edition, Feb 2, 2005 Chapter 5: CPU Scheduling Basic.
Introduction: Memory Management 2 Ideally programmers want memory that is large fast non volatile Memory hierarchy small amount of fast, expensive memory.
Design Issues of Prefetching Strategies for Heterogeneous Software DSM Author :Ssu-Hsuan Lu, Chien-Lung Chou, Kuang-Jui Wang, Hsiao-Hsi Wang, and Kuan-Ching.
1 Isolating Web Programs in Modern Browser Architectures CS6204: Cloud Environment Spring 2011.
Static Process Scheduling
Data Consolidation: A Task Scheduling and Data Migration Technique for Grid Networks Author: P. Kokkinos, K. Christodoulopoulos, A. Kretsis, and E. Varvarigos.
Distributed Scheduling Motivations: reduce response time of program execution through load balancing Goal: enable transparent execution of programs on.
Chapter 4 CPU Scheduling. 2 Basic Concepts Scheduling Criteria Scheduling Algorithms Multiple-Processor Scheduling Real-Time Scheduling Algorithm Evaluation.
 Set up is January 7, 2016  Judging is January 8, 2016.
Lecture 4 CPU scheduling. Basic Concepts Single Process  one process at a time Maximum CPU utilization obtained with multiprogramming CPU idle :waiting.
Name : Mamatha J M Seminar guide: Mr. Kemparaju. GRID COMPUTING.
1 Performance Impact of Resource Provisioning on Workflows Gurmeet Singh, Carl Kesselman and Ewa Deelman Information Science Institute University of Southern.
CPU scheduling.  Single Process  one process at a time  Maximum CPU utilization obtained with multiprogramming  CPU idle :waiting time is wasted 2.
Resource Characterization Rich Wolski, Dan Nurmi, and John Brevik Computer Science Department University of California, Santa Barbara VGrADS Site Visit.
A Comparison of File System Workloads D. Roselli J. Lorch T. Anderson University of California, Berkeley.
Optimizing Distributed Actor Systems for Dynamic Interactive Services
lecture 5: CPU Scheduling
CPU SCHEDULING.
Introduction to Load Balancing:
OpenMosix, Open SSI, and LinuxPMI
Data Structure Interview Question and Answers
Scheduling Jobs Across Geo-distributed Datacenters
Process Scheduling B.Ramamurthy 9/16/2018.
Transparent Adaptive Resource Management for Middleware Systems
CPU Scheduling G.Anuradha
Load Balancing/Sharing/Scheduling Part II
Aiman H. El-Maleh Sadiq M. Sait Syed Z. Shazli
Outline System architecture Current work Experiments Next Steps
Presentation transcript:

Research Paper Example Exploiting Process Lifetime Distributions for Dynamic Load Balancing Mor Harchol-Balter Allen Downey SIGMETRICS 2006

Components of a Research Paper Background Idea Work Results Next

Background Context of the research What was known before What is the question Positioning of our work Related work In this paper: Load balancing in a network Migration thought not to help [ELZ88] We can do better

Idea Something new –Algorithm –Observations Why the paper was written In this paper: Process lifetimes are heavy tailed This can be used to perform beneficial migrations

Work Collecting data Measurements Simulations Analysis In this paper: Collect data on process lifetimes Collect data on system overheads Perform simulations –Check competitiveness –Check sensitivity –Check optimality

Results Outcome of the simulations, analysis, etc. –Tables –Graphs –Interpretation In this paper: Better performance than competition Near optimal Not sensitive to system parameters

Next Extensions Shortcomings In this paper: More realistic memory usage model Take network into account And more…

The Abstract Bckgrnd Idea Work Results Next The abstract should mainly reflect the idea and results, with the necessary minimum of background and work. Say what it is, not that it exists!

Abstract #1 We measure the distribution of lifetimes of UNIX processes and propose a functional form that fits this distribution well. We use this functional form to derive a policy for preemptive migration, and then use a trace-driven simulator to compare our proposed policy with other preemptive migration policies, and with a non-preemptive load balancing strategy. We find that, contrary to previous reports, the performance benefits of preemptive migration are significantly greater than those of non-preemptive migration, even when the memory-transfer cost is high. Using a model of migration costs representative of current systems, we find that preemptive migration reduces the mean delay (queueing and migration) by 35-50%, compared to non- preemptive migration. Bckgrnd Idea Work Results Next

Abstract #2 We consider policies for CPU load balancing in networks of workstations. We address the question of whether preemptive migration (migrating active processes) is necessary, or whether remote execution (migrating processes only at the time of birth) is sufficient for load balancing. We show that resolving this issue is strongly tied to understanding the process lifetime distribution. Our measurements indicate that the distribution of lifetimes for a UNIX process is Pareto (heavy-tailed), with a consistent functional form over a variety of workloads. We show how to apply this distribution to derive a preemptive migration policy that requires no hand-tuned parameters. We use a trace driven simulation to show that our preemptive migration strategy is far more effective than remote execution, even when the memory transfer cost is high. Bckgrnd Idea Work Results Next

Abstract #3 We consider policies for CPU load balancing in networks of workstations, and specifically the use of preemptive migration (migrating active processes). We start by measuring the distribution of lifetimes of UNIX processes, and find that it is Pareto (heavy-tailed), with a consistent functional form over a variety of workloads. This distribution has the characteristic that processes that have already run for T time are expected to continue to run for an additional T time. Our policy uses this to identify those processes that will benefit from migration and justify the costs involved. We use a trace driven simulation to show that our preemptive migration strategy is far more effective than remote execution, even when the memory transfer cost is high. Bckgrnd Idea Work Results Next