Realistic CPU Workloads Through Host Load Trace Playback Peter A. Dinda David R. O’Hallaron Carnegie Mellon University.

Slides:



Advertisements
Similar presentations
Full-System Timing-First Simulation Carl J. Mauer Mark D. Hill and David A. Wood Computer Sciences Department University of Wisconsin—Madison.
Advertisements

A Prediction-based Real-time Scheduling Advisor Peter A. Dinda Carnegie Mellon University.
An Evaluation of Linear Models for Host Load Prediction Peter A. Dinda David R. O’Hallaron Carnegie Mellon University.
G. Alonso, D. Kossmann Systems Group
Trace-based Network Bandwidth Analysis and Prediction Yi QIAO 06/10/2002.
Enabling Efficient On-the-fly Microarchitecture Simulation Thierry Lafage September 2000.
Silberschatz, Galvin and Gagne  2002 Modified for CSCI 399, Royden, Operating System Concepts Operating Systems Lecture 19 Scheduling IV.
Workloads Experimental environment prototype real sys exec- driven sim trace- driven sim stochastic sim Live workload Benchmark applications Micro- benchmark.
SKELETON BASED PERFORMANCE PREDICTION ON SHARED NETWORKS Sukhdeep Sodhi Microsoft Corp Jaspal Subhlok University of Houston.
Synthesizing Representative I/O Workloads for TPC-H J. Zhang*, A. Sivasubramaniam*, H. Franke, N. Gautam*, Y. Zhang, S. Nagar * Pennsylvania State University.
The Case For Prediction-based Best-effort Real-time Peter A. Dinda Bruce Lowekamp Loukas F. Kallivokas David R. O’Hallaron Carnegie Mellon University.
Host Load Trace Replay Peter A. Dinda Thesis Seminar 11/23/98.
An Adaptable Benchmark for MPFS Performance Testing A Master Thesis Presentation Yubing Wang Advisor: Prof. Mark Claypool.
Modeling Host Load Peter A. Dinda Thesis Seminar 2/9/98.
Lecture 5: Learning models using EM
Responsive Interactive Applications by Dynamic Mapping of Activation Trees February 20, 1998 Peter A. Dinda School of Computer.
70-290: MCSE Guide to Managing a Microsoft Windows Server 2003 Environment Chapter 11: Monitoring Server Performance.
Understanding and Predicting Host Load Peter A. Dinda Carnegie Mellon University
OS Fall ’ 02 Performance Evaluation Operating Systems Fall 2002.
Resource Signal Prediction and Its Application to Real-time Scheduling Advisors or How to Tame Variability in Distributed Systems Peter A. Dinda Carnegie.
Performance Evaluation
Online Prediction of the Running Time Of Tasks Peter A. Dinda Department of Computer Science Northwestern University
1 Dong Lu, Peter A. Dinda Prescience Laboratory Computer Science Department Northwestern University Virtualized.
OS Fall ’ 02 Performance Evaluation Operating Systems Fall 2002.
A Prediction-based Approach to Distributed Interactive Applications Peter A. Dinda Department of Computer Science Northwestern University
A Prediction-based Real-time Scheduling Advisor Peter A. Dinda Prescience Lab Department of Computer Science Northwestern University
Load Analysis and Prediction for Responsive Interactive Applications Peter A. Dinda David R. O’Hallaron Carnegie Mellon University.
Cmpt-225 Simulation. Application: Simulation Simulation  A technique for modeling the behavior of both natural and human-made systems  Goal Generate.
The Running Time Advisor A Resource Signal-based Approach to Predicting Task Running Time and Its Applications Peter A. Dinda Carnegie Mellon University.
Alessandro Cappelletti for CTF3 collaboration 5 th May 2010 RESULTS OF BEAM BASED RF POWER PRODUCTION IN CTF3.
Session 4. Applied Regression -- Prof. Juran2 Outline for Session 4 Summary Measures for the Full Model –Top Section of the Output –Interval Estimation.
Dynamic and Decentralized Approaches for Optimal Allocation of Multiple Resources in Virtualized Data Centers Wei Chen, Samuel Hargrove, Heh Miao, Liang.
University of Zagreb MMVE 2012 workshop1 Towards Reinterpretation of Interaction Complexity for Load Prediction in Cloud-based MMORPGs Mirko Sužnjević,
1 An SLA-Oriented Capacity Planning Tool for Streaming Media Services Lucy Cherkasova, Wenting Tang, and Sharad Singhal HPLabs,USA.
Modeling and Performance Evaluation of Network and Computer Systems Introduction (Chapters 1 and 2) 10/4/2015H.Malekinezhad1.
70-290: MCSE Guide to Managing a Microsoft Windows Server 2003 Environment, Enhanced Chapter 11: Monitoring Server Performance.
Introduction to Experimental Design
Silberschatz, Galvin and Gagne  Operating System Concepts Chapter 6: CPU Scheduling Basic Concepts Scheduling Criteria Scheduling Algorithms.
Test Loads Andy Wang CIS Computer Systems Performance Analysis.
Analysis of Simulation Results Chapter 25. Overview  Analysis of Simulation Results  Model Verification Techniques  Model Validation Techniques  Transient.
Hung X. Nguyen and Matthew Roughan The University of Adelaide, Australia SAIL: Statistically Accurate Internet Loss Measurements.
70-290: MCSE Guide to Managing a Microsoft Windows Server 2003 Environment, Enhanced Chapter 11: Monitoring Server Performance.
Characterizing Failure Data in High-Performance-Computing Systems Bianca Schroeder With Garth Gibson and Gary Grider Department of Computer Science Carnegie.
© 2003, Carla Ellis Self-Scaling Benchmarks Peter Chen and David Patterson, A New Approach to I/O Performance Evaluation – Self-Scaling I/O Benchmarks,
Carnegie Mellon University Computer Science Department 1 OPEN VERSUS CLOSED: A CAUTIONARY TALE Bianca Schroeder Adam Wierman Mor Harchol-Balter Computer.
Chapter 10 Verification and Validation of Simulation Models
1 11/29/2015 Chapter 6: CPU Scheduling l Basic Concepts l Scheduling Criteria l Scheduling Algorithms l Multiple-Processor Scheduling l Real-Time Scheduling.
Chapter 3 System Performance and Models Introduction A system is the part of the real world under study. Composed of a set of entities interacting.
Shivkumar Kalyanaraman Rensselaer Polytechnic Institute 1 Based upon slides from Jay Lepreau, Utah Emulab Introduction Shiv Kalyanaraman
Measuring the Capacity of a Web Server USENIX Sympo. on Internet Tech. and Sys. ‘ Koo-Min Ahn.
Chapter 8: Simple Linear Regression Yang Zhenlin.
OPERATING SYSTEMS CS 3530 Summer 2014 Systems and Models Chapter 03.
Department of Computer Science 6 th Annual Austin CAS Conference – 24 February 2005 Ricardo Portillo, Diana Villa, Patricia J. Teller The University of.
Test Loads Andy Wang CIS Computer Systems Performance Analysis.
CPU scheduling.  Single Process  one process at a time  Maximum CPU utilization obtained with multiprogramming  CPU idle :waiting time is wasted 2.
Dynamic Resource Allocation for Shared Data Centers Using Online Measurements By- Abhishek Chandra, Weibo Gong and Prashant Shenoy.
OPERATING SYSTEMS CS 3502 Fall 2017
Framework For Exploring Interconnect Level Cache Coherency
Applied Operating System Concepts -
Andy Wang CIS 5930 Computer Systems Performance Analysis
Chapter 6: CPU Scheduling
Module 5: CPU Scheduling
3: CPU Scheduling Basic Concepts Scheduling Criteria
Chapter 6: CPU Scheduling
Operating System , Fall 2000 EA101 W 9:00-10:00 F 9:00-11:00
Chapter 6: CPU Scheduling
Module 5: CPU Scheduling
Chapter 6: CPU Scheduling
Module 5: CPU Scheduling
Srinivas Neginhal Anantharaman Kalyanaraman CprE 585: Survey Project
Presentation transcript:

Realistic CPU Workloads Through Host Load Trace Playback Peter A. Dinda David R. O’Hallaron Carnegie Mellon University

2 Talk in a Nutshell Workloads to evaluate distributed systems Prediction-based systems Shared benchmarks Reconstruct time-varying CPU contention behavior from traces of Unix load average Real (non-parametric) Reproducible Comparable Artifacts ( Playload tool Collection of host load traces

3 Outline Evaluation of distributed systems Adaptive applications Prediction systems Host load traces Synthetic Vs trace-based workloads Host load trace playback Evaluation Conclusion

4 Evaluating Distributed Systems Workloads critical in evaluation Shared benchmarks desperately needed SPEC? Synthetic versus trace-based workloads Parametric versus non-parametric Ideally: Real workloads Reproducible Comparable Sharable

5 Evaluating Prediction Systems Prediction systems model workload RPS, NWS, adaptive applications, etc. Synthetic workloads assume a model Model may be wrong or incomplete Easy to use in simulation or in a testbed Trace-based workloads assume no model However, traces may not be representative Harder to use, especially in a testbed How do we use traces in a testbed? Host load traces

6 Host Load Traces Periodically sampled Unix load average Exponentially averaged run queue length Measure of contention for CPU Complex statistical properties [SciProg99] High variability, strong autocorrelation function, self- similar, epochal behavior, … Difficult to synthesize

7 Host Load and Running Time

8 Available Host Load Traces DEC Unix 5 second exponential average Full bandwidth captured (1 Hz sample rate) Long durations Available!

9 Host Load Measurement Sample1h { Ready Queue Sample2 unknown sample rate f=2 Hz estimated exponential average, tau=5 s f=1 Hz KernelUser Trace File

10 Host Load Trace Playback h -1 Load Generator Trace File Load Measure Sample1h { Sample2 error - applied load measured load target load

11 What are h and h -1 ? run queue length trace file time constant for recorded host applied load (recovered run queue length) h: h -1 : measured load time constant for playback host

12 Load Generator “1.5 load for 1 second” “1.0 load for 1 second”... “0.5 load for 1 second” Master... Worker Processes “0.0 load for 1 second”

13 Load Generator // Split w into n cycles while (!done) { if (uniformrand(1.0) < p) compute for w/n seconds; else sleep for w/n seconds; } “p load for w seconds” done=“w seconds have elapsed” Time-based playback done=“w*p CPU seconds have been used” Work-based playback (simplified)

14 Time-based Playback External continuous 1.0 load External load amplitude modulates applied load

15 Work-based Playback External load amplitude and frequency modulates applied load External continuous 1.0 load

16 Evaluation Traces described earlier Example uses one hour 1997 axp0 trace Characterization of signals and errors Summary stats Distributions Autocorrelation Multiple platforms Digital Unix Solaris Linux FreeBSD

17 Evaluation Summary Stats

18 Evaluation on Alpha/DUX Target MeasuredError ACF Error Histogram

19 Evaluation on Sparc/Solaris Target MeasuredError ACF Error Histogram

20 Evaluation on P2/FreeBSD Target MeasuredError ACF Error Histogram

21 Evaluation on P2/Linux Target MeasuredError ACF Error Histogram

22 Conclusion CPU Workloads from traces of Unix load average Reproduce contention behavior (ignoring priorities) Real, non-parametric workloads Reproducible Artifacts ( Playload tool Collection of host load traces Future Benchmarks Priorities, memory, disk, etc. In-kernel?

23 Feedback? Use error signal to better track the load trace h -1 Load Generator Load Measure error - h -1 + z Trace File x level

24 The Problem With Feedback Feedback would try to make SUM of applied load and external load in system track the load trace External Load h -1 Load Generator Load Measure error - h -1 + z Trace File x level Applied Load Effect of Combined Load

25 Making Feedback Work External Load h -1 Load Generator Load Measure error - h -1 + z Trace File x level Signal Separation Applied Load Effect of Combined Load Estimated Effect of Applied Load Estimated Effect of External Load Load Source Models

26 Why Host Load Traces for Evaluating Distributed Systems? Real Comparable and reproducible Analogous to a SPEC benchmark Usable in simulation and experimentation Non-parametric and non-synthetic Especially important for prediction systems