1 Exploiting Nonstationarity for Performance Prediction Christopher Stewart (University of Rochester) Terence Kelly and Alex Zhang (HP Labs)

Slides:



Advertisements
Similar presentations
Topics to be discussed Introduction Performance Factors Methodology Test Process Tools Conclusion Abu Bakr Siddiq.
Advertisements

Experimental Design, Response Surface Analysis, and Optimization
Comprehensive Depiction of Configuration-dependent Performance Anomalies in Distributed Server Systems Christopher Stewart, Ming Zhong, Kai Shen, and Thomas.
Cloud Computing Resource provisioning Keke Chen. Outline  For Web applications statistical Learning and automatic control for datacenters  For data.
Copyright © 2005 Department of Computer Science CPSC 641 Winter PERFORMANCE EVALUATION Often in Computer Science you need to: – demonstrate that.
Detecting Transient Bottlenecks in n-Tier Applications through Fine- Grained Analysis Qingyang Wang Advisor: Calton Pu.
Proactive Prediction Models for Web Application Resource Provisioning in the Cloud _______________________________ Samuel A. Ajila & Bankole A. Akindele.
Adapted from Menascé & Almeida.1 Workload Characterization for the Web.
Variability in Architectural Simulations of Multi-threaded Workloads Alaa R. Alameldeen and David A. Wood University of Wisconsin-Madison
ITEC 451 Network Design and Analysis. 2 You will Learn: (1) Specifying performance requirements Evaluating design alternatives Comparing two or more systems.
Chapter 12 - Forecasting Forecasting is important in the business decision-making process in which a current choice or decision has future implications:
OS Fall ’ 02 Performance Evaluation Operating Systems Fall 2002.
Nick Trebon, Alan Morris, Jaideep Ray, Sameer Shende, Allen Malony {ntrebon, amorris, Department of.
Copyright © 1998 Wanda Kunkle Computer Organization 1 Chapter 2.1 Introduction.
1 PERFORMANCE EVALUATION H Often in Computer Science you need to: – demonstrate that a new concept, technique, or algorithm is feasible –demonstrate that.
1 Automatic Request Categorization in Internet Services Abhishek B. Sharma (USC) Collaborators: Ranjita Bhagwan (MSR, India) Monojit Choudhury (MSR, India)
Bandwidth Allocation in a Self-Managing Multimedia File Server Vijay Sundaram and Prashant Shenoy Department of Computer Science University of Massachusetts.
© 2008 Hewlett-Packard Development Company, L.P. The information contained herein is subject to change without notice Automated Workload Management in.
1 Part VI System-level Performance Models for the Web © 1998 Menascé & Almeida. All Rights Reserved.
By- Jaideep Moses, Ravi Iyer , Ramesh Illikkal and
Slides 13b: Time-Series Models; Measuring Forecast Error
Using Standard Industry Benchmarks Chapter 7 CSE807.
Performance of Web Applications Introduction One of the success-critical quality characteristics of Web applications is system performance. What.
1 Validation & Verification Chapter VALIDATION & VERIFICATION Very Difficult Very Important Conceptually distinct, but performed simultaneously.
1 An SLA-Oriented Capacity Planning Tool for Streaming Media Services Lucy Cherkasova, Wenting Tang, and Sharad Singhal HPLabs,USA.
Bottlenecks: Automated Design Configuration Evaluation and Tune.
Software Reliability SEG3202 N. El Kadri.
August 01, 2008 Performance Modeling John Meisenbacher, MasterCard Worldwide.
Modeling and Performance Evaluation of Network and Computer Systems Introduction (Chapters 1 and 2) 10/4/2015H.Malekinezhad1.
Profile Driven Component Placement for Cluster-based Online Services Christopher Stewart (University of Rochester) Kai Shen (University of Rochester) Sandhya.
1 Performance Evaluation of Computer Systems and Networks Introduction, Outlines, Class Policy Instructor: A. Ghasemi Many thanks to Dr. Behzad Akbari.
RECON: A TOOL TO RECOMMEND DYNAMIC SERVER CONSOLIDATION IN MULTI-CLUSTER DATACENTERS Anindya Neogi IEEE Network Operations and Management Symposium, 2008.
Test Loads Andy Wang CIS Computer Systems Performance Analysis.
Chapter 3 System Performance and Models. 2 Systems and Models The concept of modeling in the study of the dynamic behavior of simple system is be able.
© 2006 Hewlett-Packard Development Company, L.P. The information contained herein is subject to change without notice Profiling and Modeling Resource Usage.
1 Wenguang WangRichard B. Bunt Department of Computer Science University of Saskatchewan November 14, 2000 Simulating DB2 Buffer Pool Management.
ICOM 6115: Computer Systems Performance Measurement and Evaluation August 11, 2006.
Ó 1998 Menascé & Almeida. All Rights Reserved.1 Part V Workload Characterization for the Web (Book, chap. 6)
Queueing Models with Multiple Classes CSCI 8710 Tuesday, November 28th Kraemer.
Validating an Access Cost Model for Wide Area Applications Louiqa Raschid University of Maryland CoopIS 2001 Co-authors V. Zadorozhny, T. Zhan and L. Bright.
Chapter 10 Verification and Validation of Simulation Models
Tivoli Software © 2010 IBM Corporation 1 Using Machine Learning Techniques to Enhance The Performance of an Automatic Backup and Recovery System Amir Ronen,
Ó 1998 Menascé & Almeida. All Rights Reserved.1 Part V Workload Characterization for the Web.
Chapter 3 System Performance and Models Introduction A system is the part of the real world under study. Composed of a set of entities interacting.
Modeling Virtualized Environments in Simalytic ® Models by Computing Missing Service Demand Parameters CMG2009 Paper 9103, December 11, 2009 Dr. Tim R.
NETE4631: Network Information System Capacity Planning (2) Suronapee Phoomvuthisarn, Ph.D. /
OPERATING SYSTEMS CS 3530 Summer 2014 Systems and Models Chapter 03.
Ó 1998 Menascé & Almeida. All Rights Reserved.1 Part VI System-level Performance Models for the Web (Book, Chapter 8)
Emir Halepovic, Jeffrey Pang, Oliver Spatscheck AT&T Labs - Research
Quality Is in the Eye of the Beholder: Meeting Users ’ Requirements for Internet Quality of Service Anna Bouch, Allan Kuchinsky, Nina Bhatti HP Labs Technical.
Network Weather Service. Introduction “NWS provides accurate forecasts of dynamically changing performance characteristics from a distributed set of metacomputing.
Simulation. Types of simulation Discrete-event simulation – Used for modeling of a system as it evolves over time by a representation in which the state.
1 Evaluation of Cooperative Web Caching with Web Polygraph Ping Du and Jaspal Subhlok Department of Computer Science University of Houston presented at.
If you have a transaction processing system, John Meisenbacher
Test Loads Andy Wang CIS Computer Systems Performance Analysis.
1 Performance Impact of Resource Provisioning on Workflows Gurmeet Singh, Carl Kesselman and Ewa Deelman Information Science Institute University of Southern.
Introduction to Performance Tuning Chia-heng Tu PAS Lab Summer Workshop 2009 June 30,
Dynamic Resource Allocation for Shared Data Centers Using Online Measurements By- Abhishek Chandra, Weibo Gong and Prashant Shenoy.
CSE 340 Computer Architecture Summer 2016 Understanding Performance.
Ó 1998 Menascé & Almeida. All Rights Reserved.1 Part VI System-level Performance Models for the Web.
OPERATING SYSTEMS CS 3502 Fall 2017
Network Performance and Quality of Service
Chapter 10 Verification and Validation of Simulation Models
Computer Systems Performance Evaluation
Architecture & System Performance
Hardware Counter Driven On-the-Fly Request Signatures
Performance Evaluation of Computer Networks
Computer Systems Performance Evaluation
Performance Evaluation of Computer Networks
Presentation transcript:

1 Exploiting Nonstationarity for Performance Prediction Christopher Stewart (University of Rochester) Terence Kelly and Alex Zhang (HP Labs)

2 Motivation Enterprise applications are hard manage  Complex software hierarchy executes on (globally) distributed platforms  Application-level performance metrics are more complicated than system-level metrics  Infrastructure is fragile; system modifications (even for measurement purposes) are not always practical for real applications

3 Previous Work Performance models ease the burden of system management  Reduce complex system configurations to end-user response time or throughput prediction  Achieved via kernel modification [ barham-osdi-2004 ], runtime libraries [ chandra-eurosys-2007 ], and controlled benchmarking [stewart-nsdi-2005,urgoankar-sigmetrics-2005] Can we apply model-driven system management when intrusive measurement tools are impractical?

4 Observation Relative frequencies of transaction types in real enterprise applications are nonstationary  i.e., they change over time Nonstationarity allows model calibration using passive observations of application-level performance and system metrics

5 An Example Desire the mean value of a metric for each transaction type  Nonstationarity allows for model calibration Solve a set a linear equations: type A = 1 type B = 2 Passive observations are sufficient to calibrate performance models for real systems Passive observations ObservationMetric value# of type A requests # of type B requests

6 Outline 1. Transaction mix nonstationarity is real  Investigate 2 production enterprise applications  Implications of nonstationarity 2. A performance model for real enterprise applications 3. Performance-aware server consolidation 4. Conclusion

7 Commercial Applications Codename: VDR  Internal business-critical HP application  Services HP users and external customers  1 week trace Codename: ACME  Large Internet retailer (circa 2000)  5-day trace

8 Nonstationarity in Real Applications VDR Application  Relative frequency of the two most popular transaction types  Each point reflects an observation during a 5- minute interval Almost every ratio is represented  Transaction-type popularity is not fixed Fraction of 2 nd Most Popular Fraction of Most Popular

9 Nonstationarity in Real Applications ACME Application  Fraction of “add-to-cart” transactions in the ACME workload  Each point reflects an observation during a 5- minute window Frequencies vary by 2 orders of magnitude Time (hours)

10 Implications of Nonstationarity Performance models  A wide-range of transaction mixes is a first-order concern for real production applications  Models that consider only request rate are likely to provide poor predictive accuracy under real-world conditions

11 Implications of Nonstationarity Workload generators  Popular benchmarks (e.g., RUBiS and TPC-W) use first-order Markov models  First-order Markov models yield stationary mixes (in the long term) RUBiS browse-mix shown  Rethink workload generation Fraction of 2 nd Most Popular Fraction of Most Popular

12 Outline 1. Transaction mix nonstationarity is real 2. A performance model for real enterprise applications  Passive observations in real applications  Model design  Model validation 3. Performance-aware server consolidation 4. Conclusion

13 Model Overview Measurements under real workloads are sufficient (with some analytics) to predict application-level performance We will carefully build a model that can be calibrated from passive observations of response times and resource utilizations

14 Passive Observations Certain system metrics are easy-to-acquire and widely available in production environments  Response times, CPU, and disk utilizations are routinely collected by tools in commodity Operating Systems Passive observations from VDR (abbreviated) 5-min interval (i) Sum of Resp. (y) CPU util.Transaction Count (Per-type) type 1 (N 1 )type 2 (N 2 )type 3 (N 3 ) sec6.5% sec24% sec1.2%41914

15 Model Design Each term considers one aspect of response time The first term considers service time  N ij - The count of transaction type j in interval i   j - Typical service time of transaction type j

16 Model Design The second term considers queuing delay  U ir - The utilization of resource r at interval i  i - The arrival rate of all transactions during interval i Resource utilization is not known a priori  Independently calibrated as a function of transaction mix

17 Model Calibration For performance prediction, we must acquire  j  The second term is constant for each interval i  Solve (minimize error) a set of linear equations Regression technique: least absolute residuals (LAR)  Robust to outliers, no tunable parameters, maximizes retrospective accuracy

18 Model Validation VDR trace  ½ for calibration  ½ for prediction Our model robustly predicts past and future performance min intervals (in trace order) Sum of Response Times (sec.)

19 Model Validation VDR trace Median Error  7% calibrated set  9% predicted set ACME 12% median predictive error An accurate model from passive observations 0% 20% 40% 60% 80% 100% 0% 50% 100% 150% Absolute Percentage Error | predict – actual | / actual CDF

20 Outline 1. Transaction mix nonstationarity is real 2. Performance prediction for real enterprise applications 3. Performance-aware server consolidation  Problem statement  Extending our model for server consolidation  Validation 4. Conclusion

21 Problem Statement Performance-aware server consolidation  Given passive observations of enterprise applications running separately  Predict post-consolidation performance for each application  For this work, the hardware platform does not change

22 Performance-Aware Server Consolidation Post-consolidation performance model  Application consolidation primarily affects the queuing delay for each application  Simplifying assumption: Post-consolidation utilization is the sum of pre-consolidation utilizations

23 Validation Experimental setup  RUBiS and StockOnline  Custom nonstationary workloads Observed on ACME-variant Consolidated on VDR-variant  10-hour consolidation with 30 second measurement intervals Passively calibrated model predicts post-consolidation performance Median error 6% and 11% 0% 20% 40% 60% 80% 100% 0% 20% 40% 60% 80% Absolute Percentage Error | predict – actual | / actual CDF

24 Outline 1. Transaction mix nonstationarity is real 2. Performance prediction for real enterprise applications 3. Performance-aware server consolidation  Problem statement  Model-driven server consolidation  Validation 4. Conclusion

25 Future Work 1. Performance prediction across multi-core processor configurations Passive observations calibrate simple yet effective models of processor utilization 2. Performance anomaly depiction Predictions are used to identify situations where performance does not match model expectations [stewart-hotdep-2006, kelly-worlds-2005]

26 Take Away Points Transaction mix nonstationarity is a real phenomenon in production applications Passive observations are sufficient to calibrate performance models Passively calibrated performance models can guide system management decisions