1 Admission Control and Request Scheduling in E-Commerce Web Sites Sameh Elnikety, EPFL Erich Nahum, IBM Watson John Tracey, IBM Watson Willy Zwaenepoel,

Slides:



Advertisements
Similar presentations
Advanced Oracle DB tuning Performance can be defined in very different ways (OLTP versus DSS) Specific goals and targets must be set => clear recognition.
Advertisements

1 SLA Enforcement in the Service Cloud Anja Grünheid Fakultät für Informatik Lehrstuhl III - Datenbanksysteme.
The Effects of Wide-Area Conditions on WWW Server Performance Erich Nahum, Marcel Rosu, Srini Seshan, Jussara Almeida IBM T.J. Watson Research Center,
Exploiting Distributed Version Concurrency in a Transactional Memory Cluster Kaloian Manassiev, Madalin Mihailescu and Cristiana Amza University of Toronto,
Operating Systems Process Scheduling (Ch 3.2, )
CPU Scheduling Questions answered in this lecture: What is scheduling vs. allocation? What is preemptive vs. non-preemptive scheduling? What are FCFS,
1 Database Replication Using Generalized Snapshot Isolation Sameh Elnikety, EPFL Fernando Pedone, USI Willy Zwaenepoel, EPFL.
Benchmarking Cloud Serving Systems with YCSB Brian F. Cooper, Adam Silberstein, Erwin Tam, Raghu Ramakrishnan, Russell Sears Yahoo! Research Presenter.
Predicting Replicated Database Scalability Sameh Elnikety, Microsoft Research Steven Dropsho, Google Inc. Emmanuel Cecchet, Univ. of Mass. Willy Zwaenepoel,
Handling Web Hotspots at Dynamic Content Web Sites Using DotSlash Weibin Zhao Henning Schulzrinne Columbia University NYMAN’04.
Performance Evaluation
Handling Web Hotspots at Dynamic Content Web Sites Using DotSlash Weibin Zhao Henning Schulzrinne Columbia University Dagstuhl.
1 Multiple class queueing networks Mean Value Analysis - Open queueing networks - Closed queueing networks.
Adaptive Content Delivery for Scalable Web Servers Authors: Rahul Pradhan and Mark Claypool Presented by: David Finkel Computer Science Department Worcester.
Overview  Strong consistency  Traditional approach  Proposed approach  Implementation  Experiments 2.
Yaksha: A Self-Tuning Controller for Managing the Performance of 3-Tiered Web Sites Abhinav Kamra, Vishal Misra CS Department Columbia University Erich.
OS Fall ’ 02 Performance Evaluation Operating Systems Fall 2002.
Towards Autonomic Hosting of Multi-tier Internet Services Swaminathan Sivasubramanian, Guillaume Pierre and Maarten van Steen Vrije Universiteit, Amsterdam,
Computer Science Cataclysm: Policing Extreme Overloads in Internet Applications Bhuvan Urgaonkar and Prashant Shenoy University of Massachusetts.
Towards Eco-friendly Database Management Systems W. Lang, J. M. Patel (U Wisconsin), CIDR 2009 Shimin Chen Big Data Reading Group.
SSP Capacity Planning. One powerful machine with everything.
SEDA: An Architecture for Well-Conditioned, Scalable Internet Services
Database Replication Policies for Dynamic Content Applications Gokul Soundararajan, Cristiana Amza, Ashvin Goel University of Toronto EuroSys 2006: Leuven,
U NIVERSITY OF M ASSACHUSETTS, A MHERST – Department of Computer Science An Analytical Model for Multi-tier Internet Services and its Applications Bhuvan.
1 Scheduling Processes. 2 Processes Each process has state, that includes its text and data, procedure call stack, etc. This state resides in memory.
Profile Driven Component Placement for Cluster-based Online Services Christopher Stewart (University of Rochester) Kai Shen (University of Rochester) Sandhya.
Ideas to Improve SharePoint Usage 4. What are these 4 Ideas? 1. 7 Steps to check SharePoint Health 2. Avoid common Deployment Mistakes 3. Analyze SharePoint.
Scaling Dynamic Content Applications through Data Replication - Opportunities for Compiler Optimizations Cristiana Amza UofT.
1 Specification and Implementation of Dynamic Web Site Benchmarks Sameh Elnikety Department of Computer Science Rice University.
Network Computing Laboratory Experiment Tutorial Network Computing Lab
Architectural Characterization of an IBM RS6000 S80 Server Running TPC-W Workloads Lei Yang & Shiliang Hu Computer Sciences Department, University of.
Simulating a $2M Commercial Server on a $2K PC Alaa R. Alameldeen, Milo M.K. Martin, Carl J. Mauer, Kevin E. Moore, Min Xu, Daniel J. Sorin, Mark D. Hill.
CPU Scheduling CSCI 444/544 Operating Systems Fall 2008.
Applications Web et bases de données en grappe Séminaire InTech 3 Février 2005 – Grenoble.
Intro – Part 2 Introduction to Database Management: Ch 1 & 2.
Usenix Annual Conference, Freenix track – June 2004 – 1 : Flexible Database Clustering Middleware Emmanuel Cecchet – INRIA Julie Marguerite.
DotSlash: Handling Web Hotspots at Dynamic Content Web Sites Weibin Zhao Henning Schulzrinne Department of Computer Science Columbia.
Difference of Degradation Schemes among Operating Systems -Experimental analysis for web application servers- Hideaki Hibino*(Tokyo Tech) Kenichi Kourai.
TPC BENCHMARK W (Web Commerce) SeungLak Choi Dept. of Computer Science, KAIST.
Database Replication in Tashkent CSEP 545 Transaction Processing Sameh Elnikety.
Online Music Store. MSE Project Presentation III
A Method for Transparent Admission Control and Request Scheduling in E-Commerce Web Sites S. Elnikety, E. Nahum, J. Tracey and W. Zwaenpoel Presented By.
Design and Evaluation of a Model for Multi-tiered Internet Applications Bhuvan Urgaonkar Internship project talk – Services Management Middleware Dept,
Request Distribution in Server Clusters Krithi Ramamritham Indian Institute of Technology Bombay.
© 2006 Hewlett-Packard Development Company, L.P. The information contained herein is subject to change without notice Injecting Realistic Burstiness to.
Measuring the Capacity of a Web Server USENIX Sympo. on Internet Tech. and Sys. ‘ Koo-Min Ahn.
DotSlash – or how to deal with 15 minutes of fame Weibin Zhao Henning Schulzrinne Columbia University CATT/WICAT Annual Research Review November 14, 2003.
DynamicMR: A Dynamic Slot Allocation Optimization Framework for MapReduce Clusters Nanyang Technological University Shanjiang Tang, Bu-Sung Lee, Bingsheng.
JSP Server Integrated with Oracle8i Project2, CMSC691X Summer02 Ching-li Peng Ying Zhang.
When Average is Not Average: Large Response Time Fluctuations in n-Tier Applications Qingyang Wang, Yasuhiko Kanemasa, Calton Pu, Motoyuki Kawaba.
R. Jiménez-Peris Scalability Evaluation of the Replication Support of JOnAS, an Industrial J2EE Application Server A. Paz, F. P é rez-Sorrosal, M. Patiño-Martínez,
A Comparative Evaluation of Transparent Scaling Techniques for Dynamic Content Servers Presented by Chen Zhang Written by C. Amza, A. L. Cox,
An Architectural Evaluation of Java TPC-W Harold “Trey” Cain, Ravi Rajwar, Morris Marden, Mikko Lipasti University of Wisconsin-Madison
Admission Control and Request Scheduling in Dynamic E-Commerce Web Sites Sameh Elnikety, Erich Nahum, John Tracey, Willy Zwaenepoel C.S. Dept. EPFL IBM.
Zeta: Scheduling Interactive Services with Partial Execution Yuxiong He, Sameh Elnikety, James Larus, Chenyu Yan Microsoft Research and Microsoft Bing.
Cofax Scalability Document Version Scaling Cofax in General The scalability of Cofax is directly related to the system software, hardware and network.
Spark on Entropy : A Reliable & Efficient Scheduler for Low-latency Parallel Jobs in Heterogeneous Cloud Huankai Chen PhD Student at University of Kent.
TPC Benchmark™ W 2002 년 6 월 이상호 교수 숭실대학교 데이터베이스 연구실
SharePoint 2010 Performance and Capacity Planning Best Practices
Abhinav Kamra, Vishal Misra CS Department Columbia University
Regulating Data Flow in J2EE Application Server
PHP / MySQL Introduction
Process Scheduling B.Ramamurthy 9/16/2018.
Capacity Analysis, cont. Realistic Server Performance
Building a Database on S3
DotSlash: An Automated Web Hotspot Rescue System
Admission Control and Request Scheduling in E-Commerce Web Sites
Simulating a $2M Commercial Server on a $2K PC
Multiple-resource Request Scheduling. for Differentiated QoS
Presentation transcript:

1 Admission Control and Request Scheduling in E-Commerce Web Sites Sameh Elnikety, EPFL Erich Nahum, IBM Watson John Tracey, IBM Watson Willy Zwaenepoel, EPFL

2 Dynamic Content Is Important 1 3 2

3 Generating Dynamic Content http Database Server Web Server Dynamic Content Generator

4 Objective: Stable Performance Prevent overload (admission control) Sustain peak throughput in overload region Load Throughput Actual Ideal UnderloadoverloadSaturation

5 Objective: Improve Response Time Excess requests are queued Improve response time (request scheduling)

6 Main Idea: Gatekeeper Proxy Transparent Intercept requests to the bottleneck Maintains measurement-based estimates Performs: –Admission control –Request scheduling http Web Server Dynamic Content Generator Gate Keeper Database Server Transparently control how requests are admitted (non-invasive)

7 Admission Control Sustain peak throughput during overload Amount of work required by each request Capacity of the system

8 Estimating Work by Request Type Key Observations: –Finite number of request types –Requests of same type take same execution time –Different requests differ greatly in execution time –Online measurements Gatekeeper maintains per-request estimates http Web Server Dynamic Content Generator Gate Keeper Database Server

9 Service Time Distributions Request type is key! Online measurement

10 Estimating System Capacity Request load = –Execution time (work units of required) Database capacity = –max # work units before overload To determine system capacity –Binary search –Offline

11 req typeunits R15 R2500 R3300 …… R1 R2R R2R R1R R1 R2 R R2 Admission Control - Example 6 Table: maintain for each type, online estimates of service time (moving average)

12 Request Scheduling - Example (0+500) + (500+10) = 1010  505 (0+10) + (10+500) = 520  260

13 Request Scheduling Reduce average response time Use shortest job first (SJF) policy Reorder requests in admission queue No preemption

14 Large Variability: TPC-W Requests 95% require < 1000 ms Scheduling has high impact

15 Request Scheduling: Aging Prevent starvation Limit the delay due to scheduling Limit is X times “expected service time” Big delay due to sched.

16 Outline Motivation & Background The Gatekeeper Proxy Experimental Environment –Software & Hardware –Metrics & Methodology Results Summary and Conclusions

17 TPC-W Benchmark Transaction Processing Council (TPC-W) –Workload generator Models a large e-commerce site: –Online bookstore –Searching, browsing, buying, registration, … Persistent data –Static images on web server –All others on back-end database

18 TPC-W Benchmark - Snapshot Promotion (ad) Shopping Cart Next Interaction Image

19 TPC-W: Interactions 14 interactions, e.g.: –Home (read-only query) –Best sellers (complex) –Secure payment (ssl) –Shopping cart (update query) Scale –10,000 items –288,000 clients –350 MB database (fits in main memory) –183 MB images (in file system of web server)

20 Software GateKeeper Implemented in Java (JDBC driver) Web Server Apache App Server Jakarta Tomcat Database MySQL OS RedHat 7.2 Linux http Web Server Dynamic Content Generator Gate Keeper Database Server

21 Hardware CPU AMD Athlon 1.33 GHz Memory 768 MB Disk 60 GB, 12 msec, 5400 rpm Network 100 Mbps Ethernet http Apache Tomcat GateKeeper MySQL sql

22 http Apache Tomcat GateKeeper MySQL sql Emulated Clients Emulated Clients Client emulator – Session duration – Think time – Markov model Load is a function of the number of clients

23 Experiments Performance Metrics –Throughput (interactions/minute) –Response time (msec, submission to completion) –Examine each as a function of load (# of clients) Methodology –Average of 5 runs –100 second warm-up –600 second measurement

24 Admission Control - Throughput db processes: 345 used mem: 509 MB db processes: 233 used mem: 450 MB db processes: 49 used mem: 275 MB 3 1 2

25 Request Scheduling - Response Time

26 Request Scheduling - Analysis Waiting timeExecution (service) timeResponse time = Waiting time + Execution (service) time Large variability –Many short requests –Few very large requests

27 Request Scheduling - Explanation Average job

28 Request Scheduling - Explanation Short jobAverage jobLong job

29 Request Scheduling - Fairness Waiting timeExecution (service) timeResponse time = Waiting time + Execution (service) time Fairness trade-off –FIFO Fair: all wait for same amount of time –SJF Unfair: favors short requests Better average response time

30 Aging: Prevent Starvation ∞

31 In The Paper More results –Different bottleneck (database lock contention) –Online vs. offline measurements –DB2 Related work –Most other methods are invasive

32 Summary Presented the Gatekeeper proxy – Transparent (non-invasive) – Intercept requests – Online measurements Admission control – Consistent performance during overload – Improves throughput 10 % Request scheduling using SJF – Improves response time 14 times – Penalizes long jobs only 13 % – Aging controls penalty

33 Thank You!