Resource-Efficient and QoS-Aware Cluster Management

Slides:



Advertisements
Similar presentations
MAP REDUCE PROGRAMMING Dr G Sudha Sadasivam. Map - reduce sort/merge based distributed processing Best for batch- oriented processing Sort/merge is primitive.
Advertisements

University of Minnesota Optimizing MapReduce Provisioning in the Cloud Michael Cardosa, Aameek Singh†, Himabindu Pucha†, Abhishek Chandra
SDN + Storage.
Application-Aware Memory Channel Partitioning † Sai Prashanth Muralidhara § Lavanya Subramanian † † Onur Mutlu † Mahmut Kandemir § ‡ Thomas Moscibroda.
LIBRA: Lightweight Data Skew Mitigation in MapReduce
Resource Management §A resource can be a logical, such as a shared file, or physical, such as a CPU (a node of the distributed system). One of the functions.
Based on the text by Jimmy Lin and Chris Dryer; and on the yahoo tutorial on mapreduce at index.html
MapReduce Online Created by: Rajesh Gadipuuri Modified by: Ying Lu.
SLA-Oriented Resource Provisioning for Cloud Computing
Locality-Aware Dynamic VM Reconfiguration on MapReduce Clouds Jongse Park, Daewoo Lee, Bokyeong Kim, Jaehyuk Huh, Seungryoul Maeng.
Cloud Computing Resource provisioning Keke Chen. Outline  For Web applications statistical Learning and automatic control for datacenters  For data.
Spark: Cluster Computing with Working Sets
Look Who’s Talking: Discovering Dependencies between Virtual Machines Using CPU Utilization HotCloud 10 Presented by Xin.
Developing a Characterization of Business Intelligence Workloads for Sizing New Database Systems Ted J. Wasserman (IBM Corp. / Queen’s University) Pat.
Scaling Distributed Machine Learning with the BASED ON THE PAPER AND PRESENTATION: SCALING DISTRIBUTED MACHINE LEARNING WITH THE PARAMETER SERVER – GOOGLE,
New Challenges in Cloud Datacenter Monitoring and Management
Identifying and Incorporating Latencies in Distributed Data Mining Algorithms Michael Sevilla.
MapReduce: Hadoop Implementation. Outline MapReduce overview Applications of MapReduce Hadoop overview.
EXPOSE GOOGLE APP ENGINE AS TASKTRACKER NODES AND DATA NODES.
A Framework for Elastic Execution of Existing MPI Programs Aarthi Raveendran Tekin Bicer Gagan Agrawal 1.
Jockey Guaranteed Job Latency in Data Parallel Clusters Andrew Ferguson, Peter Bodik, Srikanth Kandula, Eric Boutin, and Rodrigo Fonseca.
A Framework for Elastic Execution of Existing MPI Programs Aarthi Raveendran Graduate Student Department Of CSE 1.
Mesos A Platform for Fine-Grained Resource Sharing in the Data Center Benjamin Hindman, Andy Konwinski, Matei Zaharia, Ali Ghodsi, Anthony Joseph, Randy.
임규찬. 1. Abstract 2. Introduction 3. Design Goals 4. Sample-Based Scheduling for Parallel Jobs 5. Implements.
Performance Prediction for Random Write Reductions: A Case Study in Modelling Shared Memory Programs Ruoming Jin Gagan Agrawal Department of Computer and.
Bi-Hadoop: Extending Hadoop To Improve Support For Binary-Input Applications Xiao Yu and Bo Hong School of Electrical and Computer Engineering Georgia.
Grid Computing at Yahoo! Sameer Paranjpye Mahadev Konar Yahoo!
Heracles: Improving Resource Efficiency at Scale ISCA’15 Stanford University Google, Inc.
U N I V E R S I T Y O F S O U T H F L O R I D A Hadoop Alternative The Hadoop Alternative Larry Moore 1, Zach Fadika 2, Dr. Madhusudhan Govindaraju 2 1.
Scale up Vs. Scale out in Cloud Storage and Graph Processing Systems
DynamicMR: A Dynamic Slot Allocation Optimization Framework for MapReduce Clusters Nanyang Technological University Shanjiang Tang, Bu-Sung Lee, Bingsheng.
COMP7330/7336 Advanced Parallel and Distributed Computing Task Partitioning Dynamic Mapping Dr. Xiao Qin Auburn University
Spark on Entropy : A Reliable & Efficient Scheduler for Low-latency Parallel Jobs in Heterogeneous Cloud Huankai Chen PhD Student at University of Kent.
Item Based Recommender System SUPERVISED BY: DR. MANISH KUMAR BAJPAI TARUN BHATIA ( ) VAIBHAV JAISWAL( )
Re-Architecting Apache Spark for Performance Understandability Kay Ousterhout Joint work with Christopher Canel, Max Wolffe, Sylvia Ratnasamy, Scott Shenker.
SketchVisor: Robust Network Measurement for Software Packet Processing
Image taken from: slideshare
Optimizing Distributed Actor Systems for Dynamic Interactive Services
TensorFlow– A system for large-scale machine learning
Efficient Multi-User Indexing for Secure Keyword Search
Introduction to Distributed Platforms
Operating Systems : Overview
Curator: Self-Managing Storage for Enterprise Clusters
Chapter 10 Data Analytics for IoT
Edinburgh Napier University
Spark Presentation.
CS 425 / ECE 428 Distributed Systems Fall 2016 Nov 10, 2016
Dynamic Graph Partitioning Algorithm
Resource Elasticity for Large-Scale Machine Learning
CS 425 / ECE 428 Distributed Systems Fall 2017 Nov 16, 2017
Computing Resource Allocation and Scheduling in A Data Center
Cloud Data Anonymization Using Hadoop Map-Reduce Framework With Qos Evaluation and Behaviour analysis PROJECT GUIDE: Ms.S.Subbulakshmi TEAM MEMBERS: A.Mahalakshmi( ).
PA an Coordinated Memory Caching for Parallel Jobs
Hadoop Clusters Tess Fulkerson.
Software Engineering Introduction to Apache Hadoop Map Reduce
A Framework for Automatic Resource and Accuracy Management in A Cloud Environment Smita Vijayakumar.
Collaborative Offloading for Distributed Mobile-Cloud Apps
湖南大学-信息科学与工程学院-计算机与科学系
Cse 344 May 4th – Map/Reduce.
CS110: Discussion about Spark
Ch 4. The Evolution of Analytic Scalability
Soft Error Detection for Iterative Applications Using Offline Training
Henge: Intent-Driven Multi-Tenant Stream Processing
Smita Vijayakumar Qian Zhu Gagan Agrawal
HPML Conference, Lyon, Sept 2018
Pei Fan*, Ji Wang, Zibin Zheng, Michael R. Lyu
Subject Name: Operating System Concepts Subject Number:
Progress Report 2014/04/23.
Cloud Computing Large-scale Resource Management
MapReduce: Simplified Data Processing on Large Clusters
Presentation transcript:

Resource-Efficient and QoS-Aware Cluster Management Quasar Resource-Efficient and QoS-Aware Cluster Management

Motivation An increasing amount of cloud computing. Most cloud facilities operate at very low utilization. CPU utilization: below 20% Reservations: up to 80%

: Resource utilization over 30 days for a large production cluster at Twitter

Challenges Oversized reservations Resource allocation Resource assignment Update for hardware and workload data Oversized reservations Many reasons: they need low-latency services, so they keep their software dedicated servers while these server operate at low utilization most of the time side-step performance unpredictability due to interference Resource allocation Right amount resouce and right kind of resouce, cpu or memory intensive mission. Resource assignment Select the server for this mission Update for hardware and workload data it is impractical for the cluster manager to analyze all possible combinations of resource allocations and assignments

Overview shifts from a reservation-centric to a performance-centric approach fast classification techniques to determine the impact of different resource allocations and assignments on workload performance performs resource allocation and assignment jointly shifts from a reservation-centric to a performance-centric approach User only provide performance constraints of the application. Throughput, latency, depending on the application type. Then Quasar determine the least amount of the available resources needed to meet performance constraints at any point. fast classification techniques to determine the impact of different resource allocations and assignments on workload performance performs resource allocation and assignment jointly right amount and specific set of resources assigned to the workload

Performance-centric User only provide performance constraints of the application. latency QoS execution time instructions-persecond For latency-critical workloads, constraints are expressed as a queries per second (QPS) target and a latency QoS constraint. For distributed frameworks like Hadoop, the constraint is execution time. For single node, single-threaded or multi-threaded workloads the constraint is a low-level metric of instructions-persecond (IPS). Once performance constraints are specified, it is up to Quasar to find a resource allocation and assignment that satisfies them

Fast and Accurate Classification

Example: Scale-up classification profile it briefly with two randomly-selected scale-up allocations. matrix A: Workloads as rows indexes Scale-up configurations as columns indexes Performance as elements collaborative filtering to estimate the missing parts. When a workload is submitted, we profile it briefly with two randomly-selected scale-up allocations. The parameters and duration of profiling depend on workload type.

Fast and Accurate Classification collaborative filtering Singular Value Decomposition (SVD) and PQ- reconstruction with SGD estimate the impact of change in resource on application performance. classifications for different workload types The input to SVD in this case is a very sparse matrix A with users as rows, movies as columns and ratings as elements. SVD decomposes A to the product of the matrix of singular values Σ that represents similarity concepts in A, the matrix of left singular vectors U that represents correlation between rows of A and similarity concepts, and the matrix of right singular vectors V that represents the correlation between columns of A and similarity concepts (A = U ·Σ·V T ). A similarity concept can be that users that liked “Lord of the Rings 1” also liked “Lord of the Rings 2” PQ-reconstruction with Stochastic Gradient Descent (SGD), a simple latent-factor model [13, 66], uses Σ, U, and V to reconstruct the missing entries in A

scale-up (amount of resources per server), scale-out (number of servers per workload), server configuration, interference(which workloads can be co-located)

Greedy Allocation and Assignment sizes the allocation based on available resources until the performance constraint is met allocate the least amount of resources needed to satisfy a workload’s performance target. 1first rank the available servers by decreasing resource quality 2sizes the allocation based on available resources until the performance constraint is met 3first increases the per-node resources (scale-up) to better pack work in few servers, and then distributes the load across machines (scaleout).

Implementation Dynamic Adaptation Side Effect Free Profiling Phase detection Allocation adjustment Side Effect Free Profiling Stragglers Phase detection: If a workload runs below its performance constraint, it either went through a phase change or was incorrectly classified or assigned. In any case, Quasar reclassifies the application at its current state and adjusts its resources as needed Side Effect Free Profiling To acquire the profiling data needed for classification, we must launch multiple copies of the incoming application. However, this may cause inconsistencies with intermediate results, duplicate entries in databases, or data corruption on file systems. To eliminate such issues, Quasar uses sandboxing for the training copies during profiling. Stragglers: In some cases, (like Hadoop or Spark), individual tasks may take much longer to complete for reasons that range from poor work partitioning to network interference and machine instability. Quasar injects two contentious microbenchmarks in the corresponding servers and reclassifies them. If the new classification differ from the original by more than 20%, we signal the task as a straggler and relaunch it on a newly assigned server.

Evaluation Single Batch Job: Efficiency: Performance: for ten Hadoop jobs, improved by an average of 29%. Efficiency: a single job running at a time on the small cluster, excuting time Analytics frameworks like Hadoop

Multiple Batch Frameworks: Average, performance improves by 27% Increases utilization, achieving 62% on average versus 34%. The cluster is shared between jobs from multiple analytics frameworks(Hadoop, Storm, and Spark)

Low-Latency Service: Throughput for HotCRP Three traffic scenarios: flat, fluctuating, and large spike Quasar: 5% runtime at minimum Queries per Second. Auto-scale: 24% runtime at minimum Queries per Second. use three traffic scenarios: flat, fluctuating, and large spike

Stateful Latency-Critical Services auto-scaling manager degrades throughput 24% for Memcached 12% for Cassandra Latency(% of request Meets latency QoS): Quasar: 98.8% for memcached, 98.6% For Cassandra. Auto-scaling: 80% for memcached, 93% For Cassandra. First, there are multiple lowlatency services. Second, these services involve significant volumes of states(memcached, Cassandra) Quasar meets latency QoS for memcached for 98.8% of requests, while auto-scaling only for 80% of requests. For Cassandra, Quasar meets the latency QoS for 98.6% of requests, while the auto-scaling for 93% of requests.

Large-Scale Cloud Provider bring everything together in a general case 1200 workloads of all types 200-node cluster of dedicated EC2 servers

Large-Scale Cloud Provider Performance: 15% of the performance target on average more than Reservation- based system with Paragon Utilization: Reservation + least-loaded manager achieves average utilization of 15%, 47% lower than Quasar Low overheads: 4%-9%

Discussion Fault tolerance Cost target Resource partitioning We use master-slave mirroring to provide fault-tolerance for the server that runs the Quasar scheduler. All system state (list of active applications, allocations, QoS guarantees) is continuously replicated and can be used by hot-standby masters. Cost target: In the future, a cost constraint, priorities, and utility functions for a workload Resource partitioning: Quasar does not explicitly partition hardware resources. Instead, it reduces interference by co-locating workloads that do not contend on the shared resources.

Conclusions moves away from the reservation-based standard for cluster management. classification techniques to analyze the impact of resource allocation and type. a greedy algorithm uses this information to allocate the necessary of resources