DynamicMR: A Dynamic Slot Allocation Optimization Framework for MapReduce Clusters Nanyang Technological University Shanjiang Tang, Bu-Sung Lee, Bingsheng.

Slides:



Advertisements
Similar presentations
Lecture 12: MapReduce: Simplified Data Processing on Large Clusters Xiaowei Yang (Duke University)
Advertisements

MapReduce Online Created by: Rajesh Gadipuuri Modified by: Ying Lu.
Locality-Aware Dynamic VM Reconfiguration on MapReduce Clouds Jongse Park, Daewoo Lee, Bokyeong Kim, Jaehyuk Huh, Seungryoul Maeng.
UC Berkeley Improving MapReduce Performance in Heterogeneous Environments Matei Zaharia, Andy Konwinski, Anthony Joseph, Randy Katz, Ion Stoica University.
UC Berkeley Improving MapReduce Performance in Heterogeneous Environments Matei Zaharia, Andy Konwinski, Anthony Joseph, Randy Katz, Ion Stoica University.
CPS216: Advanced Database Systems (Data-intensive Computing Systems) How MapReduce Works (in Hadoop) Shivnath Babu.
Take An Internal Look at Hadoop Hairong Kuang Grid Team, Yahoo! Inc
Design Patterns for Efficient Graph Algorithms in MapReduce Jimmy Lin and Michael Schatz University of Maryland MLG, January, 2014 Jaehwan Lee.
U.S. Department of the Interior U.S. Geological Survey David V. Hill, Information Dynamics, Contractor to USGS/EROS 12/08/2011 Satellite Image Processing.
Edge Based Cloud Computing as a Feasible Network Paradigm(1/27) Edge-Based Cloud Computing as a Feasible Network Paradigm Joe Elizondo and Sam Palmer.
Hadoop Ida Mele. Parallel programming Parallel programming is used to improve performance and efficiency In a parallel program, the processing is broken.
MapReduce.
Introduction to Parallel Programming MapReduce Except where otherwise noted all portions of this work are Copyright (c) 2007 Google and are licensed under.
A Dynamic MapReduce Scheduler for Heterogeneous Workloads Chao Tian, Haojie Zhou, Yongqiang He,Li Zha 簡報人:碩資工一甲 董耀文.
CS525: Special Topics in DBs Large-Scale Data Management Hadoop/MapReduce Computing Paradigm Spring 2013 WPI, Mohamed Eltabakh 1.
1 The Map-Reduce Framework Compiled by Mark Silberstein, using slides from Dan Weld’s class at U. Washington, Yaniv Carmeli and some other.
Presented By HaeJoon Lee Yanyan Shen, Beng Chin Ooi, Bogdan Marius Tudor National University of Singapore Wei Lu Renmin University Cang Chen Zhejiang University.
Storage in Big Data Systems
1 Fast Failure Recovery in Distributed Graph Processing Systems Yanyan Shen, Gang Chen, H.V. Jagadish, Wei Lu, Beng Chin Ooi, Bogdan Marius Tudor.
MapReduce: Hadoop Implementation. Outline MapReduce overview Applications of MapReduce Hadoop overview.
Hadoop/MapReduce Computing Paradigm 1 Shirish Agale.
Introduction to Hadoop and HDFS
Optimizing Cloud MapReduce for Processing Stream Data using Pipelining 作者 :Rutvik Karve , Devendra Dahiphale , Amit Chhajer 報告 : 饒展榕.
MapReduce How to painlessly process terabytes of data.
Eneryg Efficiency for MapReduce Workloads: An Indepth Study Boliang Feng Renmin University of China Dec 19.
Dominant Resource Fairness: Fair Allocation of Multiple Resource Types Ali Ghodsi, Matei Zaharia, Benjamin Hindman, Andy Konwinski, Scott Shenker, Ion.
MapReduce Kristof Bamps Wouter Deroey. Outline Problem overview MapReduce o overview o implementation o refinements o conclusion.
MC 2 : Map Concurrency Characterization for MapReduce on the Cloud Mohammad Hammoud and Majd Sakr 1.
Computer Science and Engineering Parallelizing Defect Detection and Categorization Using FREERIDE Leonid Glimcher P. 1 ipdps’05 Scaling and Parallelizing.
1 Coscheduling in Clusters: Is it a Viable Alternative? Gyu Sang Choi, Jin-Ha Kim, Deniz Ersoz, Andy B. Yoo, Chita R. Das Presented by: Richard Huang.
Using Map-reduce to Support MPMD Peng
MROrder: Flexible Job Ordering Optimization for Online MapReduce Workloads School of Computer Engineering Nanyang Technological University 30 th Aug 2013.
Department of Computer Science MapReduce for the Cell B. E. Architecture Marc de Kruijf University of Wisconsin−Madison Advised by Professor Sankaralingam.
SECTION 5: PERFORMANCE CHRIS ZINGRAF. OVERVIEW: This section measures the performance of MapReduce on two computations, Grep and Sort. These programs.
By Jeff Dean & Sanjay Ghemawat Google Inc. OSDI 2004 Presented by : Mohit Deopujari.
Shanjiang Tang, Bu-Sung Lee, Bingsheng He, Haikun Liu School of Computer Engineering Nanyang Technological University Long-Term Resource Fairness Towards.
Dynamic Slot Allocation Technique for MapReduce Clusters School of Computer Engineering Nanyang Technological University 25 th Sept 2013 Shanjiang Tang,
CS525: Big Data Analytics MapReduce Computing Paradigm & Apache Hadoop Open Source Fall 2013 Elke A. Rundensteiner 1.
C-Store: MapReduce Jianlin Feng School of Software SUN YAT-SEN UNIVERSITY May. 22, 2009.
A N I N - MEMORY F RAMEWORK FOR E XTENDED M AP R EDUCE 2011 Third IEEE International Conference on Coud Computing Technology and Science.
Hadoop/MapReduce Computing Paradigm 1 CS525: Special Topics in DBs Large-Scale Data Management Presented By Kelly Technologies
MapReduce: Simplified Data Processing on Large Clusters By Dinesh Dharme.
GPFS: A Shared-Disk File System for Large Computing Clusters Frank Schmuck & Roger Haskin IBM Almaden Research Center.
Load Rebalancing for Distributed File Systems in Clouds.
Cluster computing. 1.What is cluster computing? 2.Need of cluster computing. 3.Architecture 4.Applications of cluster computing 5.Advantages of cluster.
Next Generation of Apache Hadoop MapReduce Owen
Part III BigData Analysis Tools (YARN) Yuan Xue
INTRODUCTION TO HADOOP. OUTLINE  What is Hadoop  The core of Hadoop  Structure of Hadoop Distributed File System  Structure of MapReduce Framework.
MapReduce: Simplied Data Processing on Large Clusters Written By: Jeffrey Dean and Sanjay Ghemawat Presented By: Manoher Shatha & Naveen Kumar Ratkal.
1 Student Date Time Wei Li Nov 30, 2015 Monday 9:00-9:25am Shubbhi Taneja Nov 30, 2015 Monday9:25-9:50am Rodrigo Sanandan Dec 2, 2015 Wednesday9:00-9:25am.
PACMan: Coordinated Memory Caching for Parallel Jobs Ganesh Ananthanarayanan, Ali Ghodsi, Andrew Wang, Dhruba Borthakur, Srikanth Kandula, Scott Shenker,
Spark on Entropy : A Reliable & Efficient Scheduler for Low-latency Parallel Jobs in Heterogeneous Cloud Huankai Chen PhD Student at University of Kent.
Leverage Big Data With Hadoop Analytics Presentation by Ravi Namboori Visit
Lecture 3 – MapReduce: Implementation CSE 490h – Introduction to Distributed Computing, Spring 2009 Except as otherwise noted, the content of this presentation.
Hadoop Aakash Kag What Why How 1.
Distributed Network Traffic Feature Extraction for a Real-time IDS
Edinburgh Napier University
CS 425 / ECE 428 Distributed Systems Fall 2016 Nov 10, 2016
CS 425 / ECE 428 Distributed Systems Fall 2017 Nov 16, 2017
Introduction to MapReduce and Hadoop
PA an Coordinated Memory Caching for Parallel Jobs
Software Engineering Introduction to Apache Hadoop Map Reduce
Ministry of Higher Education
MapReduce Simplied Data Processing on Large Clusters
The Basics of Apache Hadoop
湖南大学-信息科学与工程学院-计算机与科学系
CS110: Discussion about Spark
Shanjiang Tang1, Bingsheng He2, Shuhao Zhang2,4, Zhaojie Niu3
Adaptive Data Refinement for Parallel Dynamic Programming Applications
5/7/2019 Map Reduce Map reduce.
Presentation transcript:

DynamicMR: A Dynamic Slot Allocation Optimization Framework for MapReduce Clusters Nanyang Technological University Shanjiang Tang, Bu-Sung Lee, Bingsheng He School of Computer Engineering Nanyang Technological University 22/12/2015

OutLine Background and Motivation DynamicMR Overview Experimental Evaluation Conclusion 2Nanyang Technological University22/12/2015

Big Data is Everywhere Lots of data is being collected and warehoused. – Web data, e-commerce – purchases at department/ grocery stores – Bank/Credit Card transactions – Social Network – Astronomical Image Processing – Bioinformatics. 3Nanyang Technological University22/12/2015

MapReduce is a Promising Choice A popular parallel programming model 4Nanyang Technological University Map Intermediate Result Map Reduce Output Result Reduce Output Result Reduce Output Result Reduce Output Result Final Result Map-Phase Computation Reduce-Phase Computation Input Data 22/12/2015

Hadoop Apache Hadoop is a open-source framework for reliable, scalable, and distributed computing. It implements the computational paradigm named MapReduce. – Scale up to 6,000-10,000 machines – Support for multi-tenancy Useful links: – – – 5Nanyang Technological University22/12/2015

Challenges in Distributed Environment Node failures and Stragglers (slow nodes) – Mean time between failures for 1000 nodes = 1 day  Affecting performance. Commodity network = low bandwidth – Push computation to the data (Data Locality Optimization)  Affecting performance. Resource contention in shared cluster environment – Performance isolation and fair resource sharing  Affecting performance and fairness. Performance and fairness optimization are important! 22/12/2015Nanyang Technological University6

Our Work Challenges: How to improve the performance of Hadoop while guarantee the fairness? Our Solution: DynamicMR: A Dynamic Resource Allocation System for Hadoop. – Improve the resource utilization as much as possible. – Improve the utilization efficiency as much as possible. 22/12/2015Nanyang Technological University7

OutLine Background and Motivation DynamicMR Overview Experimental Evaluation Conclusion 8Nanyang Technological University22/12/2015

Hadoop abstracts resources into map slots and reduce slots. – Configured by Hadoop administrator statically. – Resource constrain: map tasks can only use map slots, reduce tasks can only use reduce slots. Observation 1#: Poor Resource Utilization 9Nanyang Technological University22/12/ Map Phase Reduce Phase 8 map slots 4 reduce slots Slots resources are wasted during computation!

Core idea of DHSA. – Slots are generic and can be used by either map or reduce tasks, although there is a pre-configuration for the number of map and reduce slots. – Map tasks will prefer to use map slots and likewise reduce tasks prefer to use reduce slots. Technique 1#: Dynamic Hadoop Slot Allocation (DHSA) 10Nanyang Technological University22/12/ Map Phase Reduce Phase 8 map slots 4 reduce slots

Observation 2#: Speculative Execution is a Double-edged Sword Speculative Scheduling – Run a backup task for straggled task. – Pros: Can improve the performance of a single Job. – Cons: the resource utilization efficiency is reduced, especially when there are other pending tasks. 11Nanyang Technological University22/12/ straggler Backup task A Performance tradeoff for a single job and batch jobs! Benefit J1 Benefit the whole workload

Key idea of SEPB: – Instead of running speculative tasks immediately when straggler of a job is detected, we check a subset of jobs ( maxNumOfJobsCheckedForPendingTasks )for pending tasks. – If there are pending tasks, allocate pending tasks. Otherwise, allocate speculative task. Technique 2#: Speculative Execution Performance Balancing (SEPB) 12Nanyang Technological University22/12/2015 J4J3J2J1J5J6 maxNumOfJobsCheckedForPendingTasks

Observation 3#: Load Balance Requirement Harms Data Locality Load Balancing is adopted by Hadoop. – Hadoop tries to keep the load (i.e., running tasks) in each node is as close as possible. 13Nanyang Technological University22/12/2015 Load Balancing makes J 1 failed to achieve data locality!

Key idea: Improve data locality at the expense of load balance. – When there are idle slots and local data, we preschedule the task on that machine first. – Otherwise, we keep the load balance constrain. Technique 3#: Slot PreScheduling 14Nanyang Technological University22/12/2015

DynamicMR A combination of the aforementioned three techniques. – DHSA : Slot Utilization Optimization. – SEPB, Slot PreScheduling: Efficiency Optimization 15Nanyang Technological University22/12/2015 Speculative Execution Performance Balancing (SEPB) Slot PreScheduling Dynamic Hadoop Slot Allocation (DHSA) Map Task Reduce Task (1). Slot Utilization Optimization (2). Utilization Efficiency Optimization Idle Slot 1 2 3

OutLine Background and Motivation DynamicMR Overview Experimental Evaluation Conclusion 16Nanyang Technological University22/12/2015

Experimental Setup Hadoop Cluster – 10 nodes, each with two Intel X5675 CPUs (6 cores per CPU with 3.07 GHz), 24GB DDR3 memory, 56GB hard disks. Benchmark and Data Sets. 17Nanyang Technological University22/12/2015

DynamicMR Performance Evaluation 18Nanyang Technological University22/12/2015

DynamicMR VS YARN DynamicMR achieves better performance than YARN. – Benefits from the ratio control of concurrently running map and reduce tasks of DynamicMR, whereas YARN not. 19Nanyang Technological University22/12/2015

OutLine Background and Motivation DynamicMR Overview Experimental Evaluation Conclusion 20Nanyang Technological University22/12/2015

Conclusion We propose a DynamicMR framework to improve the performance of MapReduce workloads while maintaining the fairness. – Consists of three techniques: DHSA, SEPB, and Slot Prescheduling. Experimental results show that: – It improves the performance of Hadoop 46%~115% for single jobs and 49%~112% for batch jobs. – It outperforms YARN by about 2%~9% for multiple jobs. 21Nanyang Technological University22/12/2015

22Nanyang Technological University22/12/2015

DHSA Evaluation DHSA achieves a better performance than Hadoop. Hadoop is sensitive to slot configuration, whereas DHSA does not. 23Nanyang Technological University22/12/2015

SEPB Evaluation SEPB improves the performance for the whole jobs (Figure a). There is a performance tradeoff between an individual jobs and the whole jobs with SEPB (Figure b). 24Nanyang Technological University22/12/2015

Slot PreScheduling Evaluation Data Locality and Performance Improvement 25Nanyang Technological University22/12/2015