BALANCED DATA LAYOUT IN HADOOP CPS 216 Kyungmin (Jason) Lee Ke (Jessie) Xu Weiping Zhang.

Slides:



Advertisements
Similar presentations
Starfish: A Self-tuning System for Big Data Analytics.
Advertisements

MapReduce Online Tyson Condie UC Berkeley Slides by Kaixiang MO
Big Data + SDN SDN Abstractions. The Story Thus Far Different types of traffic in clusters Background Traffic – Bulk transfers – Control messages Active.
SDN + Storage.
Effective Straggler Mitigation: Attack of the Clones [1]
LIBRA: Lightweight Data Skew Mitigation in MapReduce
Based on the text by Jimmy Lin and Chris Dryer; and on the yahoo tutorial on mapreduce at index.html
MapReduce Online Created by: Rajesh Gadipuuri Modified by: Ying Lu.
MapReduce Online Veli Hasanov Fatih University.
EHarmony in Cloud Subtitle Brian Ko. eHarmony Online subscription-based matchmaking service Available in United States, Canada, Australia and United Kingdom.
Locality-Aware Dynamic VM Reconfiguration on MapReduce Clouds Jongse Park, Daewoo Lee, Bokyeong Kim, Jaehyuk Huh, Seungryoul Maeng.
SkewTune: Mitigating Skew in MapReduce Applications
Join Processing in Databases Systems with Large Main Memories
Cloud Computing Resource provisioning Keke Chen. Outline  For Web applications statistical Learning and automatic control for datacenters  For data.
CS 795 – Spring  “Software Systems are increasingly Situated in dynamic, mission critical settings ◦ Operational profile is dynamic, and depends.
Center-of-Gravity Reduce Task Scheduling to Lower MapReduce Network Traffic Mohammad Hammoud, M. Suhail Rehman, and Majd F. Sakr 1.
Towards Energy Efficient Hadoop Wednesday, June 10, 2009 Santa Clara Marriott Yanpei Chen, Laura Keys, Randy Katz RAD Lab, UC Berkeley.
Hadoop: The Definitive Guide Chap. 2 MapReduce
 Need for a new processing platform (BigData)  Origin of Hadoop  What is Hadoop & what it is not ?  Hadoop architecture  Hadoop components (Common/HDFS/MapReduce)
Towards Energy Efficient MapReduce Yanpei Chen, Laura Keys, Randy H. Katz University of California, Berkeley LoCal Retreat June 2009.
OS Spring ’ 04 Scheduling Operating Systems Spring 2004.
A Row-Permutated Data Reorganization Algorithm for Growing Server-less VoD Systems Presented by Ho Tsz Kin.
UC Berkeley Improving MapReduce Performance in Heterogeneous Environments Matei Zaharia, Andy Konwinski, Anthony Joseph, Randy Katz, Ion Stoica University.
CPS216: Advanced Database Systems (Data-intensive Computing Systems) How MapReduce Works (in Hadoop) Shivnath Babu.
Google Distributed System and Hadoop Lakshmi Thyagarajan.
Copyright © 2012 Cleversafe, Inc. All rights reserved. 1 Combining the Power of Hadoop with Object-Based Dispersed Storage.
Take An Internal Look at Hadoop Hairong Kuang Grid Team, Yahoo! Inc
Hadoop: The Definitive Guide Chap. 8 MapReduce Features
On Availability of Intermediate Data in Cloud Computations Steven Y. Ko, Imranul Hoque, Brian Cho, and Indranil Gupta Distributed Protocols Research Group.
Advanced Topics: MapReduce ECE 454 Computer Systems Programming Topics: Reductions Implemented in Distributed Frameworks Distributed Key-Value Stores Hadoop.
Rensselaer Polytechnic Institute CSC 432 – Operating Systems David Goldschmidt, Ph.D.
Map Reduce for data-intensive computing (Some of the content is adapted from the original authors’ talk at OSDI 04)
Network Aware Resource Allocation in Distributed Clouds.
CS525: Special Topics in DBs Large-Scale Data Management Hadoop/MapReduce Computing Paradigm Spring 2013 WPI, Mohamed Eltabakh 1.
MapReduce: Simplified Data Processing on Large Clusters Jeffrey Dean and Sanjay Ghemawat.
Tiziana FerrariNetwork metrics usage for optimization of the Grid1 DataGrid Project Work Package 7 Written by Tiziana Ferrari Presented by Richard Hughes-Jones.
Hadoop/MapReduce Computing Paradigm 1 Shirish Agale.
Introduction to Hadoop and HDFS
Hadoop Hardware Infrastructure considerations ©2013 OpalSoft Big Data.
MARISSA: MApReduce Implementation for Streaming Science Applications 作者 : Fadika, Z. ; Hartog, J. ; Govindaraju, M. ; Ramakrishnan, L. ; Gunter, D. ; Canon,
Papers on Storage Systems 1) Purlieus: Locality-aware Resource Allocation for MapReduce in a Cloud, SC ) Making Cloud Intermediate Data Fault-Tolerant,
GreenSched: An Energy-Aware Hadoop Workflow Scheduler
CARDIO: Cost-Aware Replication for Data-Intensive workflOws Presented by Chen He.
MC 2 : Map Concurrency Characterization for MapReduce on the Cloud Mohammad Hammoud and Majd Sakr 1.
1 Making MapReduce Scheduling Effective in Erasure-Coded Storage Clusters Runhui Li and Patrick P. C. Lee The Chinese University of Hong Kong LANMAN’15.
Tutorial: Big Data Algorithms and Applications Under Hadoop KUNPENG ZHANG SIDDHARTHA BHATTACHARYYA
1 Enabling Efficient and Reliable Transitions from Replication to Erasure Coding for Clustered File Systems Runhui Li, Yuchong Hu, Patrick P. C. Lee The.
Record Linkage in a Distributed Environment
Matchmaking: A New MapReduce Scheduling Technique
Rassul Ayani 1 Performance of parallel and distributed systems  What is the purpose of measurement?  To evaluate a system (or an architecture)  To compare.
MROrder: Flexible Job Ordering Optimization for Online MapReduce Workloads School of Computer Engineering Nanyang Technological University 30 th Aug 2013.
Dynamic Slot Allocation Technique for MapReduce Clusters School of Computer Engineering Nanyang Technological University 25 th Sept 2013 Shanjiang Tang,
CS525: Big Data Analytics MapReduce Computing Paradigm & Apache Hadoop Open Source Fall 2013 Elke A. Rundensteiner 1.
DynamicMR: A Dynamic Slot Allocation Optimization Framework for MapReduce Clusters Nanyang Technological University Shanjiang Tang, Bu-Sung Lee, Bingsheng.
IBM Research ® © 2007 IBM Corporation Introduction to Map-Reduce and Join Processing.
Copyright © 2012 Cleversafe, Inc. All rights reserved. 1 Combining the Power of Hadoop with Object-Based Dispersed Storage.
Hadoop/MapReduce Computing Paradigm 1 CS525: Special Topics in DBs Large-Scale Data Management Presented By Kelly Technologies
Using Map-reduce to Support MPMD Peng
Learn. Hadoop Online training course is designed to enhance your knowledge and skills to become a successful Hadoop developer and In-depth knowledge of.
1 VLDB, Background What is important for the user.
PACMan: Coordinated Memory Caching for Parallel Jobs Ganesh Ananthanarayanan, Ali Ghodsi, Andrew Wang, Dhruba Borthakur, Srikanth Kandula, Scott Shenker,
1 Performance Impact of Resource Provisioning on Workflows Gurmeet Singh, Carl Kesselman and Ewa Deelman Information Science Institute University of Southern.
Presenter: Yue Zhu, Linghan Zhang A Novel Approach to Improving the Efficiency of Storing and Accessing Small Files on Hadoop: a Case Study by PowerPoint.
Hadoop MapReduce Framework
PA an Coordinated Memory Caching for Parallel Jobs
MapReduce Computing Paradigm Basics Fall 2013 Elke A. Rundensteiner
The Basics of Apache Hadoop
On Spatial Joins in MapReduce
Discretized Streams: A Fault-Tolerant Model for Scalable Stream Processing Zaharia, et al (2012)
Presentation transcript:

BALANCED DATA LAYOUT IN HADOOP CPS 216 Kyungmin (Jason) Lee Ke (Jessie) Xu Weiping Zhang

Background How data is stored on HDFS affects Hadoop MapReduce performance Mapper phase: decreased performance if need to fetch input data from remote node across network Imbalance during a MapReduce workflow (output from one job used as input to next) makes problem even worse Project goal: minimize the need to fetch Map input across network by balancing input data across nodes

Previous Work Reactive Solution – HDFS Rebalancer Algorithm to rebalance data layout in HDFS based on storage utilization Reacts to already-existing data layout imbalance, would like way to prevent altogether Proactive Solution – RR Block Placement Policy On HDFS writes, choose target node in round robin fashion, so data guaranteed balance Unnecessary writes across network? Can we do better?

Balanced Block Placement Policy Do writes ‘greedily’ as long as cluster is ‘fairly balanced’ ‘Greedily’ = prioritize target nodes based on location Local node > node on rack > remote node ‘Fairly balanced’ = size of all nodes fall within a specified ranged (windowSize) Algorithm: Sort live nodes on HDFS used; threshold = max – windowSize 1 st replica: write to local node if it is below threshold or if all nodes are above threshold, otherwise write to least utilized node 2 nd replica: least utilized node that is on different rack (if possible) than 1 st replica 3 rd replica and beyond: least utilized remaining node

Test Workloads 4-node cluster Default Policy (DP) vs. Balanced Policy (BP) 2 MapReduce Jobs Balanced Sort (each reducer approx. same output size) Skewed Sort (skewed reducer output sizes) 2 Workloads Single run, vary number reducers (1, 2, 4, 10, 12) Cascaded workflow, 3 sorts in series, reducers = 10 Other parameters RF (replication factor) – 1 and 3 Speculative Execution on and off (SE vs NSE) Monitor amount of data written to node by standard deviation  higher StdDev implies more imbalance

Quick Demo on Amazon EC2

Balanced Sort Single Run DP very skewed for reducers < 4, as expected Otherwise both pretty balanced (as expected)

Skewed Sort Single Run DP significantly worse than BP RF3 show better balance than RF1 Disabling SE improves balance in BP

Balanced Sort Workflow

Skewed Sort Workflow

Performance No significant overhead/improvements observed

Speculative Execution Hadoop performance feature that runs same task on 2 nodes concurrently, uses data from task that completes first and discards the other Usually occurs toward end of a job, leading to unintended data imbalance in balanced policy Turning off speculative execution improved data balance, but in practice would like to keep this feature on for performance boost Our policy too greedy, less affected if a node writes approximately equally to all nodes  round robin Hybrid policy, some nodes run round robin and some nodes run balanced policy? Tradeoff between balance and network traffic?

Future Considerations Current implementation assumes data will be balanced throughout cluster’s lifetime What if some nodes are down for a period of time and data becomes imbalanced? Data output per job should be spread evenly, vs. overall data layout spread evenly Need additional knowledge of which job each write belongs to Effect of window size on balance/performance? Unable to test due to insufficient funds

Conclusion Implemented new block placement policy that focuses on maintaining data balance while keeping writes local as much as possible Test data showed success at maintaining data balance Greatest improvements with skewed outputs Performance not affected – would expect improvement for skewed datasets given reduction in network usage Only tested on small cluster with small datasets Should be more effective on large datasets Performance weakened by speculative execution In practice should tweak our policy to get best performance results