Piccolo – Paper Discussion Big Data Reading Group 9/20/2010.

Slides:



Advertisements
Similar presentations
Lecture 12: MapReduce: Simplified Data Processing on Large Clusters Xiaowei Yang (Duke University)
Advertisements

MapReduce Online Tyson Condie UC Berkeley Slides by Kaixiang MO
MAP REDUCE PROGRAMMING Dr G Sudha Sadasivam. Map - reduce sort/merge based distributed processing Best for batch- oriented processing Sort/merge is primitive.
Piccolo: Building fast distributed programs with partitioned tables Russell Power Jinyang Li New York University.
epiC: an Extensible and Scalable System for Processing Big Data
MapReduce Online Created by: Rajesh Gadipuuri Modified by: Ying Lu.
Cloud Computing Resource provisioning Keke Chen. Outline  For Web applications statistical Learning and automatic control for datacenters  For data.
The SMART Way to Migrate Replicated Stateful Services Jacob R. Lorch, Atul Adya, Bill Bolosky, Ronnie Chaiken, John Douceur, Jon Howell Microsoft Research.
Serverless Network File Systems. Network File Systems Allow sharing among independent file systems in a transparent manner Mounting a remote directory.
Distributed Graph Processing Abhishek Verma CS425.
Spark: Cluster Computing with Working Sets
Scaling Distributed Machine Learning with the BASED ON THE PAPER AND PRESENTATION: SCALING DISTRIBUTED MACHINE LEARNING WITH THE PARAMETER SERVER – GOOGLE,
Distributed Computations
Graph Processing Recap: data-intensive cloud computing – Just database management on the cloud – But scaling it to thousands of nodes – Handling partial.
Distributed Computations MapReduce
CPS216: Advanced Database Systems (Data-intensive Computing Systems) How MapReduce Works (in Hadoop) Shivnath Babu.
L22: SC Report, Map Reduce November 23, Map Reduce What is MapReduce? Example computing environment How it works Fault Tolerance Debugging Performance.
Lecture 2 – MapReduce CPE 458 – Parallel Programming, Spring 2009 Except as otherwise noted, the content of this presentation is licensed under the Creative.
Next Generation of Apache Hadoop MapReduce Arun C. Murthy - Hortonworks Founder and Architect Formerly Architect, MapReduce.
Take An Internal Look at Hadoop Hairong Kuang Grid Team, Yahoo! Inc
Advanced Topics: MapReduce ECE 454 Computer Systems Programming Topics: Reductions Implemented in Distributed Frameworks Distributed Key-Value Stores Hadoop.
MapReduce.
By: Jeffrey Dean & Sanjay Ghemawat Presented by: Warunika Ranaweera Supervised by: Dr. Nalin Ranasinghe.
Research on cloud computing application in the peer-to-peer based video-on-demand systems Speaker : 吳靖緯 MA0G rd International Workshop.
Map Reduce for data-intensive computing (Some of the content is adapted from the original authors’ talk at OSDI 04)
CS525: Special Topics in DBs Large-Scale Data Management Hadoop/MapReduce Computing Paradigm Spring 2013 WPI, Mohamed Eltabakh 1.
MapReduce: Simplified Data Processing on Large Clusters Jeffrey Dean and Sanjay Ghemawat.
MapReduce – An overview Medha Atre (May 7, 2008) Dept of Computer Science Rensselaer Polytechnic Institute.
Presented By HaeJoon Lee Yanyan Shen, Beng Chin Ooi, Bogdan Marius Tudor National University of Singapore Wei Lu Renmin University Cang Chen Zhejiang University.
1 Fast Failure Recovery in Distributed Graph Processing Systems Yanyan Shen, Gang Chen, H.V. Jagadish, Wei Lu, Beng Chin Ooi, Bogdan Marius Tudor.
MapReduce: Hadoop Implementation. Outline MapReduce overview Applications of MapReduce Hadoop overview.
Hadoop/MapReduce Computing Paradigm 1 Shirish Agale.
Distributed shared memory. What we’ve learnt so far  MapReduce/Dryad as a distributed programming model  Data-flow (computation as vertex, data flow.
Map Reduce: Simplified Processing on Large Clusters Jeffrey Dean and Sanjay Ghemawat Google, Inc. OSDI ’04: 6 th Symposium on Operating Systems Design.
Mesos A Platform for Fine-Grained Resource Sharing in the Data Center Benjamin Hindman, Andy Konwinski, Matei Zaharia, Ali Ghodsi, Anthony Joseph, Randy.
MapReduce How to painlessly process terabytes of data.
MapReduce M/R slides adapted from those of Jeff Dean’s.
MapReduce Kristof Bamps Wouter Deroey. Outline Problem overview MapReduce o overview o implementation o refinements o conclusion.
MapReduce and GFS. Introduction r To understand Google’s file system let us look at the sort of processing that needs to be done r We will look at MapReduce.
Fault Tolerant Extensions to Charm++ and AMPI presented by Sayantan Chakravorty Chao Huang, Celso Mendes, Gengbin Zheng, Lixia Shi.
MapReduce Algorithm Design Based on Jimmy Lin’s slides
CS525: Big Data Analytics MapReduce Computing Paradigm & Apache Hadoop Open Source Fall 2013 Elke A. Rundensteiner 1.
Big traffic data processing framework for intelligent monitoring and recording systems 學生 : 賴弘偉 教授 : 許毅然 作者 : Yingjie Xia a, JinlongChen a,b,n, XindaiLu.
C-Store: MapReduce Jianlin Feng School of Software SUN YAT-SEN UNIVERSITY May. 22, 2009.
A N I N - MEMORY F RAMEWORK FOR E XTENDED M AP R EDUCE 2011 Third IEEE International Conference on Coud Computing Technology and Science.
Hadoop/MapReduce Computing Paradigm 1 CS525: Special Topics in DBs Large-Scale Data Management Presented By Kelly Technologies
MapReduce: Simplified Data Processing on Large Clusters By Dinesh Dharme.
GPFS: A Shared-Disk File System for Large Computing Clusters Frank Schmuck & Roger Haskin IBM Almaden Research Center.
FTC-Charm++: An In-Memory Checkpoint-Based Fault Tolerant Runtime for Charm++ and MPI Gengbin Zheng Lixia Shi Laxmikant V. Kale Parallel Programming Lab.
Next Generation of Apache Hadoop MapReduce Owen
Department of Computer Science, Johns Hopkins University Pregel: BSP and Message Passing for Graph Computations EN Randal Burns 14 November 2013.
MapReduce: Simplied Data Processing on Large Clusters Written By: Jeffrey Dean and Sanjay Ghemawat Presented By: Manoher Shatha & Naveen Kumar Ratkal.
COMP7330/7336 Advanced Parallel and Distributed Computing MapReduce - Introduction Dr. Xiao Qin Auburn University
EpiC: an Extensible and Scalable System for Processing Big Data Dawei Jiang, Gang Chen, Beng Chin Ooi, Kian Lee Tan, Sai Wu School of Computing, National.
Resilient Distributed Datasets A Fault-Tolerant Abstraction for In-Memory Cluster Computing Matei Zaharia, Mosharaf Chowdhury, Tathagata Das, Ankur Dave,
Lecture 3 – MapReduce: Implementation CSE 490h – Introduction to Distributed Computing, Spring 2009 Except as otherwise noted, the content of this presentation.
TensorFlow– A system for large-scale machine learning
Distributed Programming in “Big Data” Systems Pramod Bhatotia wp
Auburn University COMP7330/7336 Advanced Parallel and Distributed Computing MapReduce - Introduction Dr. Xiao Qin Auburn.
MapReduce Computing Paradigm Basics Fall 2013 Elke A. Rundensteiner
MapReduce Simplied Data Processing on Large Clusters
湖南大学-信息科学与工程学院-计算机与科学系
Cse 344 May 4th – Map/Reduce.
Ch 4. The Evolution of Analytic Scalability
Discretized Streams: A Fault-Tolerant Model for Scalable Stream Processing Zaharia, et al (2012)
MAPREDUCE TYPES, FORMATS AND FEATURES
Cloud Computing Large-scale Resource Management
The SMART Way to Migrate Replicated Stateful Services
5/7/2019 Map Reduce Map reduce.
MapReduce: Simplified Data Processing on Large Clusters
Presentation transcript:

Piccolo – Paper Discussion Big Data Reading Group 9/20/2010

Motivation / Goals Rising demand for distributing computation PageRank, K-Means, N-Body simulation Data-centric frameworks simplify programming Existing models (e.g. MapReduce) are insufficient Designed for large scale data analysis as opposed to in-memory computation Make in-memory computations fast Enable asynchronous computation 9/20/2010 Piccolo – Paper Discussion 2

Overview Global in-memory key-value tables for sharing state Concurrently running instances of kernel applications modifying global state Locality optimized (user specified policies) Reduced synchronization (accumulation, global barriers) Checkpoint-based recovery 9/20/2010 Piccolo – Paper Discussion 3

System Design 9/20/2010 Piccolo – Paper Discussion 4

Table interface 9/20/2010 Piccolo – Paper Discussion 5

Optimization Ensure Locality Group kernel with partition Group partitions Guarantee: one partition completely on single machine Reduce Synchronization Accumulation to avoid write/write conflicts No pairwise kernel synchronization Global barriers sufficient 9/20/2010 Piccolo – Paper Discussion 6

Load balancing Assigning partitions Round robin Optimized for data location Work stealing Biggest task first (master estimates based on number of keys in partition) Master decides Restrictions Cannot kill running task (modifies shared state, restore is very expensive) Partitions need to be moved 9/20/2010 Piccolo – Paper Discussion 7

Table migration Migrate table from w a to w b Message M 1 from master to all workers All workers flush to w a All workers send all new requests to w b w b buffers all requests w a sends paused state to w b All workers ackknowledge phase 1 => master sends M 2 to w a and w b w a flushes to w b and leaves “paused” w b first works buffered requests then resumes normal operation 9/20/2010 Piccolo – Paper Discussion 8

Fault tolerance User assisted checkpoint / restore Chandy Lamport Asynchronic -> periodic Synchronic -> barrier Problem: When to start barrier checkpoint Replay log might get very long Checkpoint might not use enough free CPU time before barrier Solution: When first worker finished all his jobs No checkpoint during table migration and vice versa 9/20/2010 Piccolo – Paper Discussion 9

Applications PageRank, k-means, n-body, matrix multiplication Parallel, iterative computations Local reads + local/remote writes or local/remote reads + local writes Can be implemented as multiple MapReduce jobs Distributed web crawler Idempotent operation Cannot be realized in MapReduce 9/20/2010 Piccolo – Paper Discussion 10

Scaling 9/20/2010 Piccolo – Paper Discussion 11 Fixed input size Scaled input size

Comparison with Hadoop / MPI 9/20/2010 Piccolo – Paper Discussion 12 PageRank, k-means (Hadoop) Piccolo 4x and 11x faster For PageRank: 50% in sort Join data streams 15% (de)serialization Read/write HDFS Matrix multiplication (MPI) Piccolo 10% faster MPI waits for slowest node many times

Work stealing / slow worker / checkpoints 9/20/2010 Piccolo – Paper Discussion 13 Work stealing / slow worker PageRank has skewed partitions One slow worker (50% CPU) Checkpoints Naïve - start after all workers finished Optimized – start after first worker finished

Checkpoint limits / scalability 9/20/2010 Piccolo – Paper Discussion 14 Hypothetical data center Typical machine uptime of 1 year Worst-case scenario Optimistic? Looked different on some older slides

Distributed Crawler 9/20/2010 Piccolo – Paper Discussion Machines saturate 100Mbps There are single servers doing this Piccolo would scale higher

Summary Piccolo provides an easy to use distributed shared memory model It applies many restrictions Simple interface Reduced synchronization Relaxed consistency Accumulation Locality But it performs well Iterative computations Saves going to disk compared to MapReduce A specialized tool for data intensive in-memory computing 9/20/2010 Piccolo – Paper Discussion 16