A Fault-Tolerant Environment for Large-Scale Query Processing Mehmet Can Kurt Gagan Agrawal Department of Computer Science and Engineering The Ohio State.

Slides:



Advertisements
Similar presentations
LIBRA: Lightweight Data Skew Mitigation in MapReduce
Advertisements

MapReduce Online Created by: Rajesh Gadipuuri Modified by: Ying Lu.
HadoopDB Inneke Ponet.  Introduction  Technologies for data analysis  HadoopDB  Desired properties  Layers of HadoopDB  HadoopDB Components.
University of Minnesota CG_Hadoop: Computational Geometry in MapReduce Ahmed Eldawy* Yuan Li* Mohamed F. Mokbel*$ Ravi Janardan* * Department of Computer.
Scaling Distributed Machine Learning with the BASED ON THE PAPER AND PRESENTATION: SCALING DISTRIBUTED MACHINE LEARNING WITH THE PARAMETER SERVER – GOOGLE,
Distributed Computations
The Google File System. Why? Google has lots of data –Cannot fit in traditional file system –Spans hundreds (thousands) of servers connected to (tens.
Distributed Computations MapReduce
Jeffrey D. Ullman Stanford University.  Mining of Massive Datasets, J. Leskovec, A. Rajaraman, J. D. Ullman.  Available for free download at i.stanford.edu/~ullman/mmds.html.
Fast Subsequence Matching in Time-Series Databases Christos Faloutsos M. Ranganathan Yannis Manolopoulos Department of Computer Science and ISR University.
Distributed Data Stores – Facebook Presented by Ben Gooding University of Arkansas – April 21, 2015.
Advanced Topics: MapReduce ECE 454 Computer Systems Programming Topics: Reductions Implemented in Distributed Frameworks Distributed Key-Value Stores Hadoop.
IPDPS, Supporting Fault Tolerance in a Data-Intensive Computing Middleware Tekin Bicer, Wei Jiang and Gagan Agrawal Department of Computer Science.
Introduction to Parallel Programming MapReduce Except where otherwise noted all portions of this work are Copyright (c) 2007 Google and are licensed under.
Google MapReduce Simplified Data Processing on Large Clusters Jeff Dean, Sanjay Ghemawat Google, Inc. Presented by Conroy Whitney 4 th year CS – Web Development.
Fault-Tolerant Programming Models and Computing Frameworks Candidacy Examination 12/11/2013 Mehmet Can Kurt.
1 SciCSM: Novel Contrast Set Mining over Scientific Datasets Using Bitmap Indices Gangyi Zhu, Yi Wang, Gagan Agrawal The Ohio State University.
Map Reduce for data-intensive computing (Some of the content is adapted from the original authors’ talk at OSDI 04)
Ohio State University Department of Computer Science and Engineering Automatic Data Virtualization - Supporting XML based abstractions on HDF5 Datasets.
CS525: Special Topics in DBs Large-Scale Data Management Hadoop/MapReduce Computing Paradigm Spring 2013 WPI, Mohamed Eltabakh 1.
Ex-MATE: Data-Intensive Computing with Large Reduction Objects and Its Application to Graph Mining Wei Jiang and Gagan Agrawal.
MapReduce: Simplified Data Processing on Large Clusters Jeffrey Dean and Sanjay Ghemawat.
PAGE: A Framework for Easy Parallelization of Genomic Applications 1 Mucahid Kutlu Gagan Agrawal Department of Computer Science and Engineering The Ohio.
Presented By HaeJoon Lee Yanyan Shen, Beng Chin Ooi, Bogdan Marius Tudor National University of Singapore Wei Lu Renmin University Cang Chen Zhejiang University.
1 Fast Failure Recovery in Distributed Graph Processing Systems Yanyan Shen, Gang Chen, H.V. Jagadish, Wei Lu, Beng Chin Ooi, Bogdan Marius Tudor.
Pregel: A System for Large-Scale Graph Processing Presented by Dylan Davis Authors: Grzegorz Malewicz, Matthew H. Austern, Aart J.C. Bik, James C. Dehnert,
Cluster-based SNP Calling on Large Scale Genome Sequencing Data Mucahid KutluGagan Agrawal Department of Computer Science and Engineering The Ohio State.
Fault Tolerant Parallel Data-Intensive Algorithms Mucahid KutluGagan AgrawalOguz Kurt Department of Computer Science and Engineering The Ohio State University.
Performance Issues in Parallelizing Data-Intensive applications on a Multi-core Cluster Vignesh Ravi and Gagan Agrawal
Hadoop/MapReduce Computing Paradigm 1 Shirish Agale.
DLS on Star (Single-level tree) Networks Background: A simple network model for DLS is the star network with a master-worker platform. It consists of a.
1 A Framework for Data-Intensive Computing with Cloud Bursting Tekin Bicer David ChiuGagan Agrawal Department of Compute Science and Engineering The Ohio.
MapReduce M/R slides adapted from those of Jeff Dean’s.
Evaluating FERMI features for Data Mining Applications Masters Thesis Presentation Sinduja Muralidharan Advised by: Dr. Gagan Agrawal.
ICPP 2012 Indexing and Parallel Query Processing Support for Visualizing Climate Datasets Yu Su*, Gagan Agrawal*, Jonathan Woodring † *The Ohio State University.
Data in the Cloud – I Parallel Databases The Google File System Parallel File Systems.
EFFECTIVE LOAD-BALANCING VIA MIGRATION AND REPLICATION IN SPATIAL GRIDS ANIRBAN MONDAL KAZUO GODA MASARU KITSUREGAWA INSTITUTE OF INDUSTRIAL SCIENCE UNIVERSITY.
Performance Prediction for Random Write Reductions: A Case Study in Modelling Shared Memory Programs Ruoming Jin Gagan Agrawal Department of Computer and.
MapReduce Kristof Bamps Wouter Deroey. Outline Problem overview MapReduce o overview o implementation o refinements o conclusion.
Mehmet Can Kurt, The Ohio State University Gagan Agrawal, The Ohio State University DISC: A Domain-Interaction Based Programming Model With Support for.
HPDC 2013 Taming Massive Distributed Datasets: Data Sampling Using Bitmap Indices Yu Su*, Gagan Agrawal*, Jonathan Woodring # Kary Myers #, Joanne Wendelberger.
FREERIDE: System Support for High Performance Data Mining Ruoming Jin Leo Glimcher Xuan Zhang Ge Yang Gagan Agrawal Department of Computer and Information.
GFS. Google r Servers are a mix of commodity machines and machines specifically designed for Google m Not necessarily the fastest m Purchases are based.
Computing Scientometrics in Large-Scale Academic Search Engines with MapReduce Leonidas Akritidis Panayiotis Bozanis Department of Computer & Communication.
CCGrid, 2012 Supporting User Defined Subsetting and Aggregation over Parallel NetCDF Datasets Yu Su and Gagan Agrawal Department of Computer Science and.
Implementing Data Cube Construction Using a Cluster Middleware: Algorithms, Implementation Experience, and Performance Ge Yang Ruoming Jin Gagan Agrawal.
By Jeff Dean & Sanjay Ghemawat Google Inc. OSDI 2004 Presented by : Mohit Deopujari.
GEM: A Framework for Developing Shared- Memory Parallel GEnomic Applications on Memory Constrained Architectures Mucahid Kutlu Gagan Agrawal Department.
CS525: Big Data Analytics MapReduce Computing Paradigm & Apache Hadoop Open Source Fall 2013 Elke A. Rundensteiner 1.
MapReduce Computer Engineering Department Distributed Systems Course Assoc. Prof. Dr. Ahmet Sayar Kocaeli University - Fall 2015.
RE-PAGE: Domain-Specific REplication and PArallel Processing of GEnomic Applications 1 Mucahid Kutlu Gagan Agrawal Department of Computer Science and Engineering.
IBM Research ® © 2007 IBM Corporation Introduction to Map-Reduce and Join Processing.
R-Trees: A Dynamic Index Structure For Spatial Searching Antonin Guttman.
PDAC-10 Middleware Solutions for Data- Intensive (Scientific) Computing on Clouds Gagan Agrawal Ohio State University (Joint Work with Tekin Bicer, David.
CISC 849 : Applications in Fintech Namami Shukla Dept of Computer & Information Sciences University of Delaware iCARE : A Framework for Big Data Based.
Ohio State University Department of Computer Science and Engineering Servicing Range Queries on Multidimensional Datasets with Partial Replicas Li Weng,
Hadoop/MapReduce Computing Paradigm 1 CS525: Special Topics in DBs Large-Scale Data Management Presented By Kelly Technologies
MapReduce: Simplified Data Processing on Large Clusters By Dinesh Dharme.
Multi-dimensional Range Query Processing on the GPU Beomseok Nam Date Intensive Computing Lab School of Electrical and Computer Engineering Ulsan National.
1 Parallel Datacube Construction: Algorithms, Theoretical Analysis, and Experimental Evaluation Ruoming Jin Ge Yang Gagan Agrawal The Ohio State University.
COMP7330/7336 Advanced Parallel and Distributed Computing MapReduce - Introduction Dr. Xiao Qin Auburn University
Advanced Topics in Concurrency and Reactive Programming: Case Study – Google Cluster Majeed Kassis.
Large-scale file systems and Map-Reduce
Ge Yang Ruoming Jin Gagan Agrawal The Ohio State University
Sameh Shohdy, Yu Su, and Gagan Agrawal
Li Weng, Umit Catalyurek, Tahsin Kurc, Gagan Agrawal, Joel Saltz
On Spatial Joins in MapReduce
February 26th – Map/Reduce
Cse 344 May 4th – Map/Reduce.
Automatic and Efficient Data Virtualization System on Scientific Datasets Li Weng.
Presentation transcript:

A Fault-Tolerant Environment for Large-Scale Query Processing Mehmet Can Kurt Gagan Agrawal Department of Computer Science and Engineering The Ohio State University HiPC’12 Pune, India 1

Motivation “big data” problem – Walmart handles 1 million customer transaction every hour, estimated data volume is 2.5 Petabytes. – Facebook handles more than 40 billion images – LSST generates 6 petabytes every year massive parallelism is the key HiPC’12 Pune, India 2

Motivation Mean-Time To Failure (MTTF) decreases Typical first year for a new cluster* – 1000 individual machine failures – 1 PDU failure (~ machines suddenly disappear) – 20 rack failures (40-80 machines disappear, 1-6 hours to get back) HiPC’12 Pune, India 3 * taken from Jeff Dean’s talk in Google IO (

Our Work supporting fault-tolerant query processing and data analysis for a massive scientific dataset focusing on two specific query types: 1.Range Queries on Spatial datasets 2.Aggregation Queries on Point datasets supported failure types: single-machine failures and rack failures HiPC’12 Pune, India 4 * rack: a number of machines connected to the same hardware (network switch, …)

Our Work Primary Goals 1)high efficiency of execution when there are no failures (indexing if applicable, ensuring load-balance) 2)handling failures efficiently up to a certain number of nodes (low-overhead fault tolerance through data replication) 3)a modest slowdown in processing times when recovered from a failure (preserving load-balance) HiPC’12 Pune, India 5

Range Queries on Spatial Data nature of the task: – each data object is a rectangle in 2D space – each query is defined with a rectangle – return intersecting data rectangles computational model: – master/worker model – master serves as coordinator – each worker responsible for a portion of data HiPC’12 Pune, India 6 Y X query data worker query master

Range Queries on Spatial Data data organization: – chunk is the smallest data unit – create chunks by grouping data objects together – assign chunks to workers in round-robin fashion HiPC’12 Pune, India 7 Y X chunk 1 chunk 2 chunk 3 worker chunk 4 worker * actual number of chunks depends on chunk size parameter.

Range Queries on Spatial Data ensuring load-balance: – enumerate & sort data objects according to Hilbert Space-Filling Curve, then pack sorted data objects into chunks spatial index support: – Hilbert R-Tree deployed on master node – leaf nodes correspond to data chunks – initial filtering at master, tells workers which chunks to look HiPC’12 Pune, India o1o1 o4o4 o3o3 o8o8 o6o6 o5o5 o2o2 o7o7 sorted objects: o 1, o 3, o 8, o 6, o 2, o 7, o 4, o 5 chunk 1chunk 2chunk 3chunk 4

Range Queries on Spatial Data Fault-Tolerance Support – Sub-chunk Replication: step1: divide data chunks into k sub-chunks step2: distribute sub-chunks in round-robin fashion HiPC’12 Pune, India 9 Worker 1Worker 2 Worker 3Worker 4 chunk1chunk2 chunk3chunk4 chunk1,1chunk1,2 step1 chunk2,1chunk2,2 step1 chunk3,1chunk3,2 step1 chunk4,1chunk4,2 step1 * rack-failure: same approach, but distribute sub-chunks to nodes in different rack k = 2

Range Queries on Spatial Data Fault-Tolerance Support - Bookkeeping: – add a sub-leaf level to the bottom of Hilbert R-Tree – Hilbert R-Tree both as a filtering structure and failure management tool HiPC’12 Pune, India 10

Aggregation Queries on Point Data nature of the task: – each data object is a point in 2D space – each query is defined with a dimension (X or Y), and aggregation function (SUM, AVG, …) computational model: – master/worker model – divide space into M partitions – no indexing support – standard 2-phase algorithm: local and global aggregation HiPC’12 Pune, India 11 worker 1 worker 2 worker 3 worker 4 X Y partial result in worker 2 M = 4

Aggregation Queries on Point Data reducing communication volume – initial partitioning scheme has a direct impact – have insights about data and query workload: P(X) and P(Y) = probability of aggregation along X and Y-axis |r x | and |r y | = range of X and Y coordinates expected communication volume V comm defined as: Goal: choose a partitioning scheme (c v and c h ) that minimizes V comm HiPC’12 Pune, India 12

Aggregation Queries on Point Data Fault-Tolerance Support – Sub-partition Replication: step1: divide each partition evenly into M’ sub-partitions step2: send each of M’ sub-partitions to a different worker node Important questions: 1)how many sub-partitions (M’)? 2)how to divide a partition (cv’ and ch’) ? 3)where to send each sub-partition? (random vs. rule-based) HiPC’12 Pune, India 13 Y X M’ = 4 ch’ = 2 cv’ = 2 a better distribution reduces comm. overhead rule-based selection: assign to nodes which share the same coordinate- range

Experiments local cluster with nodes – two quad-core 2.53 GHz Xeon(R) processors with 12 GB RAM entire system implemented in C by using MPI-library range queries: – comparison with chunk replication scheme – 32 GB spatial data – 1000 queries are run, and aggregate time is reported aggregation queries: – comparison with partition replication scheme – 24 GB point data 64 nodes used, unless noted otherwise HiPC’12 Pune, India 14

Experiments: Range Queries Optimal Chunk Size SelectionScalability HiPC’12 Pune, India 15 - Execution Times with No Replication and No Failures (chunk size = 10000)

Experiments: Range Queries Single-Machine FailureRack Failure HiPC’12 Pune, India 16 -Execution Times under Failure Scenarios (64 workers in total) -k is the number of sub-chunks for a chunk

Experiments: Aggregation Queries Effect of Partitioning Scheme On Normal Execution Single-Machine Failure HiPC’12 Pune, India 17 P(X) = P(Y) = 0.5, |r x | = |r y | = P(X) = P(Y) = 0.5, |r x | = |r y | =

Conclusion a fault-tolerant environment that can process – range queries on spatial data and aggregation queries on point data – but, proposed approaches can be extended for other type of queries and analysis tasks high efficiency under normal execution sub-chunk and sub-partition replications – preserve load-balance in presence of failures, and hence – outperform traditional replication schemes HiPC’12 Pune, India 18

Thank you for listening … Questions HiPC’12 Pune, India 19