1 A Framework for Data-Intensive Computing with Cloud Bursting Tekin Bicer David ChiuGagan Agrawal Department of Compute Science and Engineering The Ohio.

Slides:



Advertisements
Similar presentations
IPDPS Boston Integrating Online Compression to Accelerate Large-Scale Data Analytics Applications Tekin Bicer, Jian Yin, David Chiu, Gagan Agrawal.
Advertisements

University of Minnesota Optimizing MapReduce Provisioning in the Cloud Michael Cardosa, Aameek Singh†, Himabindu Pucha†, Abhishek Chandra
Piccolo: Building fast distributed programs with partitioned tables Russell Power Jinyang Li New York University.
SLA-Oriented Resource Provisioning for Cloud Computing
Locality-Aware Dynamic VM Reconfiguration on MapReduce Clouds Jongse Park, Daewoo Lee, Bokyeong Kim, Jaehyuk Huh, Seungryoul Maeng.
Virtual Machine Usage in Cloud Computing for Amazon EE126: Computer Engineering Connor Cunningham Tufts University 12/1/14 “Virtual Machine Usage in Cloud.
Cloud Computing Resource provisioning Keke Chen. Outline  For Web applications statistical Learning and automatic control for datacenters  For data.
Dinker Batra CLUSTERING Categories of Clusters. Dinker Batra Introduction A computer cluster is a group of linked computers, working together closely.
IMapReduce: A Distributed Computing Framework for Iterative Computation Yanfeng Zhang, Northeastern University, China Qixin Gao, Northeastern University,
Authors: Thilina Gunarathne, Tak-Lon Wu, Judy Qiu, Geoffrey Fox Publish: HPDC'10, June 20–25, 2010, Chicago, Illinois, USA ACM Speaker: Jia Bao Lin.
Presented by Sujit Tilak. Evolution of Client/Server Architecture Clients & Server on different computer systems Local Area Network for Server and Client.
MapReduce in the Clouds for Science CloudCom 2010 Nov 30 – Dec 3, 2010 Thilina Gunarathne, Tak-Lon Wu, Judy Qiu, Geoffrey Fox {tgunarat, taklwu,
Computer Science and Engineering A Middleware for Developing and Deploying Scalable Remote Mining Services P. 1DataGrid Lab A Middleware for Developing.
MATE-EC2: A Middleware for Processing Data with Amazon Web Services Tekin Bicer David Chiu* and Gagan Agrawal Department of Compute Science and Engineering.
ANL Chicago Elastic and Efficient Execution of Data- Intensive Applications on Hybrid Cloud Tekin Bicer Computer Science and Engineering Ohio State.
Analysis of Remote Sensing Quantitative Inversion in Cloud Computing Jing Dong, Yong Xue, Ziqiang Chen, Hui Xu, Yingjie Li Institute of Remote Sensing.
IPDPS, Supporting Fault Tolerance in a Data-Intensive Computing Middleware Tekin Bicer, Wei Jiang and Gagan Agrawal Department of Computer Science.
Ch 4. The Evolution of Analytic Scalability
A Brief Overview by Aditya Dutt March 18 th ’ Aditya Inc.
Designing Efficient Systems Services and Primitives for Next-Generation Data-Centers K. Vaidyanathan, S. Narravula, P. Balaji and D. K. Panda Network Based.
Department of Computer Science Engineering SRM University
Cloud Computing 1. Outline  Introduction  Evolution  Cloud architecture  Map reduce operation  Platform 2.
Marcos Dias de Assunção 1,2, Alexandre di Costanzo 1 and Rajkumar Buyya 1 1 Department of Computer Science and Software Engineering 2 National ICT Australia.
CS525: Special Topics in DBs Large-Scale Data Management Hadoop/MapReduce Computing Paradigm Spring 2013 WPI, Mohamed Eltabakh 1.
Ex-MATE: Data-Intensive Computing with Large Reduction Objects and Its Application to Graph Mining Wei Jiang and Gagan Agrawal.
1 Quincy: Fair Scheduling for Distributed Computing Clusters Michael Isard, Vijayan Prabhakaran, Jon Currey, Udi Wieder, Kunal Talwar, and Andrew Goldberg.
Performance Issues in Parallelizing Data-Intensive applications on a Multi-core Cluster Vignesh Ravi and Gagan Agrawal
EXPOSE GOOGLE APP ENGINE AS TASKTRACKER NODES AND DATA NODES.
Introduction to Hadoop and HDFS
An Autonomic Framework in Cloud Environment Jiedan Zhu Advisor: Prof. Gagan Agrawal.
1 Time & Cost Sensitive Data-Intensive Computing on Hybrid Clouds Tekin Bicer David ChiuGagan Agrawal Department of Compute Science and Engineering The.
A Framework for Elastic Execution of Existing MPI Programs Aarthi Raveendran Tekin Bicer Gagan Agrawal 1.
CCGrid 2014 Improving I/O Throughput of Scientific Applications using Transparent Parallel Compression Tekin Bicer, Jian Yin and Gagan Agrawal Ohio State.
Large Scale Sky Computing Applications with Nimbus Pierre Riteau Université de Rennes 1, IRISA INRIA Rennes – Bretagne Atlantique Rennes, France
A Framework for Elastic Execution of Existing MPI Programs Aarthi Raveendran Graduate Student Department Of CSE 1.
Porting Irregular Reductions on Heterogeneous CPU-GPU Configurations Xin Huo, Vignesh T. Ravi, Gagan Agrawal Department of Computer Science and Engineering.
Euro-Par, A Resource Allocation Approach for Supporting Time-Critical Applications in Grid Environments Qian Zhu and Gagan Agrawal Department of.
Data-Intensive and High Performance Computing on Cloud Environments Gagan Agrawal 1.
A Map-Reduce System with an Alternate API for Multi-Core Environments Wei Jiang, Vignesh T. Ravi and Gagan Agrawal.
A Hierarchical MapReduce Framework Yuan Luo and Beth Plale School of Informatics and Computing, Indiana University Data To Insight Center, Indiana University.
Computer Science and Engineering Parallelizing Defect Detection and Categorization Using FREERIDE Leonid Glimcher P. 1 ipdps’05 Scaling and Parallelizing.
CCGrid, 2012 Supporting User Defined Subsetting and Aggregation over Parallel NetCDF Datasets Yu Su and Gagan Agrawal Department of Computer Science and.
Optimizing MapReduce for GPUs with Effective Shared Memory Usage Department of Computer Science and Engineering The Ohio State University Linchuan Chen.
Compiler and Runtime Support for Enabling Generalized Reduction Computations on Heterogeneous Parallel Configurations Vignesh Ravi, Wenjing Ma, David Chiu.
Department of Computer Science MapReduce for the Cell B. E. Architecture Marc de Kruijf University of Wisconsin−Madison Advised by Professor Sankaralingam.
Virtualization and Databases Ashraf Aboulnaga University of Waterloo.
CS525: Big Data Analytics MapReduce Computing Paradigm & Apache Hadoop Open Source Fall 2013 Elke A. Rundensteiner 1.
DynamicMR: A Dynamic Slot Allocation Optimization Framework for MapReduce Clusters Nanyang Technological University Shanjiang Tang, Bu-Sung Lee, Bingsheng.
Elastic Cloud Caches for Accelerating Service-Oriented Computations Gagan Agrawal Ohio State University Columbus, OH David Chiu Washington State University.
Supporting Load Balancing for Distributed Data-Intensive Applications Leonid Glimcher, Vignesh Ravi, and Gagan Agrawal Department of ComputerScience and.
Rapid Tomographic Image Reconstruction via Large-Scale Parallelization Ohio State University Computer Science and Engineering Dep. Gagan Agrawal Argonne.
PDAC-10 Middleware Solutions for Data- Intensive (Scientific) Computing on Clouds Gagan Agrawal Ohio State University (Joint Work with Tekin Bicer, David.
OSU – CSE 2014 Supporting Data-Intensive Scientific Computing on Bandwidth and Space Constrained Environments Tekin Bicer Department of Computer Science.
MATE-CG: A MapReduce-Like Framework for Accelerating Data-Intensive Computations on Heterogeneous Clusters Wei Jiang and Gagan Agrawal.
Hadoop/MapReduce Computing Paradigm 1 CS525: Special Topics in DBs Large-Scale Data Management Presented By Kelly Technologies
A Two-phase Execution Engine of Reduce Tasks In Hadoop MapReduce XiaohongZhang*GuoweiWang* ZijingYang*YangDing School of Computer Science and Technology.
AUTO-GC: Automatic Translation of Data Mining Applications to GPU Clusters Wenjing Ma Gagan Agrawal The Ohio State University.
3/12/2013Computer Engg, IIT(BHU)1 CLOUD COMPUTING-2.
Evaluating and Optimizing Indexing Schemes for a Cloud-based Elastic Key- Value Store Apeksha Shetty and Gagan Agrawal Ohio State University David Chiu.
Auburn University COMP7500 Advanced Operating Systems I/O-Aware Load Balancing Techniques (2) Dr. Xiao Qin Auburn University.
Accelerating MapReduce on a Coupled CPU-GPU Architecture
Tools and Techniques for Processing (and Management) of Data
Optimizing MapReduce for GPUs with Effective Shared Memory Usage
Ch 4. The Evolution of Analytic Scalability
Wei Jiang Advisor: Dr. Gagan Agrawal
Data-Intensive Computing: From Clouds to GPU Clusters
Syllabus and Introduction Keke Chen
Yi Wang, Wei Jiang, Gagan Agrawal
Resource Allocation for Distributed Streaming Applications
L. Glimcher, R. Jin, G. Agrawal Presented by: Leo Glimcher
Presentation transcript:

1 A Framework for Data-Intensive Computing with Cloud Bursting Tekin Bicer David ChiuGagan Agrawal Department of Compute Science and Engineering The Ohio State University School of Engineering and Computer Science Washington State University † † Cluster Texas Austin

Outline Introduction Motivation Challenges MATE-EC2 MATE-EC2 and Cloud Bursting Experiments Conclusion 2 Cluster Texas Austin

Data-Intensive and Cloud Comp. Data-Intensive Computing – Need for large storage, processing and bandwidth – Traditionally on supercomputers or local clusters Resources can be exhausted Cloud Environments – Pay-as-you-go model – Availability of elastic storage and processing e.g. AWS, Microsoft Azure, Google Apps etc. – Unavailability of high performance inter-connect Cluster Compute Instances, Cluster GPU instances Cluster Texas Austin

Cloud Bursting - Motivation In-house dedicated machines –Demand for more resources –Workload might vary in time Cloud resources Collaboration between local and remote resources –Local resources: base workload –Cloud resources: extra workload from users 4 Cluster Texas Austin

Cloud Bursting - Challenges Cooperation of the resources –Minimizing the system overhead –Distribution of the data –Job assignments Determining workload 5 Cluster Texas Austin

Outline Introduction Motivation Challenges MATE MATE-EC2 and Cloud Bursting Experiments Conclusion 6 Cluster Texas Austin

MATE vs. Map-Reduce Processing Structure 7 Reduction Object represents the intermediate state of the execution Reduce func. is commutative and associative Sorting, grouping.. overheads are eliminated with red. func/obj. Cluster Texas Austin

MATE on Amazon EC2 Data organization –Metadata information –Three levels: Buckets/Files, Chunks and Units Chunk Retrieval –S3: Threaded Data Retrieval –Local: Cont. read –Selective Job Assignment Load Balancing and handling heterogeneity –Pooling mechanism 8 Cluster Texas Austin

MATE-EC2 Processing Flow for AWS C 0 C 5 C n Computing Layer Job Scheduler Job Pool Request Job from Master NodeC 0 is assigned as job Retrieve chunk pieces and Write them into the buffer T 0 T 1 T 2 T 3 Pass retrieved chunk to Computing Layer and process Request another job C 5 is assigned as a job Retrieve the new job EC2 Slave Node S3 Data Object EC2 Master Node 9

System Overview for Cloud Bursting (1) Local cluster(s) and Cloud Environment Map-Reduce type of processing All the clusters connect to a centralized node – Coarse grained job assignment – Consideration of locality Each clusters has a Master node – Fine grained job assignment Work Stealing Cluster Texas Austin 10

System Overview for Cloud Bursting(2) Cluster Texas Austin 11

Experiments 2 geographically distributed clusters –Cloud: EC2 instances running on Virginia –Local: Campus cluster (Columbus, OH) 3 applications with 120GB of data –Kmeans: k=1000; Knn: k=1000; PageRank: 50x10 links w/ 9.2x10 edges Goals: –Evaluating the system overhead with different job distributions –Evaluating the scalability of the system 12 Cluster Texas Austin 68

System Overhead: K-Means 13 Cluster Texas Austin Env-*Global Reduction Idle TimeTotal SlowdownStolen # Jobs (960) localEC2 50/ (0.5%)0 33/ (5.9%)128 17/ (10.4%)240

System Overhead: PageRank 14 Cluster Texas Austin Env-*Global Reduction Idle TimeTotal SlowdownStolen # Jobs (960) localEC2 50/ (10.5%)0 33/ (18.9%)112 17/ (30.8%)240

Scalability: K-Means 15 Cluster Texas Austin

Scalability: PageRank 16 Cluster Texas Austin

Conclusion MATE-EC2 is a data intensive middleware developed for Cloud Bursting Hybrid cloud is new – Most of Map-Reduce implementations consider local cluster(s); no known system for cloud bursting Our results show that – Inter-cluster comm. overhead is low in most data-intensive app. – Job distribution is important – Overall slowdown is modest even the disproportion in data dist. increases; our system is scalable 17

Thanks Any Questions? 18 Cluster Texas Austin

System Overhead: KNN 19 Cluster Texas Austin Env-*Global Reduction Idle TimeTotal Slowdown Stolen # Jobs (960) localEC2 50/ (1.7%)0 33/ (15.4%)64 17/ (45.9%)128

Scalability: KNN 20 Cluster Texas Austin

Future Work Cloud bursting can answer user requirements (De)allocate resources on cloud Time constraint – Given time, minimize the cost on cloud Cost constraint – Given cost, minimize the execution time Cluster Texas Austin

References The Cost of Doing Science on the Cloud (Deelman et. Al.; SC’08) Data Sharing Options for Scientific Workflow on Amazon EC2 (Deelman et. Al.; SC’10) Amazon S3 for Science Grids: A viable solution? (Palankar et. al.; DADC’08) Evaluating the Cost Benefit of Using Cloud Computing to Extend the Capacity of Clusters. (Assuncao et. al.; HPDC’09) Elastic Site: Using Clouds to Elastically Extend Site Resources (Marshall et. al.; CCGRID’10) Towards Optimizing Hadoop Provisioning in the Cloud. (Kambatla et. Al.; HotCloud’09) Cluster Texas Austin 22