Supporting Load Balancing for Distributed Data-Intensive Applications Leonid Glimcher, Vignesh Ravi, and Gagan Agrawal Department of ComputerScience and.

Slides:



Advertisements
Similar presentations
SkewReduce YongChul Kwon Magdalena Balazinska, Bill Howe, Jerome Rolia* University of Washington, *HP Labs Skew-Resistant Parallel Processing of Feature-Extracting.
Advertisements

“Compiler and Runtime Support for Enabling Generalized Reduction Computations on Heterogeneous Parallel Computations” By Ravi, Ma, Chiu, & Agrawal Presented.
1 Introduction to Load Balancing: l Definition of Distributed systems. Collection of independent loosely coupled computing resources. l Load Balancing.
Computer Science and Engineering A Middleware for Developing and Deploying Scalable Remote Mining Services P. 1DataGrid Lab A Middleware for Developing.
MATE-EC2: A Middleware for Processing Data with Amazon Web Services Tekin Bicer David Chiu* and Gagan Agrawal Department of Compute Science and Engineering.
IPDPS, Supporting Fault Tolerance in a Data-Intensive Computing Middleware Tekin Bicer, Wei Jiang and Gagan Agrawal Department of Computer Science.
Word Wide Cache Distributed Caching for the Distributed Enterprise.
Scheduling in Heterogeneous Grid Environments: The Effects of Data Migration Leonid Oliker, Hongzhang Shan Future Technology Group Lawrence Berkeley Research.
Venkatram Ramanathan 1. Motivation Evolution of Multi-Core Machines and the challenges Background: MapReduce and FREERIDE Co-clustering on FREERIDE Experimental.
Exploiting Domain-Specific High-level Runtime Support for Parallel Code Generation Xiaogang Li Ruoming Jin Gagan Agrawal Department of Computer and Information.
Ex-MATE: Data-Intensive Computing with Large Reduction Objects and Its Application to Graph Mining Wei Jiang and Gagan Agrawal.
PAGE: A Framework for Easy Parallelization of Genomic Applications 1 Mucahid Kutlu Gagan Agrawal Department of Computer Science and Engineering The Ohio.
Active Monitoring in GRID environments using Mobile Agent technology Orazio Tomarchio Andrea Calvagna Dipartimento di Ingegneria Informatica e delle Telecomunicazioni.
Performance Issues in Parallelizing Data-Intensive applications on a Multi-core Cluster Vignesh Ravi and Gagan Agrawal
1 Configurable Security for Scavenged Storage Systems NetSysLab The University of British Columbia Abdullah Gharaibeh with: Samer Al-Kiswany, Matei Ripeanu.
Ohio State University Department of Computer Science and Engineering 1 Cyberinfrastructure for Coastal Forecasting and Change Analysis Gagan Agrawal Hakan.
An Autonomic Framework in Cloud Environment Jiedan Zhu Advisor: Prof. Gagan Agrawal.
1 Time & Cost Sensitive Data-Intensive Computing on Hybrid Clouds Tekin Bicer David ChiuGagan Agrawal Department of Compute Science and Engineering The.
A Framework for Elastic Execution of Existing MPI Programs Aarthi Raveendran Tekin Bicer Gagan Agrawal 1.
CCGrid 2014 Improving I/O Throughput of Scientific Applications using Transparent Parallel Compression Tekin Bicer, Jian Yin and Gagan Agrawal Ohio State.
1 A Framework for Data-Intensive Computing with Cloud Bursting Tekin Bicer David ChiuGagan Agrawal Department of Compute Science and Engineering The Ohio.
A Framework for Elastic Execution of Existing MPI Programs Aarthi Raveendran Graduate Student Department Of CSE 1.
Porting Irregular Reductions on Heterogeneous CPU-GPU Configurations Xin Huo, Vignesh T. Ravi, Gagan Agrawal Department of Computer Science and Engineering.
Evaluating FERMI features for Data Mining Applications Masters Thesis Presentation Sinduja Muralidharan Advised by: Dr. Gagan Agrawal.
Euro-Par, A Resource Allocation Approach for Supporting Time-Critical Applications in Grid Environments Qian Zhu and Gagan Agrawal Department of.
Data-Intensive Computing: From Multi-Cores and GPGPUs to Cloud Computing and Deep Web Gagan Agrawal u.
Integrating and Optimizing Transactional Memory in a Data Mining Middleware Vignesh Ravi and Gagan Agrawal Department of ComputerScience and Engg. The.
A Map-Reduce System with an Alternate API for Multi-Core Environments Wei Jiang, Vignesh T. Ravi and Gagan Agrawal.
Data-Intensive Computing: From Clouds to GPUs Gagan Agrawal June 1,
Computer Science and Engineering Predicting Performance for Grid-Based P. 1 IPDPS’07 A Performance Prediction Framework.
Computer Science and Engineering Parallelizing Defect Detection and Categorization Using FREERIDE Leonid Glimcher P. 1 ipdps’05 Scaling and Parallelizing.
FREERIDE: System Support for High Performance Data Mining Ruoming Jin Leo Glimcher Xuan Zhang Ge Yang Gagan Agrawal Department of Computer and Information.
CCGrid 2014 Improving I/O Throughput of Scientific Applications using Transparent Parallel Compression Tekin Bicer, Jian Yin and Gagan Agrawal Ohio State.
High-level Interfaces and Abstractions for Data-Driven Applications in a Grid Environment Gagan Agrawal Department of Computer Science and Engineering.
CCGrid, 2012 Supporting User Defined Subsetting and Aggregation over Parallel NetCDF Datasets Yu Su and Gagan Agrawal Department of Computer Science and.
Optimizing MapReduce for GPUs with Effective Shared Memory Usage Department of Computer Science and Engineering The Ohio State University Linchuan Chen.
Data-Intensive Computing: From Clouds to GPUs Gagan Agrawal December 3,
Compiler and Runtime Support for Enabling Generalized Reduction Computations on Heterogeneous Parallel Configurations Vignesh Ravi, Wenjing Ma, David Chiu.
1 Supporting Dynamic Migration in Tightly Coupled Grid Applications Liang Chen Qian Zhu Gagan Agrawal Computer Science & Engineering The Ohio State University.
DynamicMR: A Dynamic Slot Allocation Optimization Framework for MapReduce Clusters Nanyang Technological University Shanjiang Tang, Bu-Sung Lee, Bingsheng.
Computer Science and Engineering FREERIDE-G: A Grid-Based Middleware for Scalable Processing of Remote Data Leonid Glimcher Gagan Agrawal.
RE-PAGE: Domain-Specific REplication and PArallel Processing of GEnomic Applications 1 Mucahid Kutlu Gagan Agrawal Department of Computer Science and Engineering.
Grid Appliance The World of Virtual Resource Sharing Group # 14 Dhairya Gala Priyank Shah.
Rapid Tomographic Image Reconstruction via Large-Scale Parallelization Ohio State University Computer Science and Engineering Dep. Gagan Agrawal Argonne.
PDAC-10 Middleware Solutions for Data- Intensive (Scientific) Computing on Clouds Gagan Agrawal Ohio State University (Joint Work with Tekin Bicer, David.
System Support for High Performance Data Mining Ruoming Jin Leo Glimcher Xuan Zhang Gagan Agrawal Department of Computer and Information Sciences Ohio.
High-level Interfaces for Scalable Data Mining Ruoming Jin Gagan Agrawal Department of Computer and Information Sciences Ohio State University.
Ohio State University Department of Computer Science and Engineering Servicing Range Queries on Multidimensional Datasets with Partial Replicas Li Weng,
Data Consolidation: A Task Scheduling and Data Migration Technique for Grid Networks Author: P. Kokkinos, K. Christodoulopoulos, A. Kretsis, and E. Varvarigos.
AUTO-GC: Automatic Translation of Data Mining Applications to GPU Clusters Wenjing Ma Gagan Agrawal The Ohio State University.
Research Overview Gagan Agrawal Associate Professor.
Euro-Par, HASTE: An Adaptive Middleware for Supporting Time-Critical Event Handling in Distributed Environments ICAC 2008 Conference June 2 nd,
Porting Irregular Reductions on Heterogeneous CPU-GPU Configurations Xin Huo Vignesh T. Ravi Gagan Agrawal Department of Computer Science and Engineering,
Architecture for Resource Allocation Services Supporting Interactive Remote Desktop Sessions in Utility Grids Vanish Talwar, HP Labs Bikash Agarwalla,
System Support for High Performance Scientific Data Mining Gagan Agrawal Ruoming Jin Raghu Machiraju S. Parthasarathy Department of Computer and Information.
Computer Science and Engineering Parallelizing Feature Mining Using FREERIDE Leonid Glimcher P. 1 ipdps’04 Scaling and Parallelizing a Scientific Feature.
Optimizing Distributed Actor Systems for Dynamic Interactive Services
A Dynamic Scheduling Framework for Emerging Heterogeneous Systems
Introduction to Load Balancing:
Author: Ragalatha P, Manoj Challa, Sundeep Kumar. K
Accelerating MapReduce on a Coupled CPU-GPU Architecture
Supporting Fault-Tolerance in Streaming Grid Applications
Optimizing MapReduce for GPUs with Effective Shared Memory Usage
Data-Intensive Computing: From Clouds to GPU Clusters
An Adaptive Middleware for Supporting Time-Critical Event Response
Resource Allocation in a Middleware for Streaming Data
A Grid-Based Middleware for Scalable Processing of Remote Data
Resource Allocation for Distributed Streaming Applications
FREERIDE: A Framework for Rapid Implementation of Datamining Engines
L. Glimcher, R. Jin, G. Agrawal Presented by: Leo Glimcher
Presentation transcript:

Supporting Load Balancing for Distributed Data-Intensive Applications Leonid Glimcher, Vignesh Ravi, and Gagan Agrawal Department of ComputerScience and Engg. The Ohio State University Columbus, Ohio

Outline Introduction Motivation FREERIDE-G Processing Structure Run-time Load Balancing System Experimental Results Conclusions December 24, 20152

Introduction Growing abundance of data –Sensors, scientific simulations and business transactions Data Analysis –Translate raw data into knowledge Grid/Cloud Computing –Enables distributed processing December 24, 20153

Motivation Resources are geographically distributed –Data nodes –Compute nodes –Middleware user Remote data analysis is important Heterogeneity of resources –Difference in network bandwidth –Difference in compute power December 24, Data Nodes Compute Nodes Middleware user Grid/Cloud Environment

FREERIDE-G Processing Structure (Framework for Rapid Implementation of Datamining Engines – Grid) December 24, While( ) { forall( data instances d) { (I, d’) = process(d) R(I) = R(I) op d’ } ……. } A Map-reduce like system Remote data analysis Middleware API Process Reduce Global Combine Reduction Object

A Real-time Grid/Cloud Scenario December 24, A B C D Compute Data

Run-time Load Balancing December 24, Two factors of load imbalance Computational factor, w1 Remote data transfer (wait time), w2 Case 1: w1 > w2 Case 2: w2 > w1 We use sum of weights to account for both the components

Dynamic Load Balancing Algorithm December 24, Consider every chunk, Ci Calculate Compute cost, Cc Calculate Data transfer cost, Tc Input Bandwidth matrix, W1 & W2 Total cost = W1*Cc + W2*Tc If Total cost < Min Update Min Assign Ci to Pj

Experimental Setup Settings Organizational Grid Wide Area Network (WAN) Goals are to evaluate Scalability Dynamic Load balancing overhead Adaptability to scenarios –compute bound, –I/O bound, –WAN setting Applications K-means Vortex Detection December 24, 20159

10 Scalability and Overhead of Dynamic Balancing Vortex detection 14.8 GB data Organizational setting Bandwidth –50mb/sec –100mb/sec 31% benefit Overhead within 10% December 24,

Model Adaptability – Compute Bound Scenario Kmeans clustering 25.6 GB data Bandwidth –50 MB –200 MB Best result combination skewed towards work load component Initial (unbalanced) overhead 57% over balanced Dynamic overhead 5% over balanced December 24, Ideal Case Dynamic case Compute Data transfer

Model Adaptability – I/O Bound Scenario December 24, Kmeans clustering 25.6 GB data Bandwidth –15 mb/s –60 mb/s Best result combination skewed towards data transfer component Initial (unbalanced) overhead 40% over balanced Dynamic overhead 4% over balanced

Model Adaptability – WAN setting Vortex Detection 14.6 GB Best result combination results in lowest overhead (favoring data delivery component) Unbalanced configuration 20% overhead over balanced Our approach Overhead reduced to 8% December 24,

Conclusions Dynamic load balancing solution for grid environments Both workload and data transfer factors are important Scalability is good and overheads are within 10% Adaptable to compute-bound, I/O bound, and WAN settings December 24,

December 24, Thank You! Questions? Contacts: Leonid Glimcher Vignesh Ravi- Gagan Agrawal-

P. 16 DataGrid Lab Setup 1: Organizational Grid Data hosted on Opteron 250’s Processed on Opteron 254’s 2 clusters connected through two 10 GB optical fibers Both clusters within same city (0.5 mile apart) Evaluating: Scalability Adaptability Integration overhead Compute cluster (cse-ri) Repository cluster (bmi-ri)

P. 17 DataGrid Lab Setup 2: WAN Data Repository: Opteron 250’s (OSU) Opteron 258’s (Kent St) Processed on Opteron 254’s No dedicated link between processing and repository clusters Evaluating: Scalability Adaptability Compute cluster (OSU ) Repository cluster (Kent ST) Repository cluster (OSU)

FREERIDE-G System Design December 24,