Computer Science and Engineering Predicting Performance for Grid-Based P. 1 IPDPS’07 A Performance Prediction Framework.

Slides:



Advertisements
Similar presentations
Designing Services for Grid-based Knowledge Discovery A. Congiusta, A. Pugliese, Domenico Talia, P. Trunfio DEIS University of Calabria ITALY
Advertisements

IPDPS Boston Integrating Online Compression to Accelerate Large-Scale Data Analytics Applications Tekin Bicer, Jian Yin, David Chiu, Gagan Agrawal.
SkewReduce YongChul Kwon Magdalena Balazinska, Bill Howe, Jerome Rolia* University of Washington, *HP Labs Skew-Resistant Parallel Processing of Feature-Extracting.
1 Virtual Machine Resource Monitoring and Networking of Virtual Machines Ananth I. Sundararaj Department of Computer Science Northwestern University July.
Homework 2 In the docs folder of your Berkeley DB, have a careful look at documentation on how to configure BDB in main memory. In the docs folder of your.
Real Parallel Computers. Background Information Recent trends in the marketplace of high performance computing Strohmaier, Dongarra, Meuer, Simon Parallel.
Computer Science and Engineering A Middleware for Developing and Deploying Scalable Remote Mining Services P. 1DataGrid Lab A Middleware for Developing.
MATE-EC2: A Middleware for Processing Data with Amazon Web Services Tekin Bicer David Chiu* and Gagan Agrawal Department of Compute Science and Engineering.
Real Parallel Computers. Modular data centers Background Information Recent trends in the marketplace of high performance computing Strohmaier, Dongarra,
IPDPS, Supporting Fault Tolerance in a Data-Intensive Computing Middleware Tekin Bicer, Wei Jiang and Gagan Agrawal Department of Computer Science.
Exploiting Domain-Specific High-level Runtime Support for Parallel Code Generation Xiaogang Li Ruoming Jin Gagan Agrawal Department of Computer and Information.
Ohio State University Department of Computer Science and Engineering Automatic Data Virtualization - Supporting XML based abstractions on HDF5 Datasets.
Profiling Grid Data Transfer Protocols and Servers George Kola, Tevfik Kosar and Miron Livny University of Wisconsin-Madison USA.
PAGE: A Framework for Easy Parallelization of Genomic Applications 1 Mucahid Kutlu Gagan Agrawal Department of Computer Science and Engineering The Ohio.
Preparing for the Poster Session Gagan Agrawal. Outline Background on the proposal Overall research focus Equipment requested Preparing for the Site Visit.
A Metadata Based Approach For Supporting Subsetting Queries Over Parallel HDF5 Datasets Vignesh Santhanagopalan Graduate Student Department Of CSE.
Performance Issues in Parallelizing Data-Intensive applications on a Multi-core Cluster Vignesh Ravi and Gagan Agrawal
Ohio State University Department of Computer Science and Engineering 1 Cyberinfrastructure for Coastal Forecasting and Change Analysis Gagan Agrawal Hakan.
Ohio State University Department of Computer Science and Engineering 1 Supporting SQL-3 Aggregations on Grid-based Data Repositories Li Weng, Gagan Agrawal,
The Data Grid: Towards an Architecture for the Distributed Management and Analysis of Large Scientific Dataset Caitlin Minteer & Kelly Clynes.
1 Time & Cost Sensitive Data-Intensive Computing on Hybrid Clouds Tekin Bicer David ChiuGagan Agrawal Department of Compute Science and Engineering The.
A Framework for Elastic Execution of Existing MPI Programs Aarthi Raveendran Tekin Bicer Gagan Agrawal 1.
Porting Irregular Reductions on Heterogeneous CPU-GPU Configurations Xin Huo, Vignesh T. Ravi, Gagan Agrawal Department of Computer Science and Engineering.
Shared Memory Parallelization of Decision Tree Construction Using a General Middleware Ruoming Jin Gagan Agrawal Department of Computer and Information.
Evaluating FERMI features for Data Mining Applications Masters Thesis Presentation Sinduja Muralidharan Advised by: Dr. Gagan Agrawal.
Euro-Par, A Resource Allocation Approach for Supporting Time-Critical Applications in Grid Environments Qian Zhu and Gagan Agrawal Department of.
Data-Intensive Computing: From Multi-Cores and GPGPUs to Cloud Computing and Deep Web Gagan Agrawal u.
Performance Prediction for Random Write Reductions: A Case Study in Modelling Shared Memory Programs Ruoming Jin Gagan Agrawal Department of Computer and.
Data-Intensive Computing: From Clouds to GPUs Gagan Agrawal June 1,
PREDIcT: Towards Predicting the Runtime of Iterative Analytics Adrian Popescu 1, Andrey Balmin 2, Vuk Ercegovac 3, Anastasia Ailamaki
Computer Science and Engineering Parallelizing Defect Detection and Categorization Using FREERIDE Leonid Glimcher P. 1 ipdps’05 Scaling and Parallelizing.
FREERIDE: System Support for High Performance Data Mining Ruoming Jin Leo Glimcher Xuan Zhang Ge Yang Gagan Agrawal Department of Computer and Information.
High-level Interfaces and Abstractions for Data-Driven Applications in a Grid Environment Gagan Agrawal Department of Computer Science and Engineering.
Data-Intensive Computing: From Clouds to GPUs Gagan Agrawal December 3,
Compiler and Runtime Support for Enabling Generalized Reduction Computations on Heterogeneous Parallel Configurations Vignesh Ravi, Wenjing Ma, David Chiu.
1 Supporting Dynamic Migration in Tightly Coupled Grid Applications Liang Chen Qian Zhu Gagan Agrawal Computer Science & Engineering The Ohio State University.
Implementing Data Cube Construction Using a Cluster Middleware: Algorithms, Implementation Experience, and Performance Ge Yang Ruoming Jin Gagan Agrawal.
Ohio State University Department of Computer Science and Engineering An Approach for Automatic Data Virtualization Li Weng, Gagan Agrawal et al.
DynamicMR: A Dynamic Slot Allocation Optimization Framework for MapReduce Clusters Nanyang Technological University Shanjiang Tang, Bu-Sung Lee, Bingsheng.
Supporting Load Balancing for Distributed Data-Intensive Applications Leonid Glimcher, Vignesh Ravi, and Gagan Agrawal Department of ComputerScience and.
Computer Science and Engineering FREERIDE-G: A Grid-Based Middleware for Scalable Processing of Remote Data Leonid Glimcher Gagan Agrawal.
RE-PAGE: Domain-Specific REplication and PArallel Processing of GEnomic Applications 1 Mucahid Kutlu Gagan Agrawal Department of Computer Science and Engineering.
PDAC-10 Middleware Solutions for Data- Intensive (Scientific) Computing on Clouds Gagan Agrawal Ohio State University (Joint Work with Tekin Bicer, David.
MATE-CG: A MapReduce-Like Framework for Accelerating Data-Intensive Computations on Heterogeneous Clusters Wei Jiang and Gagan Agrawal.
System Support for High Performance Data Mining Ruoming Jin Leo Glimcher Xuan Zhang Gagan Agrawal Department of Computer and Information Sciences Ohio.
High-level Interfaces for Scalable Data Mining Ruoming Jin Gagan Agrawal Department of Computer and Information Sciences Ohio State University.
Ohio State University Department of Computer Science and Engineering Servicing Range Queries on Multidimensional Datasets with Partial Replicas Li Weng,
Exploiting Computing Power of GPU for Data Mining Application Wenjing Ma, Leonid Glimcher, Gagan Agrawal.
Packet Size optimization for Supporting Coarse-Grained Pipelined Parallelism Wei Du Gagan Agrawal Ohio State University.
AUTO-GC: Automatic Translation of Data Mining Applications to GPU Clusters Wenjing Ma Gagan Agrawal The Ohio State University.
System Support for High Performance Scientific Data Mining Gagan Agrawal Ruoming Jin Raghu Machiraju S. Parthasarathy Department of Computer and Information.
Servicing Seismic and Oil Reservoir Simulation Data through Grid Data Services Sivaramakrishnan Narayanan, Tahsin Kurc, Umit Catalyurek and Joel Saltz.
Computer Science and Engineering Parallelizing Feature Mining Using FREERIDE Leonid Glimcher P. 1 ipdps’04 Scaling and Parallelizing a Scientific Feature.
Clouds , Grids and Clusters
A Dynamic Scheduling Framework for Emerging Heterogeneous Systems
Liang Chen Advisor: Gagan Agrawal Computer Science & Engineering
Supporting Fault-Tolerance in Streaming Grid Applications
Year 2 Updates.
Communication and Memory Efficient Parallel Decision Tree Construction
Soft Error Detection for Iterative Applications Using Offline Training
Data-Intensive Computing: From Clouds to GPU Clusters
Bin Ren, Gagan Agrawal, Brad Chamberlain, Steve Deitz
Resource Allocation in a Middleware for Streaming Data
Compiler Supported Coarse-Grained Pipelined Parallelism: Why and How
A Grid-Based Middleware for Scalable Processing of Remote Data
Resource Allocation for Distributed Streaming Applications
FREERIDE: A Framework for Rapid Implementation of Datamining Engines
FREERIDE: A Framework for Rapid Implementation of Datamining Engines
Cluster Computers.
L. Glimcher, R. Jin, G. Agrawal Presented by: Leo Glimcher
Presentation transcript:

Computer Science and Engineering Predicting Performance for Grid-Based P. 1 IPDPS’07 A Performance Prediction Framework for Grid- Based Data Mining Applications Leonid Glimcher Gagan Agrawal

Computer Science and Engineering Predicting Performance for Grid-Based P. 2 IPDPS’07 Motivating Scenario Data Repository Clusters Compute Clusters User ? 3 stages: Disk i/o, Network, Compute.

Computer Science and Engineering Predicting Performance for Grid-Based P. 3 IPDPS’07 Remote Data Analysis Remote data analysis –Grid is a good fit –Details can be very tedious Middleware abstracts away lots of development details Resource selection – crucial to performance Performance prediction facilitates resource selection

Computer Science and Engineering Predicting Performance for Grid-Based P. 4 IPDPS’07 Presentation Road Map Problem statement and motivation Middleware background Our performance prediction approach Experimental evaluation Related work Conclusions

Computer Science and Engineering Predicting Performance for Grid-Based P. 5 IPDPS’07 Problem Statement Given:  Parallel data processing application  Execution time break-down (profile)  Configurations of available computing resources  Dataset replicas in different size repositories Predict application execution time in order to select right dataset replica and resource configuration

Computer Science and Engineering Predicting Performance for Grid-Based P. 6 IPDPS’07 FREERIDE-G Design

Computer Science and Engineering Predicting Performance for Grid-Based P. 7 IPDPS’07 FREERIDE-G Processing KEY observation: most data mining algorithms follow canonical loop Middleware API: Subset of data to be processed Reduction object Local and global reduction operations Iterator While( ) { forall( data instances d) { I = process(d) R(I) = R(I) op d } ……. }

Computer Science and Engineering Predicting Performance for Grid-Based P. 8 IPDPS’07 Performance Prediction Approach 3 Phases of execution: –Retrieval at data server –Data delivery to compute node –Parallel processing at compute node Special processing structure: –Generalized reduction T exec = T disk + T network + T compute

Computer Science and Engineering Predicting Performance for Grid-Based P. 9 IPDPS’07 Needed profile information Numbers of  storage nodes (n)  compute nodes (c) Available bandwidth between these (b), in profile configuration Execution time breakdown:  data retrieval (t d )  network communication (t n )  data processing (t c ) components Dataset size (s) Reduction object information:  maximum size  communication time Global reduction time

Computer Science and Engineering Predicting Performance for Grid-Based P. 10 IPDPS’07 Data Retrieval and Communication Time Data Retrieval: Dataset size (s) and number of data hosts (n) for base profile and predicted configuration (s’ and n’). Used to scale t d. Data Communication: Also need dataset size and number of data hosts, as well as bandwidth (b and b’). Used to scale t n.

Computer Science and Engineering Predicting Performance for Grid-Based P. 11 IPDPS’07 Initial Data Processing Time Prediction Dataset size (s) and number of compute nodes (c): base profile (s,c) predicted profile (s’, c’) Used to scale up t c. Limitations – not modeling: Inter-processor communication time Global reduction time

Computer Science and Engineering Predicting Performance for Grid-Based P. 12 IPDPS’07 Modeling Interprocessor Communication Parallel computation involves communication of reduction object Communication time (T ro ) Reduction object size (r) Interprocessor bandwidth (w) Latency (l) Reduction object size either remains constant or scales linearly

Computer Science and Engineering Predicting Performance for Grid-Based P. 13 IPDPS’07 Modeling Global Reduction Global reduction time (T g ) is also serialized Depending on application, global reduction time: –Scales linearly with number of nodes but is constant independent of size –Stays constant independent of number of nodes, but scales linearly with data size

Computer Science and Engineering Predicting Performance for Grid-Based P. 14 IPDPS’07 Modeling Across Heterogeneous Clusters Need scaling factors for all 3 stages of computation (from a set of representative applications).

Computer Science and Engineering Predicting Performance for Grid-Based P. 15 IPDPS’07 FREERIDE-G Applications Data mining: K-means clustering KNN search EM clustering Scientific data processing: Vortex extraction (right) Molecular defect detection and categorization

Computer Science and Engineering Predicting Performance for Grid-Based P. 16 IPDPS’07 Experimental Setup Base: 700 MHz Pentiums connected through Myrinet LaNai 7.0 Heterogeneous prediction: 2.4 GHz Opteron 250’s connected through Infiniband (1Gb) Goal – to correctly model changes in: 1.Parallel configuration 2.Dataset size 3.Network bandwidth 4.Underlying resources

Computer Science and Engineering Predicting Performance for Grid-Based P. 17 IPDPS’07 Modeling Parallel Performance Errors for 3 approaches for: 1.Vortex detection, base: 1-1 configuration 710 MB dataset 2.Defect detection, base: 1-1 configuration 130 MB dataset Results: modeling reduction pays off accurate predictions

Computer Science and Engineering Predicting Performance for Grid-Based P. 18 IPDPS’07 Modeling Dataset Size Errors for 1 (best) approach for: 1.EM clustering (1.4 GB), base: 1-1 configuration 350 MB dataset 2.Defect detection (1.8 GB), base: 1-1 configuration 130 MB dataset Results: biggest error when number of data nodes is same as number of compute nodes accurate predictions

Computer Science and Engineering Predicting Performance for Grid-Based P. 19 IPDPS’07 Impact of Network Bandwidth Errors for 1 (best) approach for: 1.EM clustering (250 Kbps), base: 1-1 configuration 500 Kbps 2.Defect detection (250 Kbps), base: 1-1 configuration 500 Kbps Results: biggest error when number of data nodes is same as number of compute nodes Modeling reduction is most accurate

Computer Science and Engineering Predicting Performance for Grid-Based P. 20 IPDPS’07 Predictions for different type of cluster Errors for 1 (best) approach for: 1.Defect detection (1.8 GB), base: 1-1 configuration 710 MB dataset 2.EM clustering (700 MB), base: 8-8 configuration 350 MB dataset Results: Scaling factors different Largest error when predicted configuration has same number of compute nodes as base

Computer Science and Engineering Predicting Performance for Grid-Based P. 21 IPDPS’07 Existing Work 3 broad categories for resource allocation:  Heuristic approach to mapping  Prediction through modeling: Statistical estimation/prediction Analytical modeling of parallel application  Simulation based performance prediction

Computer Science and Engineering Predicting Performance for Grid-Based P. 22 IPDPS’07 Summary Performance prediction approach Exploits similarities in application processing structure to come up with very accurate results Approach accurately models changes in: –Computing configuration –Dataset size –Network bandwidth –Underlying compute resources