LDBC-Benchmarking Graph-Processing Platforms: A Vision Benchmarking Graph-Processing Platforms: A Vision (A SPEC Research Group Process) Delft University.

Slides:



Advertisements
Similar presentations
LIBRA: Lightweight Data Skew Mitigation in MapReduce
Advertisements

1 HYRISE – A Main Memory Hybrid Storage Engine By: Martin Grund, Jens Krüger, Hasso Plattner, Alexander Zeier, Philippe Cudre-Mauroux, Samuel Madden, VLDB.
IMapReduce: A Distributed Computing Framework for Iterative Computation Yanfeng Zhang, Northeastern University, China Qixin Gao, Northeastern University,
Authors: Thilina Gunarathne, Tak-Lon Wu, Judy Qiu, Geoffrey Fox Publish: HPDC'10, June 20–25, 2010, Chicago, Illinois, USA ACM Speaker: Jia Bao Lin.
MSSG: A Framework for Massive-Scale Semantic Graphs Timothy D. R. Hartley, Umit Catalyurek, Fusun Ozguner, Andy Yoo, Scott Kohn, Keith Henderson Dept.
1 A Performance Study of Grid Workflow Engines Alexandru Iosup and Dick Epema PDS Group Delft University of Technology The Netherlands Corina Stratan Parallel.
Design Patterns for Efficient Graph Algorithms in MapReduce Jimmy Lin and Michael Schatz University of Maryland Tuesday, June 29, 2010 This work is licensed.
A Lightweight Infrastructure for Graph Analytics Donald Nguyen Andrew Lenharth and Keshav Pingali The University of Texas at Austin.
Big Data and Hadoop and DLRL Introduction to the DLRL Hadoop Cluster Sunshin Lee and Edward A. Fox DLRL, CS, Virginia Tech 21 May 2015 presentation for.
Google Distributed System and Hadoop Lakshmi Thyagarajan.
BiGraph BiGraph: Bipartite-oriented Distributed Graph Partitioning for Big Learning Jiaxin Shi Rong Chen, Jiaxin Shi, Binyu Zang, Haibing Guan Institute.
Design Patterns for Efficient Graph Algorithms in MapReduce Jimmy Lin and Michael Schatz University of Maryland MLG, January, 2014 Jaehwan Lee.
TPB Models Development Status Report Presentation to the Travel Forecasting Subcommittee Ron Milone National Capital Region Transportation Planning Board.
Venkatram Ramanathan 1. Motivation Evolution of Multi-Core Machines and the challenges Background: MapReduce and FREERIDE Co-clustering on FREERIDE Experimental.
11 If you were plowing a field, which would you rather use? Two oxen, or 1024 chickens? (Attributed to S. Cray) Abdullah Gharaibeh, Lauro Costa, Elizeu.
Bottlenecks: Automated Design Configuration Evaluation and Tune.
Performance Tuning on Multicore Systems for Feature Matching within Image Collections Xiaoxin Tang*, Steven Mills, David Eyers, Zhiyi Huang, Kai-Cheung.
Emalayan Vairavanathan
Ex-MATE: Data-Intensive Computing with Large Reduction Objects and Its Application to Graph Mining Wei Jiang and Gagan Agrawal.
Presented by CH.Anusha.  Apache Hadoop framework  HDFS and MapReduce  Hadoop distributed file system  JobTracker and TaskTracker  Apache Hadoop NextGen.
Presented By HaeJoon Lee Yanyan Shen, Beng Chin Ooi, Bogdan Marius Tudor National University of Singapore Wei Lu Renmin University Cang Chen Zhejiang University.
Storage in Big Data Systems
1 Fast Failure Recovery in Distributed Graph Processing Systems Yanyan Shen, Gang Chen, H.V. Jagadish, Wei Lu, Beng Chin Ooi, Bogdan Marius Tudor.
DisCo: Distributed Co-clustering with Map-Reduce S. Papadimitriou, J. Sun IBM T.J. Watson Research Center Speaker: 吳宏君 陳威遠 洪浩哲.
X-Stream: Edge-Centric Graph Processing using Streaming Partitions
Performance Issues in Parallelizing Data-Intensive applications on a Multi-core Cluster Vignesh Ravi and Gagan Agrawal
Introduction to Apache Hadoop Zibo Wang. Introduction  What is Apache Hadoop?  Apache Hadoop is a software framework which provides open source libraries.
f ACT s  Data intensive applications with Petabytes of data  Web pages billion web pages x 20KB = 400+ terabytes  One computer can read
CCGrid 2014 Improving I/O Throughput of Scientific Applications using Transparent Parallel Compression Tekin Bicer, Jian Yin and Gagan Agrawal Ohio State.
Porting Irregular Reductions on Heterogeneous CPU-GPU Configurations Xin Huo, Vignesh T. Ravi, Gagan Agrawal Department of Computer Science and Engineering.
Mining High Utility Itemset in Big Data
Shared Memory Parallelization of Decision Tree Construction Using a General Middleware Ruoming Jin Gagan Agrawal Department of Computer and Information.
Eneryg Efficiency for MapReduce Workloads: An Indepth Study Boliang Feng Renmin University of China Dec 19.
Evaluation of Agent Teamwork High Performance Distributed Computing Middleware. Solomon Lane Agent Teamwork Research Assistant October 2006 – March 2007.
Achieving Scalability, Performance and Availability on Linux with Oracle 9iR2-RAC Grant McAlister Senior Database Engineer Amazon.com Paper
Indexing HDFS Data in PDW: Splitting the data from the index VLDB2014 WSIC、Microsoft Calvin
Alastair Duncan STFC Pre Coffee talk STFC July 2014 The Trials and Tribulations and ultimate success of parallelisation using Hadoop within the SCAPE project.
Autonomic scheduling of tasks from data parallel patterns to CPU/GPU core mixes Published in: High Performance Computing and Simulation (HPCS), 2013 International.
MC 2 : Map Concurrency Characterization for MapReduce on the Cloud Mohammad Hammoud and Majd Sakr 1.
CCGrid 2014 Improving I/O Throughput of Scientific Applications using Transparent Parallel Compression Tekin Bicer, Jian Yin and Gagan Agrawal Ohio State.
Computing Scientometrics in Large-Scale Academic Search Engines with MapReduce Leonidas Akritidis Panayiotis Bozanis Department of Computer & Communication.
ICN Baseline Scenarios draft-pentikousis-icn-scenarios-04 K. Pentikousis (Ed.), B. Ohlman, D. Corujo, G. Boggia, G. Tyson, E. Davies, P. Mahadevan, S.
U N I V E R S I T Y O F S O U T H F L O R I D A Hadoop Alternative The Hadoop Alternative Larry Moore 1, Zach Fadika 2, Dr. Madhusudhan Govindaraju 2 1.
Is Your Graph Algorithm Eligible for Nondeterministic Execution? Zhiyuan Shao, Lin Hou, Yan Ai, Yu Zhang and Hai Jin Services Computing Technology and.
Hadoop IT Services Hadoop Users Forum CERN October 7 th,2015 CERN IT-D*
Scale up Vs. Scale out in Cloud Storage and Graph Processing Systems
DynamicMR: A Dynamic Slot Allocation Optimization Framework for MapReduce Clusters Nanyang Technological University Shanjiang Tang, Bu-Sung Lee, Bingsheng.
Performance and Energy Efficiency Evaluation of Big Data Systems Presented by Yingjie Shi Institute of Computing Technology, CAS
Data Structures and Algorithms in Parallel Computing Lecture 7.
1 If you were plowing a field, which would you rather use? Two oxen, or 1024 chickens? (Attributed to S. Cray)
What we know or see What’s actually there Wikipedia : In information technology, big data is a collection of data sets so large and complex that it.
Load Rebalancing for Distributed File Systems in Clouds.
Next Generation of Apache Hadoop MapReduce Owen
Distributed Process Discovery From Large Event Logs Sergio Hernández de Mesa {
LIOProf: Exposing Lustre File System Behavior for I/O Middleware
Implementation of Classifier Tool in Twister Magesh khanna Vadivelu Shivaraman Janakiraman.
EpiC: an Extensible and Scalable System for Processing Big Data Dawei Jiang, Gang Chen, Beng Chin Ooi, Kian Lee Tan, Sai Wu School of Computing, National.
Computer Science and Engineering Parallelizing Feature Mining Using FREERIDE Leonid Glimcher P. 1 ipdps’04 Scaling and Parallelizing a Scientific Feature.
Large Scale Semantic Data Integration and Analytics through Cloud: A Case Study in Bioinformatics Tat Thang Parallel and Distributed Computing Centre,
Chenning Xie+, Rong Chen+, Haibing Guan*, Binyu Zang+ and Haibo Chen+
Organizations Are Embracing New Opportunities
Scaling Spark on HPC Systems
GRAPHALYTICS A Big Data Benchmark for Graph-Processing Platforms
Hadoop Clusters Tess Fulkerson.
CS110: Discussion about Spark
Degree-aware Hybrid Graph Traversal on FPGA-HMC Platform
Many-Core Graph Workload Analysis
The Performance of Big Data Workloads in Cloud Datacenters
TensorFlow: A System for Large-Scale Machine Learning
Assoc. Prof. Marc FRÎNCU, PhD. Habil.
Presentation transcript:

LDBC-Benchmarking Graph-Processing Platforms: A Vision Benchmarking Graph-Processing Platforms: A Vision (A SPEC Research Group Process) Delft University of Technology, University of Amsterdam, VU University Amsterdam, Intel Corporation Yong Guo, Ana Lucia Varbanescu, Alexandru Iosup, Claudio Martella, Theodore L. Willke 1

The data deluge: large-scale graphs 2

Platform diversity Platform: the combined hardware, software, and programming system that is being used to complete a graph processing task. 3 YARN Trinity

What is the performance of these platforms? Graph500 Single application (BFS), Single class of synthetic datasets Few existing platform-centric comparative studies Prove the superiority of a given system, limited set of metrics GreenGraph500, GraphBench, XGDBench Representativeness, systems covered, metrics, … 4 Performance Metrics Graph Diversity Algorithm Diversity Our vision: a benchmarking suite for graph processing across all platforms

A Call to Arms Defining workloads Understanding the metrics, datasets, and algorithms used in practice: fill in our survey Evaluating and reporting on various platforms 5 Join us within the SPEC RG Cloud Working Group rg-cloud-working-group.html

Our Vision for Benchmarking Graph-Processing Platforms Methodological challenges 1. Evaluation process 2. Selection and design of performance metrics 3. Dataset selection 4. Algorithm coverage 6 Practical challenges 5. Scalability of evaluation, selection processes 6. Portability 7. Result reporting Y. Guo, A. L. Varbanescu, A. Iosup, C. Martella, T. L. Willke: Benchmarking graph-processing platforms: a vision. ICPE 2014:

Selection and Design of Performance Metrics for Graph Processing Raw processing power Execution time Actual computation time Edges/Vertices per second Resource utilization CPU, memory, network Scalability Horizontal vs. vertical Strong vs. weak Overhead Data ingestion time Overhead time 7 Challenge 2. Metric selection Elasticity (?)

Dataset Selection: Application Domains Number of vertices, edges, link density, size, directivity, etc. The Game Trace Archive Challenge 3. Dataset selection

Graph-Processing Algorithms Literature survey of of metrics, datasets, and algorithms 10 top research conferences: SIGMOD, VLDB, HPDC … Key word: graph processing, social network 2009–2013, 124 articles 9 ClassExamples% Graph StatisticsDiameter, PageRank16.1 Graph TraversalBFS, SSSP, DFS46.3 Connected ComponentReachability, BiCC13.4 Community DetectionClustering, Nearest Neighbor5.4 Graph EvolutionForest Fire Model, PAM4.0 OtherSampling, Partitioning14.8 Y. Guo, M. Biczak, A. L. Varbanescu, A. Iosup, C. Martella, and T. L. Willke. How Well do Graph-Processing Platforms Perform? An Empirical Performance Evaluation and Analysis,IPDPS’14.’14.

Platforms we have evaluated Distributed or non-distributed Graph-specific or generic 10 YARN Non-distributed (Graph-specific) Distributed (Generic) Distributed (Graph-specific) Challenge 6. Portability Y. Guo, M. Biczak, A. L. Varbanescu, A. Iosup, C. Martella, and T. L. Willke. How Well do Graph-Processing Platforms Perform? An Empirical Performance Evaluation and Analysis,IPDPS’14.’14.

BFS: results for all platforms, all graphs 11 Challenge 7. Result reporting Y. Guo, M. Biczak, A. L. Varbanescu, A. Iosup, C. Martella, and T. L. Willke. How Well do Graph-Processing Platforms Perform? An Empirical Performance Evaluation and Analysis,IPDPS’14.’14.

Key Findings From the Study of 6 Platforms Towards Benchmarking Graph-Processing Platforms Performance is function of (Dataset, Algorithm, Platform, Deployment) Previous performance studies may lead to tunnel vision Platforms have their own drawbacks (crashes, long execution time, tuning, etc.) Best-performing is not only low response time Ease-of-use of a platform is very important Some platforms can scale up reasonably with cluster size (horizontally) or number of cores (vertically) Strong vs weak scaling still a challenge—workload scaling tricky 12 Y. Guo, M. Biczak, A. L. Varbanescu, A. Iosup, C. Martella, and T. L. Willke. How Well do Graph-Processing Platforms Perform? An Empirical Performance Evaluation and Analysis,IPDPS’14.’14.

Thank you for your attention! Comments? Questions? Suggestions? Yong Guo Parallel and Distributed Systems Group Delft University of Technology 13 Join us, join the SPEC Research Group Fill in our survey on Big Data Use Cases Ask about other results

14

Experimental Setup DAS4: a multi-cluster Dutch grid/cloud Intel Xeon 2.4 GHz CPU (dual quad-core, 12 MB cache) Memory 24 GB 1 Gbit/s Ethernet network Size Most experiments take 20 working machines Up to 50 working machines HDFS used here as distributed file systems 15 Y. Guo, M. Biczak, A. L. Varbanescu, A. Iosup, C. Martella, T. L. Willke, How Well do Graph-Processing Platforms Perform? An Empirical Performance Evaluation and Analysis, IPDPS 2014

Design of a Survey In total, 33 questions minutes avg to finish the survey, at most 40 minutes Includes 4 major parts Institution profile—Who processes Big Data? System infrastructure—On what infrastructure? Data collection, storage, and process—Which workloads, metrics, etc.? What are the challenges? Framework stacks—Who do they rely on? 16 Survey on Big Data Use Cases

Our method: A benchmark suite Identifying the performance aspects and metrics of interest Defining and selecting representative datasets and algorithms Implementing, configuring, and executing the tests Analyzing the results 17 Challenge 1. Evaluation process

Selection of algorithms A1: General Statistics (STATS: # vertices and edges, LCC) Single step, low processing, decision-making A2: Breadth First Search (BFS) Iterative, low processing, building block A3: Connected Component (CONN) Iterative, medium processing, building block A4: Community Detection (CD) Iterative, medium or high processing, social network A5: Graph Evolution (EVO) Iterative (multi-level), high processing, prediction 18 Challenge 4. Algorithm selection

Scalability: BFS on Friendster Using more computing machines/cores can reduce execution time Tuning needed, e.g., for GraphLab, split large input file into number of chunks equal to the number of machines 19 Horizontal Vertical

The CPU utilization: computing node 20 YARN and Hadoop exhibit obvious volatility The CPU utilization of graph-specific platforms is lower Generic Graph-specific

Overhead: BFS on DotaLeague 21 The percentage of overhead time of generic platforms is smaller The percentage of overhead time is diverse across the platforms, algorithms, and graphs

Additional Overheads Data ingestion time Data ingestion Batch system: one ingestion, multiple processing Transactional system: one ingestion, one processing Data ingestion matters even for batch systems September 13, AmazonDotaLeagueFriendster HDFS1 second7 seconds5 minutes Neo4J4 hours6 daysn/a Guo, Biczak, Varbanescu, Iosup, Martella, Willke. How Well do Graph-Processing Platforms Perform? An Empirical Performance Evaluation and Analysis