Pwrake: An extensible parallel and distributed flexible workflow management tool Masahiro Tanaka and Osamu Tatebe University of Tsukuba PRAGMA18 3-4 March.

Slides:



Advertisements
Similar presentations
Gfarm v2 and CSF4 Osamu Tatebe University of Tsukuba Xiaohui Wei Jilin University SC08 PRAGMA Presentation at NCHC booth Nov 19,
Advertisements

National Institute of Advanced Industrial Science and Technology Ninf-G - Core GridRPC Infrastructure Software OGF19 Yoshio Tanaka (AIST) On behalf.
A Workflow Engine with Multi-Level Parallelism Supports Qifeng Huang and Yan Huang School of Computer Science Cardiff University
Copyright © 2008 SAS Institute Inc. All rights reserved. SAS and all other SAS Institute Inc. product or service names are registered trademarks or trademarks.
Parallel Processing with OpenMP
A Dynamic World, what can Grids do for Multi-Core computing? Daniel Goodman, Anne Trefethen and Douglas Creager
Setting up of condor scheduler on computing cluster Raman Sehgal NPD-BARC.
Recent Development of Gfarm File System Osamu Tatebe University of Tsukuba PRAGMA Institute on Implementation: Avian Flu Grid with Gfarm, CSF4 and OPAL.
26-Jun-15 Rake. rake and make A program can consist of many source code files This is always true in Rails! The files may need to be compiled in a certain.
A Source-to-Source OpenACC compiler for CUDA Akihiro Tabuchi †1 Masahiro Nakao †2 Mitsuhisa Sato †1 †1. Graduate School of Systems and Information Engineering,
Parallel Data Analysis from Multicore to Cloudy Grids Indiana University Geoffrey Fox, Xiaohong Qiu, Scott Beason, Seung-Hee.
U.S. Department of the Interior U.S. Geological Survey David V. Hill, Information Dynamics, Contractor to USGS/EROS 12/08/2011 Satellite Image Processing.
Optimization of application in virtual laboratory Optimization of application in virtual laboratory constructing workflows based on application sources.
Operating System Program 5 I/O System DMA Device Driver.
ADLB Update Recent and Current Adventures with the Asynchronous Dynamic Load Balancing Library Rusty Lusk Mathematics and Computer Science Division Argonne.
TeraGrid Gateway User Concept – Supporting Users V. E. Lynch, M. L. Chen, J. W. Cobb, J. A. Kohl, S. D. Miller, S. S. Vazhkudai Oak Ridge National Laboratory.
CLARIN tools for workflows Overview. Objective of this document  Determine which are the responsibilities of the different components of CLARIN workflows.
Authors: Weiwei Chen, Ewa Deelman 9th International Conference on Parallel Processing and Applied Mathmatics 1.
Fabien Viale 1 Matlab & Scilab Applications to Finance Fabien Viale, Denis Caromel, et al. OASIS Team INRIA -- CNRS - I3S.
2012/06/22 Contents  GPU (Graphic Processing Unit)  CUDA Programming  Target: Clustering with Kmeans  How to use.
WORK ON CLUSTER HYBRILIT E. Aleksandrov 1, D. Belyakov 1, M. Matveev 1, M. Vala 1,2 1 Joint Institute for nuclear research, LIT, Russia 2 Institute for.
ICOM 5995: Performance Instrumentation and Visualization for High Performance Computer Systems Lecture 7 October 16, 2002 Nayda G. Santiago.
Grid Data Management A network of computers forming prototype grids currently operate across Britain and the rest of the world, working on the data challenges.
LOGO Scheduling system for distributed MPD data processing Gertsenberger K. V. Joint Institute for Nuclear Research, Dubna.
Ex-MATE: Data-Intensive Computing with Large Reduction Objects and Its Application to Graph Mining Wei Jiang and Gagan Agrawal.
GXP Make Kenjiro Taura. GXP make What is it? –Another parallel & distributed make What is it useful for? –Running jobs with dependencies in parallel –Running.
Architectural Support for Fine-Grained Parallelism on Multi-core Architectures Sanjeev Kumar, Corporate Technology Group, Intel Corporation Christopher.
1 Dryad Distributed Data-Parallel Programs from Sequential Building Blocks Michael Isard, Mihai Budiu, Yuan Yu, Andrew Birrell, Dennis Fetterly of Microsoft.
DOMENICO TALIA (joint work with M. Cannataro, A. Congiusta, P. Trunfio) DEIS University of Calabria ITALY Grid-Based Data Mining and.
MARISSA: MApReduce Implementation for Streaming Science Applications 作者 : Fadika, Z. ; Hartog, J. ; Govindaraju, M. ; Ramakrishnan, L. ; Gunter, D. ; Canon,
1 Rake. 2 Automated Build Any non-trivial project needs facility to automate builds –Routine common tasks that need to be carried out several times a.
SALSASALSASALSASALSA Design Pattern for Scientific Applications in DryadLINQ CTP DataCloud-SC11 Hui Li Yang Ruan, Yuduo Zhou Judy Qiu, Geoffrey Fox.
Evaluation of Agent Teamwork High Performance Distributed Computing Middleware. Solomon Lane Agent Teamwork Research Assistant October 2006 – March 2007.
Condor Week 2005Optimizing Workflows on the Grid1 Optimizing workflow execution on the Grid Gaurang Mehta - Based on “Optimizing.
A Hierarchical MapReduce Framework Yuan Luo and Beth Plale School of Informatics and Computing, Indiana University Data To Insight Center, Indiana University.
LOGO PROOF system for parallel MPD event processing Gertsenberger K. V. Joint Institute for Nuclear Research, Dubna.
Data-Intensive Computing: From Clouds to GPUs Gagan Agrawal June 1,
Intermediate Condor: Workflows Rob Quick Open Science Grid Indiana University.
10/02/2012CS4230 CS4230 Parallel Programming Lecture 11: Breaking Dependences and Task Parallel Algorithms Mary Hall October 2,
Debugging parallel programs. Breakpoint debugging Probably the most widely familiar method of debugging programs is breakpoint debugging. In this method,
LOGO Development of the distributed computing system for the MPD at the NICA collider, analytical estimations Mathematical Modeling and Computational Physics.
Data-Intensive Computing: From Clouds to GPUs Gagan Agrawal December 3,
Department of Computer Science MapReduce for the Cell B. E. Architecture Marc de Kruijf University of Wisconsin−Madison Advised by Professor Sankaralingam.
David Adams ATLAS DIAL: Distributed Interactive Analysis of Large datasets David Adams BNL August 5, 2002 BNL OMEGA talk.
CUDA Basics. Overview What is CUDA? Data Parallelism Host-Device model Thread execution Matrix-multiplication.
RE-PAGE: Domain-Specific REplication and PArallel Processing of GEnomic Applications 1 Mucahid Kutlu Gagan Agrawal Department of Computer Science and Engineering.
LQCD Workflow Project L. Piccoli October 02, 2006.
PDAC-10 Middleware Solutions for Data- Intensive (Scientific) Computing on Clouds Gagan Agrawal Ohio State University (Joint Work with Tekin Bicer, David.
A N I N - MEMORY F RAMEWORK FOR E XTENDED M AP R EDUCE 2011 Third IEEE International Conference on Coud Computing Technology and Science.
Super Scaling PROOF to very large clusters Maarten Ballintijn, Kris Gulbrandsen, Gunther Roland / MIT Rene Brun, Fons Rademakers / CERN Philippe Canal.
AUTO-GC: Automatic Translation of Data Mining Applications to GPU Clusters Wenjing Ma Gagan Agrawal The Ohio State University.
GPU Programming Contest. Contents Target: Clustering with Kmeans How to use toolkit1.0 Towards the fastest program.
STAR Scheduler Gabriele Carcassi STAR Collaboration.
Holding slide prior to starting show. Processing Scientific Applications in the JINI-Based OGSA-Compliant Grid Yan Huang.
1 An unattended, fault-tolerant approach for the execution of distributed applications Manuel Rodríguez-Pascual, Rafael Mayo-García CIEMAT Madrid, Spain.
High-throughput parallel pipelined data processing system for remote Earth sensing big data in the clouds Высокопроизводительная параллельно-конвейерная.
Apache Ignite Compute Grid Research Corey Pentasuglia.
SparkBWA: Speeding Up the Alignment of High-Throughput DNA Sequencing Data - Aditi Thuse.
How to download, configure and run a mapReduce program In a cloudera VM Presented By: Mehakdeep Singh Amrit Singh Chaggar Ranjodh Singh.
Conception of parallel algorithms
A task-based implementation for GeantV
Integration of Singularity With Makeflow
Year 2 Updates.
Intel® Parallel Studio and Advisor
Rake 4-Dec-18.
Wide Area Workload Management Work Package DATAGRID project
Overview of Workflows: Why Use Them?
rvGAHP – Push-Based Job Submission Using Reverse SSH Connections
Implementation of a De-blocking Filter and Optimization in PLX
L. Glimcher, R. Jin, G. Agrawal Presented by: Leo Glimcher
Presentation transcript:

Pwrake: An extensible parallel and distributed flexible workflow management tool Masahiro Tanaka and Osamu Tatebe University of Tsukuba PRAGMA March 2010

Workflow Systems PRAGMA March 2010 Visual Workflow Creation Is EASY but has many LIMITATIONS!

Montage Astrophysics workflow Flexible Task dependency Loop & Conditions Parallel & Remote execution availability from single host to Cluster & Grid PRAGMA March 2010

Pwrake = Rake + Parallel Workflow extension Rake – Ruby version of make – Much powerful description than Makefile – Just specify input files and output files, that it! Pwrake – Parallel workflow extension – If execution fails, pwrake again – Extensible Mounting Gfarm file system on the remote node Gfarm file affinity scheduling PRAGMA March 2010

Rake syntax = Ruby syntax file “prog” => [“a.o”, “b.o”] do sh “cc –o prog a.o b.o” end Ruby method defined in Rake Ruby method defined in Rake Ruby code block enclosed by do … end or {…}. not executed on task definition, but passed to the file method and executed as a task action Ruby code block enclosed by do … end or {…}. not executed on task definition, but passed to the file method and executed as a task action Key-value argument to file method task_name => prerequisites Key-value argument to file method task_name => prerequisites PRAGMA March 2010

Pwrake implementation PwMultitask class Prerequisite Tasks Prerequisite Tasks SSH connection Task1 Task2 Task3 Task Queue Thread Queue for remote executions … worker thread1 worker thread2 worker thread3 remote host1 remote host2 remote host3 enqueue dequeue able to extend for affinity scheduling able to extend for affinity scheduling PRAGMA March 2010

Benefit of Pwrake Rakefile is evaluated as a Ruby script. With Ruby’s scripting power, ANY TASK and DEPENDENCY can be defined. PRAGMA March 2010

Example of Rake (1) File Dependency: – Not suffix-based dependency How do you define these tasks? A00A01A02A03 B00B01B02 … … PRAGMA March 2010

Comparison of task definition Make: B00: A00 A01 prog A00 A01 > B00 B01: A01 A02 prog A01 A02 > B01 B02: A02 A03 prog A02 A03 > B02 …… Rake: for i in "00".."10" file “B#{i}" => [“A#{i}",“A#{i.succ}"] {|t| sh "prog #{t.prerequisites.join(' ')} > #{t.name}" } end PRAGMA March 2010

Example of Rake (2) File dependency is given as a list written in a file: $ cat depend_list dif_1_2.fits image1.fits image2.fits dif_1_3.fits image1.fits image3.fits dif_2_3.fits image2.fits image3.fits... How do you write this? image1 dif_1_2 … … image2image3 dif_1_3dif_2_3 PRAGMA March 2010

Dependency is given as a file list Make: – Needs other script to convert file list to Makefile Rake: open("depend_list") { |f| f.readlines.each { |line| name, file1, file2 = line.split file name => [file1,file2] do |t| sh “prog #{t.prerequisites.join(' ')} #{t.name}" end } PRAGMA March 2010

Performance measurement Workflow: – Montage a tool to combine astronomical images. Input data: – 3.3 GB (1,580 files) of 2MASS All sky survey Used cluster: sitecorenodesmemory Univ of Tsukubaquad84GB AISTdual82GB PRAGMA March 2010

Performance of Montage workflow 1 node 4 cores 2 nodes 8 cores 4 nodes 16 cores 8 nodes 32 cores 1-site 2 sites 16 nodes 48 cores PRAGMA March 2010

Performance of Montage workflow 1 node 4 cores 2 nodes 8 cores 4 nodes 16 cores 8 nodes 32 cores 1-site 2 sites 16 nodes 48 cores NFS PRAGMA March 2010

Performance of Montage workflow 1 node 4 cores 2 nodes 8 cores 4 nodes 16 cores 8 nodes 32 cores 1-site 2 sites 16 nodes 48 cores Gfarm without affinity scheduling, initial files are not distributed Gfarm without affinity scheduling, initial files are not distributed PRAGMA March 2010

Performance of Montage workflow 1 node 4 cores 2 nodes 8 cores 4 nodes 16 cores 8 nodes 32 cores 1-site 2 sites 16 nodes 48 cores Gfarm with affinity scheduling, initial files are not distributed Gfarm with affinity scheduling, initial files are not distributed 14% speedup 14% speedup PRAGMA March 2010

Performance of Montage workflow 1 node 4 cores 2 nodes 8 cores 4 nodes 16 cores 8 nodes 32 cores 1-site 2 sites 16 nodes 48 cores Gfarm with affinity scheduling, initial files are distributed Gfarm with affinity scheduling, initial files are distributed 20% speedup 20% speedup PRAGMA March 2010

Performance of Montage workflow 1 node 4 cores 2 nodes 8 cores 4 nodes 16 cores 8 nodes 32 cores 1-site 2 sites 16 nodes 48 cores Gfarm with affinity scheduling, initial files are distributed Gfarm with affinity scheduling, initial files are distributed PRAGMA March 2010

Performance of Montage workflow 1 node 4 cores 2 nodes 8 cores 4 nodes 16 cores 8 nodes 32 cores 1-site 2 sites 16 nodes 48 cores Gfarm with affinity scheduling, initial files are optimally allocated Gfarm with affinity scheduling, initial files are optimally allocated PRAGMA March 2010

Conclusion Pwrake, a parallel and distributed flexible workflow management tool, is proposed. Pwrake is extensible, and has flexible and powerful workflow language to describe scientific workflow. We demonstrate a practical e-Science data-intensive workflow in Astronomical data analysis on Gfarm file system in wide area environment. Extending a scheduling algorithm to be aware of file locations, 20% of speed up was observed using 8 nodes (32 cores) in a PC cluster. Using two PC clusters located at different sites, the file location aware scheduling and appropriate input data placement showed scalable speedup. PRAGMA March 2010