Building Scalable Scientific Applications with Work Queue Douglas Thain and Dinesh Rajan University of Notre Dame Applied Cyber Infrastructure Concepts.

Slides:



Advertisements
Similar presentations
1 Real-World Barriers to Scaling Up Scientific Applications Douglas Thain University of Notre Dame Trends in HPDC Workshop Vrije University, March 2012.
Advertisements

Lobster: Personalized Opportunistic Computing for CMS at Large Scale Douglas Thain (on behalf of the Lobster team) University of Notre Dame CVMFS Workshop,
BXGrid: A Data Repository and Computing Grid for Biometrics Research Hoang Bui University of Notre Dame 1.
Overview of Wisconsin Campus Grid Dan Bradley Center for High-Throughput Computing.
Experience with Adopting Clouds at Notre Dame Douglas Thain University of Notre Dame IEEE CloudCom, November 2010.
DV/dt : Accelerating the Rate of Progress Towards Extreme Scale Collaborative Science Miron Livny (UW), Ewa Deelman (USC/ISI), Douglas Thain (ND), Frank.
Introduction to Scalable Programming using Makeflow and Work Queue Dinesh Rajan and Mike Albrecht University of Notre Dame October 24 and November 7, 2012.
Scaling Up Without Blowing Up Douglas Thain University of Notre Dame.
RCAC Research Computing Presents: DiaGird Overview Tuesday, September 24, 2013.
1 Condor Compatible Tools for Data Intensive Computing Douglas Thain University of Notre Dame Condor Week 2011.
1 High Throughput Scientific Computing with Condor: Computer Science Challenges in Large Scale Parallelism Douglas Thain University of Notre Dame UAB 27.
1 Opportunities and Dangers in Large Scale Data Intensive Computing Douglas Thain University of Notre Dame Large Scale Data Mining Workshop at SIGKDD August.
1 Scaling Up Data Intensive Science with Application Frameworks Douglas Thain University of Notre Dame Michigan State University September 2011.
1 Models and Frameworks for Data Intensive Cloud Computing Douglas Thain University of Notre Dame IDGA Cloud Computing 8 February 2011.
1 Scaling Up Data Intensive Science to Campus Grids Douglas Thain Clemson University 25 Septmber 2009.
Work Queue: A Scalable Master/Worker Framework Peter Bui June 29, 2010.
Getting Beyond the Filesystem: New Models for Data Intensive Scientific Computing Douglas Thain University of Notre Dame HEC FSIO Workshop 6 August 2009.
Introduction to Makeflow Li Yu University of Notre Dame 1.
Cloud based Dynamic workflow with QOS for Mass Spectrometry Data Analysis Thesis Defense: Ashish Nagavaram Graduate student Computer Science and Engineering.
Building Scalable Elastic Applications using Makeflow Dinesh Rajan and Douglas Thain University of Notre Dame Tutorial at CCGrid, May Delft, Netherlands.
Building Scalable Scientific Applications using Makeflow Dinesh Rajan and Peter Sempolinski University of Notre Dame.
Building Scalable Applications on the Cloud with Makeflow and Work Queue Douglas Thain and Patrick Donnelly University of Notre Dame Science Cloud Summer.
Introduction to Makeflow and Work Queue CSE – Cloud Computing – Fall 2014 Prof. Douglas Thain.
CONDOR DAGMan and Pegasus Selim Kalayci Florida International University 07/28/2009 Note: Slides are compiled from various TeraGrid Documentations.
Portable Resource Management for Data Intensive Workflows Douglas Thain University of Notre Dame.
Apache Airavata GSOC Knowledge and Expertise Computational Resources Scientific Instruments Algorithms and Models Archived Data and Metadata Advanced.
Design of an Active Storage Cluster File System for DAG Workflows Patrick Donnelly and Douglas Thain University of Notre Dame 2013 November 18 th DISCS-2013.
Elastic Applications in the Cloud Dinesh Rajan University of Notre Dame CCL Workshop, June 2012.
Massively Parallel Ensemble Methods Using Work Queue Badi’ Abdul-Wahid Department of Computer Science University of Notre Dame CCL Workshop 2012.
Toward a Common Model for Highly Concurrent Applications Douglas Thain University of Notre Dame MTAGS Workshop 17 November 2013.
Introduction to Work Queue Applications CSE – Cloud Computing – Fall 2014 Prof. Douglas Thain.
Building Scalable Scientific Applications with Makeflow Douglas Thain and Dinesh Rajan University of Notre Dame Applied Cyber Infrastructure Concepts University.
Building Scalable Scientific Applications using Makeflow Dinesh Rajan and Douglas Thain University of Notre Dame.
The Cooperative Computing Lab  We collaborate with people who have large scale computing problems in science, engineering, and other fields.  We operate.
Introduction to Scalable Programming using Work Queue Dinesh Rajan and Ben Tovar University of Notre Dame October 10, 2013.
Distributed Framework for Automatic Facial Mark Detection Graduate Operating Systems-CSE60641 Nisha Srinivas and Tao Xu Department of Computer Science.
1 Computational Abstractions: Strategies for Scaling Up Applications Douglas Thain University of Notre Dame Institute for Computational Economics University.
Introduction to Work Queue Applications Applied Cyberinfrastructure Concepts Course University of Arizona 2 October 2014 Douglas Thain and Nicholas Hazekamp.
Enabling Grids for E-sciencE EGEE-III INFSO-RI Using DIANE for astrophysics applications Ladislav Hluchy, Viet Tran Institute of Informatics Slovak.
Grid Computing at Yahoo! Sameer Paranjpye Mahadev Konar Yahoo!
Building Scalable Scientific Applications with Work Queue Douglas Thain and Dinesh Rajan University of Notre Dame Applied Cyber Infrastructure Concepts.
Review of Condor,SGE,LSF,PBS
Introduction to Makeflow and Work Queue Prof. Douglas Thain, University of Notre Dame
Analyzing LHC Data on 10K Cores with Lobster and Work Queue Douglas Thain (on behalf of the Lobster Team)
Introduction to Scalable Programming using Work Queue Dinesh Rajan and Mike Albrecht University of Notre Dame October 24 and November 7, 2012.
A N I N - MEMORY F RAMEWORK FOR E XTENDED M AP R EDUCE 2011 Third IEEE International Conference on Coud Computing Technology and Science.
Building Scalable Elastic Applications using Work Queue Dinesh Rajan and Douglas Thain University of Notre Dame Tutorial at CCGrid, May Delft,
Demonstration of Scalable Scientific Applications Peter Sempolinski and Dinesh Rajan University of Notre Dame.
Massively Parallel Molecular Dynamics Using Adaptive Weighted Ensemble Badi’ Abdul-Wahid PI: Jesús A. Izaguirre CCL Workshop 2013.
Geant4 GRID production Sangwan Kim, Vu Trong Hieu, AD At KISTI.
Introduction to Makeflow and Work Queue Nicholas Hazekamp and Ben Tovar University of Notre Dame XSEDE 15.
1 An unattended, fault-tolerant approach for the execution of distributed applications Manuel Rodríguez-Pascual, Rafael Mayo-García CIEMAT Madrid, Spain.
Instrumenting Badi Abdul-Wahid, RJ Nowling CSE Operating Systems Professor Striegel.
Introduction to Makeflow and Work Queue
The Accelerated Weighted Ensemble
Integrated genome analysis using
Scaling Up Scientific Workflows with Makeflow
Introduction to Makeflow and Work Queue
Integration of Singularity With Makeflow
Introduction to Makeflow and Work Queue
Haiyan Meng and Douglas Thain
Introduction to Makeflow and Work Queue
Weaving Abstractions into Workflows
Introduction to Makeflow and Work Queue
CS110: Discussion about Spark
Introduction to Makeflow and Work Queue with Containers
BXGrid: A Data Repository and Computing Grid for Biometrics Research
What’s New in Work Queue
Creating Custom Work Queue Applications
Condor-G Making Condor Grid Enabled
Presentation transcript:

Building Scalable Scientific Applications with Work Queue Douglas Thain and Dinesh Rajan University of Notre Dame Applied Cyber Infrastructure Concepts University of Arizona, September 13, 2012

Go to: Click “Lecture and Tutorial for Applied CI Concepts Class”

Recap: Makeflow

4 An Old Idea: Makefiles part1 part2 part3: input.data split.py./split.py input.data out1: part1 mysim.exe./mysim.exe part1 > out1 out2: part2 mysim.exe./mysim.exe part2 > out2 out3: part3 mysim.exe./mysim.exe part3 > out3 result: out1 out2 out3 join.py./join.py out1 out2 out3 > result

5 Makeflow = Make + Workflow Makeflow LocalCondorTorque Work Queue Provides portability across batch systems. Enable parallelism (but not too much!) Fault tolerance at multiple scales. Data and resource management.

Makeflow + Work Queue

Private Cluster Campus Condor Pool Public Cloud Provider FutureGrid Torque Cluster Makefile Makeflow Local Files and Programs Makeflow + Batch System makeflow –T torque makeflow –T condor ???

FutureGrid Torque Cluster Campus Condor Pool Public Cloud Provider Private Cluster Makefile Makeflow Local Files and Programs Makeflow + Work Queue W W W ssh WW W W torque_submit_workers W W W condor_submit_workers W W W Thousands of Workers in a Personal Cloud submit tasks

Makeflow and Work Queue with Project Names Start Makeflow with a project name: % makeflow –T wq –p 0 –a –N arizona-class sims.mf Listening for workers on port XYZ… Start one worker: % work_queue_worker -a -N arizona-class Start many workers: % torque_submit_workers –a –N arizona-class 5

Advantages of Using Work Queue Harness multiple resources simultaneously. Hold on to cluster nodes to execute multiple tasks rapidly. (ms/task instead of min/task) Scale resources up and down as needed. Better management of data, with local caching for data intensive tasks. Matching of tasks to nodes with data.

Makeflow Great for static workflows! - Tasks/dependencies are known beforehand - DAGs Great for file-based workflows! - Input: files, Output: files

What if my workflow is dynamic? Write a program using the Work Queue API

13 Work Queue API use work_queue; queue = work_queue_create(); while( not done ) { while (more work ready) { task = work_queue_task_create(); // add some details to the task work_queue_submit(queue, task); } task = work_queue_wait(queue); // process the completed task }

14 worker sim in.txtout.txt put sim.exe put in.txt exec sim.exe out.txt get out.txt 1000s of workers dispatched to clusters, clouds, & grids Work Queue System Work Queue Library Work Queue Program C / Python / Perl cache recently used files

Run One Task in Python from work_queue import * queue = WorkQueue( port = 0 ) queue.specify_name( “myproject” ); task = Task(“sim.exe –p 50 in.dat >out.txt”) ### Missing: Specify files needed by the task. queue.submit( task ) While not queue.empty(): task = queue.wait(60)

Run One Task in Perl use work_queue; $queue = work_queue_create( 0 ); work_queue_specify_name( “myproject” ); $task = work_queue_task_create(“sim.exe –p 50 in.dat >out.txt”); ### Missing: Specify files needed by the task. work_queue_submit( $queue, $task ); while(!work_queue_empty($queue)) { $task = work_queue_wait( $queue, 60 ); if($task) work_queue_task_delete( $task ); }

Run One Task in C #include “work_queue.h” struct work_queue *queue; struct work_queue_task *task; queue = work_queue_create( 0 ); work_queue_specify_name( “myproject” ); task = work_queue_task_create(“sim.exe –p 50 in.dat >out.txt”); /// Missing: Specify files needed by the task. work_queue_submit( queue, task ); while(!work_queue_empty(queue)) { task = work_queue_wait( queue, 60 ); if(task) work_queue_task_delete( task ); }

Python: Specify Files for a Task task.specify_file( “in.dat”, ”in.dat”, WORK_QUEUE_INPUT, cache = False ) task.specify_file( “calib.dat”, ”calib.dat”, WORK_QUEUE_INPUT, cache = False ) task.specify_file( “out.txt”, ”out.txt”, WORK_QUEUE_OUTPUT, cache = False ) task.specify_file( “sim.exe”, ”sim.exe”, WORK_QUEUE_INPUT, cache = True ) sim.exe in.dat calib.dat out.txt sim.exe in.dat –p 50 > out.txt

Perl: Specify Files for a Task sim.exe in.dat calib.dat out.txt sim.exe in.dat –p 50 > out.txt work_queue_task_specify_file( $task,“in.dat”,”in.dat”, $WORK_QUEUE_INPUT, $WORK_QUEUE_NOCACHE ); work_queue_task_specify_file( $task,“calib.dat”,”calib.dat”, $WORK_QUEUE_INPUT, $WORK_QUEUE_NOCACHE ); work_queue_task_specify_file( $task,“out.txt”,”out.txt”, $WORK_QUEUE_OUTPUT, $WORK_QUEUE_NOCACHE ); work_queue_task_specify_file( $task,“sim.exe”,”sim.exe”, $WORK_QUEUE_INPUT, $WORK_QUEUE_CACHE );

C: Specify Files for a Task work_queue_task_specify_file( task,“in.dat”,”in.dat”, WORK_QUEUE_INPUT, WORK_QUEUE_NOCACHE ); work_queue_task_specify_file( task,“calib.dat”,”calib.dat”, WORK_QUEUE_INPUT, WORK_QUEUE_CACHE ); work_queue_task_specify_file( task,“out.txt”,”out.txt”, WORK_QUEUE_OUTPUT, WORK_QUEUE_NOCACHE ); work_queue_task_specify_file( task,“sim.exe”,”sim.exe”, WORK_QUEUE_INPUT, WORK_QUEUE_CACHE ); sim.exe in.dat calib.dat out.txt sim.exe in.dat –p 50 > out.txt

You must state all the files needed by the command.

Start workers for your Work Queue program Start one local worker: work_queue_worker -a -N myproject Submit workers to Torque: torque_submit_workers -a –N myproject 25 Submit workers to Condor: condor_submit_workers -a –N myproject 25 Submit workers to SGE: sge_submit_workers -a –N myproject 25

Sample Applications of Work Queue

Genome Assembly Christopher Moretti, Andrew Thrasher, Li Yu, Michael Olson, Scott Emrich, and Douglas Thain, A Framework for Scalable Genome Assembly on Clusters, Clouds, and Grids, IEEE Transactions on Parallel and Distributed Systems, 2012 Using WQ, we could assemble a human genome in 2.5 hours on a collection of clusters, clouds, and grids with a speedup of 952X. SAND filter master SAND align master Celera Consensus W W W W W W W Sequence Data Modified Celera Assembler

Replica Exchange T=10KT=20K T=30KT=40K Replica Exchange Work Queue Simplified Algorithm: Submit N short simulations at different temps. Wait for all to complete. Select two simulations to swap. Continue all of the simulations. Dinesh Rajan, Anthony Canino, Jesus A Izaguirre, and Douglas Thain, Converting A High Performance Application to an Elastic Cloud Application, Cloud Com 2011.

Adaptive Weighted Ensemble 26 Proteins fold into a number of distinctive states, each of which affects its function in the organism. How common is each state? How does the protein transition between states? How common are those transitions?

27 Simplified Algorithm: Simplified Algorithm: N short simulations in various states. –Submit N short simulations in various states. for them to finish. –Wait for them to finish. –When done, record all state transitions. –If too many are in one state, redistribute them. –Has enough data been collected? –Yes : Stop –No : Go back to the beginning. AWE Using Work Queue

AWE on Clusters, Clouds, and Grids 28

New Pathway Found! 29 Joint work with computational biologists: Folding Proteins at 500 ns/hour with Work Queue, eScience 2012

Private Cluster Campus Condor Pool Public Cloud Provider Shared SGE Cluster Elastic Application Stack W W WW W W W W W Work Queue Library All-PairsWavefrontMakeflow Custom Apps Thousands of Workers in a Personal Cloud W

Next Steps: Tutorial: Basic examples of how to write and use Work Queue on Future Grid. Homework: Problems to work on this week.

Go to: Click “Lecture and Tutorial for Applied CI Concepts Class”