28 April, 2005ISGC 2005, Taiwan The Efficient Handling of BLAST Applications on the GRID Hurng-Chun Lee 1 and Jakub Moscicki 2 1 Academia Sinica Computing.

Slides:



Advertisements
Similar presentations
Enabling Cost-Effective Resource Leases with Virtual Machines Borja Sotomayor University of Chicago Ian Foster Argonne National Laboratory/
Advertisements

A Workflow Engine with Multi-Level Parallelism Supports Qifeng Huang and Yan Huang School of Computer Science Cardiff University
Maria Grazia Pia Simulation in a Distributed Computing Environment Simulation in a Distributed Computing Environment S. Guatelli 1, A. Mantero 1, P. Mendez.
CERN LCG Overview & Scaling challenges David Smith For LCG Deployment Group CERN HEPiX 2003, Vancouver.
WHAT IS AN OPERATING SYSTEM? An interface between users and hardware - an environment "architecture ” Allows convenient usage; hides the tedious stuff.
Parasol Architecture A mild case of scary asynchronous system stuff.
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
490dp Synchronous vs. Asynchronous Invocation Robert Grimm.
1: Operating Systems Overview
TurboBLAST: A Parallel Implementation of BLAST Built on the TurboHub Bin Gan CMSC 838 Presentation.
Scripting Languages For Virtual Worlds. Outline Necessary Features Classes, Prototypes, and Mixins Static vs. Dynamic Typing Concurrency Versioning Distribution.
Workload Management Massimo Sgaravatto INFN Padova.
07/14/08. 2 Points Introduction. Cluster and Supercomputers. Cluster Types and Advantages. Our Cluster. Cluster Performance. Cluster Computer for Basic.
Client/Server Grid applications to manage complex workflows Filippo Spiga* on behalf of CRAB development team * INFN Milano Bicocca (IT)
LHC Experiment Dashboard Main areas covered by the Experiment Dashboard: Data processing monitoring (job monitoring) Data transfer monitoring Site/service.
ATIF MEHMOOD MALIK KASHIF SIDDIQUE Improving dependability of Cloud Computing with Fault Tolerance and High Availability.
Ch 4. The Evolution of Analytic Scalability
DIANE Overview Germán Carrera, Alfredo Solano (CNB/CSIC) EMBRACE COURSE Monday 19th of February to Friday 23th. CNB-CSIC Madrid.
Applying Twister to Scientific Applications CloudCom 2010 Indianapolis, Indiana, USA Nov 30 – Dec 3, 2010.
PROOF: the Parallel ROOT Facility Scheduling and Load-balancing ACAT 2007 Jan Iwaszkiewicz ¹ ² Gerardo Ganis ¹ Fons Rademakers ¹ ¹ CERN PH/SFT ² University.
Tools and software process for the FLP prototype B. von Haller 9. June 2015 CERN.
KARMA with ProActive Parallel Suite 12/01/2009 Air France, Sophia Antipolis Solutions and Services for Accelerating your Applications.
LOGO OPERATING SYSTEM Dalia AL-Dabbagh
Operating System Review September 10, 2012Introduction to Computer Security ©2004 Matt Bishop Slide #1-1.
DISTRIBUTED COMPUTING
Cluster Reliability Project ISIS Vanderbilt University.
Flexibility and user-friendliness of grid portals: the PROGRESS approach Michal Kosiedowski
March 3rd, 2006 Chen Peng, Lilly System Biology1 Cluster and SGE.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks EGEE Application Case Study: Distributed.
Engr. M. Fahad Khan Lecturer Software Engineering Department University Of Engineering & Technology Taxila.
DIANE Project CHEP 03 DIANE Distributed Analysis Environment for semi- interactive simulation and analysis in Physics Jakub T. Moscicki,
Evaluation of Agent Teamwork High Performance Distributed Computing Middleware. Solomon Lane Agent Teamwork Research Assistant October 2006 – March 2007.
Efficient Data Accesses for Parallel Sequence Searches Heshan Lin (NCSU) Xiaosong Ma (NCSU & ORNL) Praveen Chandramohan (ORNL) Al Geist (ORNL) Nagiza Samatova.
Stuart Wakefield Imperial College London Evolution of BOSS, a tool for job submission and tracking W. Bacchi, G. Codispoti, C. Grandi, INFN Bologna D.
Enabling Grids for E-sciencE EGEE-III INFSO-RI Using DIANE for astrophysics applications Ladislav Hluchy, Viet Tran Institute of Informatics Slovak.
1 Large-Scale Profile-HMM on the Grid Laurent Falquet Swiss Institute of Bioinformatics CH-1015 Lausanne, Switzerland Borrowed from Heinz Stockinger June.
S. Guatelli, A. Mantero, J. Moscicki, M. G. Pia Geant4 medical simulations in a distributed computing environment 4th Workshop on Geant4 Bio-medical Developments.
Running BLAST on the cluster system over the Pacific Rim.
CASTOR evolution Presentation to HEPiX 2003, Vancouver 20/10/2003 Jean-Damien Durand, CERN-IT.
CERN IT Department CH-1211 Geneva 23 Switzerland t CF Computing Facilities Agile Infrastructure Monitoring CERN IT/CF.
CIS250 OPERATING SYSTEMS Chapter One Introduction.
Development of e-Science Application Portal on GAP WeiLong Ueng Academia Sinica Grid Computing
DOE Network PI Meeting 2005 Runtime Data Management for Data-Intensive Scientific Applications Xiaosong Ma NC State University Joint Faculty: Oak Ridge.
AFS/OSD Project R.Belloni, L.Giammarino, A.Maslennikov, G.Palumbo, H.Reuter, R.Toebbicke.
Susanna Guatelli Geant4 in a Distributed Computing Environment S. Guatelli 1, P. Mendez Lorenzo 2, J. Moscicki 2, M.G. Pia 1 1. INFN Genova, Italy, 2.
CERN Certification & Testing LCG Certification & Testing Team (C&T Team) Marco Serra - CERN / INFN Zdenek Sekera - CERN.
Distributed parallel processing analysis framework for Belle II and Hyper Suprime-Cam MINEO Sogo (Univ. Tokyo), ITOH Ryosuke, KATAYAMA Nobu (KEK), LEE.
Breaking the frontiers of the Grid R. Graciani EGI TF 2012.
Zach Miller Computer Sciences Department University of Wisconsin-Madison Supporting the Computation Needs.
Geant4 GRID production Sangwan Kim, Vu Trong Hieu, AD At KISTI.
Grid Services for Digital Archive Tao-Sheng Chen Academia Sinica Computing Centre
DIRAC: Workload Management System Garonne Vincent, Tsaregorodtsev Andrei, Centre de Physique des Particules de Marseille Stockes-rees Ian, University of.
Introduction to Distributed Platforms
(on behalf of the POOL team)
Design rationale and status of the org.glite.overlay component
Joseph JaJa, Mike Smorul, and Sangchul Song
Simulation in a Distributed Computing Environment
Sharing Memory: A Kernel Approach AA meeting, March ‘09 High Performance Computing for High Energy Physics Vincenzo Innocente July 20, 2018 V.I. --
LHCb Computing Model and Data Handling Angelo Carbone 5° workshop italiano sulla fisica p-p ad LHC 31st January 2008.
WP1 activity, achievements and plans
New Mexico State University
Applying Twister to Scientific Applications
SDM workshop Strawman report History and Progress and Goal.
Ch 4. The Evolution of Analytic Scalability
Support for ”interactive batch”
Simulation in a Distributed Computing Environment
Future EU Grid Projects
Parallel System for BLAST
Simulation in a Distributed Computing Environment
Database System Architectures
Presentation transcript:

28 April, 2005ISGC 2005, Taiwan The Efficient Handling of BLAST Applications on the GRID Hurng-Chun Lee 1 and Jakub Moscicki 2 1 Academia Sinica Computing Centre, Taiwan 2 CERN IT-GD-ED, Switzerland

28 April, 2005ISGC 2005, Taiwan Outline The consideration of distributing BLAST jobs The master-worker computing model of BLAST –mpiBLAST The Gridified BLAST –mpiBLAST-g2 vs. DIANE-BLAST Summary

28 April, 2005ISGC 2005, Taiwan The considerations of distributing BLAST jobs BLAST has been widely and routinely used for sequence analysis The essential component in most of bioinformatics and life science applications Problem Complexity ~ O(S q xS d ) –S q : The query size –S d : The database size In most cases, S d >> S q –e.g. S q ~ O(MB), S d ~ O(GB) –The cost of moving query is lower Database management, storage and sharing issues –Replication, Archive –Privacy, Security Other perspective for service providing –scalability, robustness

28 April, 2005ISGC 2005, Taiwan The master-worker model of BLAST Database splitting is the easiest way to distribute BLAST jobs Fragmented databases for avoiding the memory swapping Each sub task can be 100% independent Each worker requests the tasks from master (pull model) and runs the normal BLAST search The individual result can be easily merged by master process Report generation (BioSeq fetching) Multi-query blast search can be easily split to multiple independent single-query blast search by a trivial script –Master-worker model can also be applied in each single-query search Database Master workers DB Fragments Task list Job requesting Result merging formatdb blast search BioSeq fetching

28 April, 2005ISGC 2005, Taiwan mpiBLAST LANL, US The MPI implementation of BLAST master-worker model Advantages –High throughput –Load Balancing Running in local cluster –Performance and Problem size still be limited by local computing power –Simultaneous I/O to centralized database causes the performance bottleneck –Database sharing is still difficult

28 April, 2005ISGC 2005, Taiwan mpiBLAST-g2 ASCC, Taiwan and PRAGMA A GT2-enabled parallel BLAST runs on Grid –GT2 GASSCOPY API –MPICH-g2 The enhancement from mpiBLAST by ASCC Performing cross cluster scheme of job execution Performing remote database sharing Help Tools for –database replication –automatic resource specification and job submission (with static resource table) –multi-query job splitting and result merging Close link with mpiBLAST development team –The new patches of mpiBLAST can be quickly applied in mpiBLAST-g2

28 April, 2005ISGC 2005, Taiwan SC2004 mpiBLAST-g2 demonstration KISTI

28 April, 2005ISGC 2005, Taiwan mpiBLAST-g2 current deployment -- From PRAGMA GOC

28 April, 2005ISGC 2005, Taiwan mpiBLAST-g2 Performance Evaluation (perfect case) Elapsed timeSpeedup Database: est_human ~ 3.5 GBytes Queries: 441 test sequences ~ 300 KBytes Overall speedup is approximately linear — Searching + Merging — BioSeq fetching — Overall

28 April, 2005ISGC 2005, Taiwan mpiBLAST-g2 Performance Evaluation (worse case) Elapsed timeSpeedup Database: drosophila NT ~ 122 MBytes Queries: 441 test sequences ~ 300 KBytes The overall speedup is limited by the unscalable BioSeq fetching — Searching + Merging — BioSeq fetching — Overall

28 April, 2005ISGC 2005, Taiwan Issues of mpiBLAST-g2 Single error will crash the whole job –The MPICH nature –Error might be due to the transient problem on the loosely coupled Grid environment MPI Job will be started only when all resources are available –Different level of resource availability Error recovery is required for –providing a robust application service on the Grid –efficiently using the Grid resources Asynchronous task dispatching/pulling to use the available resources immediately

28 April, 2005ISGC 2005, Taiwan The DIANE DIstributed ANalysis Environment Lightweight distributed framework for parallel scientific applications in master-worker model –A perfect match of the mpiBLAST computing model Current applications –BLAST for Genomic Sequence Analysis (DIANE-BLAST) –Geant4 Simulation for Radiotherapy and Astrophysics –Image Rendering –Data Analysis for High Energy Physics

28 April, 2005ISGC 2005, Taiwan DIANE Features Rapid prototyping –Python and CORBA Error recovery –Heartbeat worker health check –Resubmission of failed tasks –User defined error recovery method No need of outbound connectivity –Proxy of workers with only private IP Job submitters for –Simple fork –Condor, LSF, SGE, PBS –GT2, LCG, gLite Pull Model Batch and Interactive Distributed workers planner integrator

28 April, 2005ISGC 2005, Taiwan DIANE-BLAST implementation Splitting mpiBLAST-g2 to DIANE components –Master (Planner and Integrator), Worker Wrapping each component with Python –Hooking core BLAST C libraries with python swig Implementing the DIANE GT2 job submitter –For running workers on the GT2-enabled clusters Reusing the deployed databases for mpiBLAST-g2

28 April, 2005ISGC 2005, Taiwan mpiBLAST-g2 vs. DIANE-BLAST The Speedup Query –Drosophila chromosome 4 –size: 1.2 Mbps DB –Drosophila nucleotide sequence database –size: 1170 seq. 122 Mbps –no. fragments: 32 Computing Resource –Available # of CPU: 12 –PIII 1.4GHz –1GByte Memory Speedup of mpiBLAST-g2 Speedup of DIANE-BLAST

28 April, 2005ISGC 2005, Taiwan mpiBLAST-g2 vs. DIANE-BLAST The Worker Lifeline DIANE-BLAST task dispatching Handled by DIANE’s task thread Due to the bugs in the current DIANE release DIANE-BLAST task dispatching Handled by DIANE’s task thread Due to the bugs in the current DIANE release mpiBLAST-g2 task dispatching mpiBLAST-g2 task handling logic mpiBLAST-g2 task dispatching mpiBLAST-g2 task handling logic

28 April, 2005ISGC 2005, Taiwan mpiBLAST-g2 vs. DIANE-BLAST Overall Comparisons mpiBLAST-g2 –Master-Worker model implemented by using MPICH-g2 libraries –Gridification efforts Implementing database sharing with GASSCOPY API Recompilation with MPICH-g2 and GT2 libraries –Error recovery Need the fault-tolerance MPI –Cross cluster computation Requiring outbound connectivity on each worker –Performance/Throughput In cluster performance is as well as the original mpiBLAST DIANE-BLAST –Pluggable application for DIANE Master-Worker framework –Gridification efforts Through the gridified DIANE framework –Error recovery Task resubmission Tracking the health of each worker –Cross cluster computation Using proxy for workers with private IPs –Performance/Throughput Performance can be tuned by controlling the job thread

28 April, 2005ISGC 2005, Taiwan Summary Two grid-enabled BLAST implementations (mpiBLAST-g2 and DIANE-BLAST) were introduced for efficient handling the BLAST jobs on the Grid Both implementations are based on the Master-Worker model for distributing BLAST jobs on the Grid The mpiBLAST-g2 has good scalability and speedup in some cases –Require the fault-tolerance MPI implementation for error recovery –In the unscalable cases, BioSeq fetching is the bottleneck DIANE-BLAST provides flexible mechanism for error recovery –Any master-worker workflow can be easily plugged into this framework –The job thread control should be improved to achieving the good performance and scalability

28 April, 2005ISGC 2005, Taiwan Thanks for your attention!!