XSEDE 13 July 24, 2013. Galaxy Team: PSC Team:

Slides:



Advertisements
Similar presentations
Grid Resource Allocation Management (GRAM) GRAM provides the user to access the grid in order to run, terminate and monitor jobs remotely. The job request.
Advertisements

Natasha Pavlovikj, Kevin Begcy, Sairam Behera, Malachy Campbell, Harkamal Walia, Jitender S.Deogun University of Nebraska-Lincoln Evaluating Distributed.
1 OBJECTIVES To generate a web-based system enables to assemble model configurations. to submit these configurations on different.
High Performance Computing Course Notes Grid Computing.
Condor-G: A Computation Management Agent for Multi-Institutional Grids James Frey, Todd Tannenbaum, Miron Livny, Ian Foster, Steven Tuecke Reporter: Fu-Jiun.
A Computation Management Agent for Multi-Institutional Grids
(e)Science-Driven, Production- Quality, Distributed Grid and Cloud Data Infrastructure for the Transformative, Disruptive, Revolutionary, Next-Generation.
Milos Kobliha Alejandro Cimadevilla Luis de Alba Parallel Computing Seminar GROUP 12.
The new The new MONARC Simulation Framework Iosif Legrand  California Institute of Technology.
UNICORE UNiform Interface to COmputing REsources Olga Alexandrova, TITE 3 Daniela Grudinschi, TITE 3.
UMIACS PAWN, LPE, and GRASP data grids Mike Smorul.
Web-based Portal for Discovery, Retrieval and Visualization of Earth Science Datasets in Grid Environment Zhenping (Jane) Liu.
IPlant Collaborative Tools and Services Workshop iPlant Collaborative Tools and Services Workshop Overview of Atmosphere.
Client/Server Grid applications to manage complex workflows Filippo Spiga* on behalf of CRAB development team * INFN Milano Bicocca (IT)
Building Data-intensive Pipelines Ravi K Madduri Argonne National Lab University of Chicago.
The SAM-Grid Fabric Services Gabriele Garzoglio (for the SAM-Grid team) Computing Division Fermilab.
Apache Airavata GSOC Knowledge and Expertise Computational Resources Scientific Instruments Algorithms and Models Archived Data and Metadata Advanced.
 Cloud computing  Workflow  Workflow lifecycle  Workflow design  Workflow tools : xcp, eucalyptus, open nebula.
Customized cloud platform for computing on your terms !
Data Management Kelly Clynes Caitlin Minteer. Agenda Globus Toolkit Basic Data Management Systems Overview of Data Management Data Movement Grid FTP Reliable.
Chapter 6 Operating System Support. This chapter describes how middleware is supported by the operating system facilities at the nodes of a distributed.
Presenter: Dipesh Gautam.  Introduction  Why Data Grid?  High Level View  Design Considerations  Data Grid Services  Topology  Grids and Cloud.
Digital Object Architecture
IPlant Collaborative Tools and Services Workshop iPlant Collaborative Tools and Services Workshop Overview of Atmosphere.
Flexibility and user-friendliness of grid portals: the PROGRESS approach Michal Kosiedowski
GT Components. Globus Toolkit A “toolkit” of services and packages for creating the basic grid computing infrastructure Higher level tools added to this.
The Data Grid: Towards an Architecture for the Distributed Management and Analysis of Large Scientific Dataset Caitlin Minteer & Kelly Clynes.
Grid Resource Allocation and Management (GRAM) Execution management Execution management –Deployment, scheduling and monitoring Community Scheduler Framework.
INFSO-RI Module 01 ETICS Overview Etics Online Tutorial Marian ŻUREK Baltic Grid II Summer School Vilnius, 2-3 July 2009.
CSF4 Meta-Scheduler Name: Zhaohui Ding, Xiaohui Wei
The PROGRESS Grid Service Provider Maciej Bogdański Portals & Portlets 2003 Edinburgh, July 14th-17th.
Introduction. Readings r Coulouris, Dollimore and Kindberg Distributed Systems: Concepts and Design Edn. 3 m Note: All figures from this book.
Database Design and Management CPTG /23/2015Chapter 12 of 38 Functions of a Database Store data Store data School: student records, class schedules,
GEM Portal and SERVOGrid for Earthquake Science PTLIU Laboratory for Community Grids Geoffrey Fox, Marlon Pierce Computer Science, Informatics, Physics.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
PROGRESS: ICCS'2003 GRID SERVICE PROVIDER: How to improve flexibility of grid user interfaces? Michał Kosiedowski.
Cracow Grid Workshop October 2009 Dipl.-Ing. (M.Sc.) Marcus Hilbrich Center for Information Services and High Performance.
Tool Integration with Data and Computation Grid GWE - “Grid Wizard Enterprise”
Grid Computing at Yahoo! Sameer Paranjpye Mahadev Konar Yahoo!
Ames Research CenterDivision 1 Information Power Grid (IPG) Overview Anthony Lisotta Computer Sciences Corporation NASA Ames May 2,
9 Systems Analysis and Design in a Changing World, Fourth Edition.
Policy Based Data Management Data-Intensive Computing Distributed Collections Grid-Enabled Storage iRODS Reagan W. Moore 1.
NA-MIC National Alliance for Medical Image Computing UCSD: Engineering Core 2 Portal and Grid Infrastructure.
9 Systems Analysis and Design in a Changing World, Fourth Edition.
Review of Condor,SGE,LSF,PBS
Overview of Privilege Project at Fermilab (compilation of multiple talks and documents written by various authors) Tanya Levshina.
Cole David Ronnie Julio. Introduction Globus is A community of users and developers who collaborate on the use and development of open source software,
Campus grids: e-Infrastructure within a University Mike Mineter National e-Science Centre 14 February 2006.
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
Some comments on Portals and Grid Computing Environments PTLIU Laboratory for Community Grids Geoffrey Fox, Marlon Pierce Computer Science, Informatics,
Distributed Data for Science Workflows Data Architecture Progress Report December 2008.
Globus and PlanetLab Resource Management Solutions Compared M. Ripeanu, M. Bowman, J. Chase, I. Foster, M. Milenkovic Presented by Dionysis Logothetis.
Tool Integration with Data and Computation Grid “Grid Wizard 2”
Vignesh Ravindran Sankarbala Manoharan. Infrastructure As A Service (IAAS) is a model that is used to deliver a platform virtualization environment with.
ATLAS-specific functionality in Ganga - Requirements for distributed analysis - ATLAS considerations - DIAL submission from Ganga - Graphical interfaces.
PROGRESS: GEW'2003 Using Resources of Multiple Grids with the Grid Service Provider Michał Kosiedowski.
MSF and MAGE: e-Science Middleware for BT Applications Sep 21, 2006 Jaeyoung Choi Soongsil University, Seoul Korea
Building Preservation Environments with Data Grid Technology Reagan W. Moore Presenter: Praveen Namburi.
SAM architecture EGEE 07 Service Availability Monitor for the LHC experiments Simone Campana, Alessandro Di Girolamo, Nicolò Magini, Patricia Mendez Lorenzo,
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI EGI Services for Distributed e-Infrastructure Access Tiziana Ferrari on behalf.
IPlant Collaborative Tools and Services Workshop iPlant Collaborative Tools and Services Workshop Overview of Atmosphere.
PaaS services for Computing and Storage
CLOUD ARCHITECTURE Many organizations and researchers have defined the architecture for cloud computing. Basically the whole system can be divided into.
StoRM: a SRM solution for disk based storage systems
Joseph JaJa, Mike Smorul, and Sangchul Song
Tools and Services Workshop Overview of Atmosphere
THE STEPS TO MANAGE THE GRID
Outline Midterm results summary Distributed file systems – continued
Storing and Accessing G-OnRamp’s Assembly Hubs outside of Galaxy
Mats Rynge USC Information Sciences Institute
Presentation transcript:

XSEDE 13 July 24, 2013

Galaxy Team: PSC Team:

 Galaxy: The Need, the Framework, the Popularity Bottleneck  Distributing Galaxy work and data flows to XSEDE systems: first steps  The Future: Galaxy Gateway(s)

643 HiSeqs = 6.5 Pb/year

Biology has rapidly become data intensive, and dependent on computational methods How can we ensure that these methods are accessible to researchers?...while also ensuring that scientific results remain reproducible?

A free (for everyone) web service integrating a wealth of tools, compute resources, terabytes of reference data and permanent storage Open-source software allowing anyone to freely deploy or extend this platform A community of users and developers

 Mostly command line tools, a declarative XML description of the interface, how to generate a command line  Designed to be as easy as possible for tool authors, while still allowing rigorous reasoning  Workflows can be constructed from scratch or extracted from existing analysis histories  Facilitate reuse, as well as providing precise reproducibility of a complex analysis

Describe analysis tool behavior abstractly Analysis environment automatically and transparently tracks details Workflow system for complex analysis, constructed explicitly or automatically Pervasive sharing, and publication of documents with integrated analysis

 Entire Galaxy workflows or component tasks.  Especially, tasks that require HPC, e.g. de-novo assembly applications Velvet (of genome) and Trinity (of transcriptome) to PSC Blacklight (up to 16 TB of coherent shared memory per process).  Should be transparent to the user of usegalaxy.org.

 Data Migration: Galaxy currently relies on a shared filesystem between the instance host and the execution server to store the reference and user data required by the workflow. This is implemented via NFS.  Remote Job Submission: Galaxy job execution currently requires a direct interface with the resource manager on the execution server.

Penn State PSC Data SuperCell Data Generation Nodes Storage 10gigE link Transferred 470TB in 21 days from PSU to PSC (average ~22TB/day; peak 40 TB/day) rsync used to initially stage and synchronize subsequent updates Data copy maintained in PSC in /arc file system available from compute nodes

Penn State PSC /galaxys2 Data Generation and Processing Nodes Data Generation and Processing Nodes SLASH2 Wide-Area Common File system GalaxyFS Access is identical from PSU and PSC to the shared dataset via /galaxys2 SLASH2 file system handles consistency and multiple residency coherency and presence Local copies are maintained for performance Jobs run on PSC compute resources such as Blacklight and at PSU

Metadata Server (MDS) I/O Servers (IOS) Clients I/O Servers (IOS) I/O Servers (IOS) I/O Servers (IOS) One at PSU and one at PSC for performance Converts pathnames to object IDs Schedules updates when copies become inconsistent Consistency protocol to avoid incoherent data Residency and network scheduling policies enforced Clients are compute resources & dedicated front ends Dataset residency requests issued from administrators and/or users I/O servers are very lightweight Can use most backing file systems (ZFS, ext4fs, etc.) READ and WRITE Data consistency updates All other file ops (RENAME, SYMLINK, etc.)

 Created a new Galaxy job-running plugin for submission via a remote shell program and PSC “SIMON” Torque command line: CLI runner.  Velvet and Trinity have been incorporated into the Galaxy web platform using Galaxy’s XML interface.  Test jobs have been successfully submitted from Penn State and executed on Blacklight using the data replicated via SLASH2 from Penn State to PSC.

 Integrate this work with the production public Galaxy site, usegalaxy.org  Dynamic job submission, allowing the selection of appropriate remote or local resources (cores, memory, walltime, etc.) based on individual job requirements, using an Open Grid Services Architecture Basic Execution Service compatible service, such as Unicore.

 Galaxy-controlled data management, to intelligently create replicas as close as possible to the compute resource that will use the data.  Authentication with Galaxy instances, using XSEDE or other credentials, e.g., InCommon/CILogon.  Additional data transfer capabilities in Galaxy, such as IRODS and Globus Online.

Ultimately, we envision that any Galaxy instance (in any lab, not just Galaxy Main) will be able to spawn jobs, access data, and share data on external infrastructure whether this is an XSEDE resource, a cluster of Amazon EC2 machines, a remote storage array, etc.