Introduction The Open Science Grid (OSG) is a consortium of more than 100 institutions including universities, national laboratories, and computing centers.

Slides:



Advertisements
Similar presentations
Network II.5 simulator ..
Advertisements

Dominik Stokłosa Pozna ń Supercomputing and Networking Center, Supercomputing Department INGRID 2008 Lacco Ameno, Island of Ischia, ITALY, April 9-11 Workflow.
A Workflow Engine with Multi-Level Parallelism Supports Qifeng Huang and Yan Huang School of Computer Science Cardiff University
Current Testbed : 100 GE 2 sites (NERSC, ANL) with 3 nodes each. Each node with 4 x 10 GE NICs Measure various overheads from protocols and file sizes.
 Contributing >30% of throughput to ATLAS and CMS in Worldwide LHC Computing Grid  Reliant on production and advanced networking from ESNET, LHCNET and.
High Performance Computing Course Notes Grid Computing.
A Computation Management Agent for Multi-Institutional Grids
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
GridFlow: Workflow Management for Grid Computing Kavita Shinde.
DataGrid Kimmo Soikkeli Ilkka Sormunen. What is DataGrid? DataGrid is a project that aims to enable access to geographically distributed computing power.
Feb 23, 20111/22 LSST Simulations on OSG Feb 23, 2011 Gabriele Garzoglio for the OSG Task Force on LSST Computing Division, Fermilab Overview The Open.
April 2009 OSG Grid School - RDU 1 Open Science Grid John McGee – Renaissance Computing Institute University of North Carolina, Chapel.
Minerva Infrastructure Meeting – October 04, 2011.
Ian M. Fisk Fermilab February 23, Global Schedule External Items ➨ gLite 3.0 is released for pre-production in mid-April ➨ gLite 3.0 is rolled onto.
Assessment of Core Services provided to USLHC by OSG.
Authors: Weiwei Chen, Ewa Deelman 9th International Conference on Parallel Processing and Applied Mathmatics 1.
GlobusWorld 2012: Experience with EXPERIENCE WITH GLOBUS ONLINE AT FERMILAB Gabriele Garzoglio Computing Sector Fermi National Accelerator.
Open Science Grid For CI-Days Internet2: Fall Member Meeting, 2007 John McGee – OSG Engagement Manager Renaissance Computing Institute.
Dr. Thomas Hacker Co-Leader for Information Technology George E. Brown Network for Earthquake Engineering Simulation Assistant Professor, Computer & Information.
OSG Public Storage and iRODS
National Center for Supercomputing Applications Observational Astronomy NCSA projects radio astronomy: CARMA & SKA optical astronomy: DES & LSST access:
DOE BER Climate Modeling PI Meeting, Potomac, Maryland, May 12-14, 2014 Funding for this study was provided by the US Department of Energy, BER Program.
Central Reconstruction System on the RHIC Linux Farm in Brookhaven Laboratory HEPIX - BNL October 19, 2004 Tomasz Wlodek - BNL.
D0 SAM – status and needs Plagarized from: D0 Experiment SAM Project Fermilab Computing Division.
Open Science Grid For CI-Days Elizabeth City State University Jan-2008 John McGee – OSG Engagement Manager Manager, Cyberinfrastructure.
4th EGEE user forum / OGF 25 Catania, TheoSSA on AstroGrid-D Iliya Nickelt (AIP / GAVO), Thomas Rauch (IAAT / GAVO), Harry Enke (AIP.
CHEP'07 September D0 data reprocessing on OSG Authors Andrew Baranovski (Fermilab) for B. Abbot, M. Diesburg, G. Garzoglio, T. Kurca, P. Mhashilkar.
CCGrid 2014 Improving I/O Throughput of Scientific Applications using Transparent Parallel Compression Tekin Bicer, Jian Yin and Gagan Agrawal Ohio State.
GlobusWorld 2012: Experience with EXPERIENCE WITH GLOBUS ONLINE AT FERMILAB Gabriele Garzoglio Computing Sector Fermi National Accelerator.
Grid Technologies  Slide text. What is Grid?  The World Wide Web provides seamless access to information that is stored in many millions of different.
Current Testbed : 100 GE 2 sites (NERSC, ANL) with 3 nodes each. Each node with 4 x 10 GE NICs Measure various overheads from protocols and file sizes.
Sep 21, 20101/14 LSST Simulations on OSG Sep 21, 2010 Gabriele Garzoglio for the OSG Task Force on LSST Computing Division, Fermilab Overview OSG Engagement.
CSIU Submission of BLAST jobs via the Galaxy Interface Rob Quick Open Science Grid – Operations Area Coordinator Indiana University.
Grid Workload Management Massimo Sgaravatto INFN Padova.
Instrumentation of the SAM-Grid Gabriele Garzoglio CSC 426 Research Proposal.
10/24/2015OSG at CANS1 Open Science Grid Ruth Pordes Fermilab
LSST: Preparing for the Data Avalanche through Partitioning, Parallelization, and Provenance Kirk Borne (Perot Systems Corporation / NASA GSFC and George.
November SC06 Tampa F.Fanzago CRAB a user-friendly tool for CMS distributed analysis Federica Fanzago INFN-PADOVA for CRAB team.
Tier-2  Data Analysis  MC simulation  Import data from Tier-1 and export MC data CMS GRID COMPUTING AT THE SPANISH TIER-1 AND TIER-2 SITES P. Garcia-Abia.
Ames Research CenterDivision 1 Information Power Grid (IPG) Overview Anthony Lisotta Computer Sciences Corporation NASA Ames May 2,
1 Development of a High-Throughput Computing Cluster at Florida Tech P. FORD, R. PENA, J. HELSBY, R. HOCH, M. HOHLMANN Physics and Space Sciences Dept,
Ruth Pordes November 2004TeraGrid GIG Site Review1 TeraGrid and Open Science Grid Ruth Pordes, Fermilab representing the Open Science.
HIGUCHI Takeo Department of Physics, Faulty of Science, University of Tokyo Representing dBASF Development Team BELLE/CHEP20001 Distributed BELLE Analysis.
US-CMS T2 Centers US-CMS Tier 2 Report Patricia McBride Fermilab GDB Meeting August 31, 2007 Triumf - Vancouver.
Scheduling MPI Workflow Applications on Computing Grids Juemin Zhang, Waleed Meleis, and David Kaeli Electrical and Computer Engineering Department, Northeastern.
CMS Usage of the Open Science Grid and the US Tier-2 Centers Ajit Mohapatra, University of Wisconsin, Madison (On Behalf of CMS Offline and Computing Projects)
Super Computing 2000 DOE SCIENCE ON THE GRID Storage Resource Management For the Earth Science Grid Scientific Data Management Research Group NERSC, LBNL.
Final Implementation of a High Performance Computing Cluster at Florida Tech P. FORD, X. FAVE, K. GNANVO, R. HOCH, M. HOHLMANN, D. MITRA Physics and Space.
Computing Sector, Fermi National Accelerator Laboratory 4/12/12GlobusWorld 2012: Experience with
Università di Perugia Enabling Grids for E-sciencE Status of and requirements for Computational Chemistry NA4 – SA1 Meeting – 6 th April.
Parag Mhashilkar Computing Division, Fermi National Accelerator Laboratory.
April 25, 2006Parag Mhashilkar, Fermilab1 Resource Selection in OSG & SAM-On-The-Fly Parag Mhashilkar Fermi National Accelerator Laboratory Condor Week.
Considering Time in Designing Large-Scale Systems for Scientific Computing Nan-Chen Chen 1 Sarah S. Poon 2 Lavanya Ramakrishnan 2 Cecilia R. Aragon 1,2.
July 26, 2007Parag Mhashilkar, Fermilab1 DZero On OSG: Site And Application Validation Parag Mhashilkar, Fermi National Accelerator Laboratory.
Northwest Indiana Computational Grid Preston Smith Rosen Center for Advanced Computing Purdue University - West Lafayette West Lafayette Calumet.
Building on virtualization capabilities for ExTENCI Carol Song and Preston Smith Rosen Center for Advanced Computing Purdue University ExTENCI Kickoff.
Open Science Grid Consortium Storage on Open Science Grid Placing, Using and Retrieving Data on OSG Resources Abhishek Singh Rana OSG Users Meeting July.
Parag Mhashilkar (Fermi National Accelerator Laboratory)
Scientific Computing at Fermilab Lothar Bauerdick, Deputy Head Scientific Computing Division 1 of 7 10k slot tape robots.
VO Experiences with Open Science Grid Storage OSG Storage Forum | Wednesday September 22, 2010 (10:30am)
INTRODUCTION TO XSEDE. INTRODUCTION  Extreme Science and Engineering Discovery Environment (XSEDE)  “most advanced, powerful, and robust collection.
OSG User Support OSG User Support March 13, 2013 Chander Sehgal
Accessing the VI-SEEM infrastructure
LQCD Computing Operations
 YongPyong-High Jan We appreciate that you give an opportunity to have this talk. Our Belle II computing group would like to report on.
Overview of Workflows: Why Use Them?
High Throughput Computing for Astronomers
The LHCb Computing Data Challenge DC06
Presentation transcript:

Introduction The Open Science Grid (OSG) is a consortium of more than 100 institutions including universities, national laboratories, and computing centers. OSG fosters scientific research and knowledge supporting the computational activities of more than 80 communities. Because of its federating model, during a cycle of low resource utilization at an institution, communities have the ability of utilizing the idle resources “opportunistically” via a common interface. [1] The OSG User Support group helps communities port their computing operations to OSG. This poster describes application requirements for some new communities and the processes and workflows that User Support chose to meet those requirements. Other communities with similar requirements could consider similar solutions. Workflow Metrics The following workflow metrics help illustrate the similarities between different communities: Wall time: Aggregate execution time of the computational campaign Data: Aggregated transfer size of input, output and the executable Workflow Steps: A workflow is a logical unit of computation, composed of one or more data processing steps. Each step runs as an independent Grid job. Job Count: the number of jobs in the campaign. Computing Model OSG primarily supplies Distributed High Throughput Computing (DHTC) resources. These are optimized for running applications that require a lot of compute time, but do not require time critical communication between different instances of these applications. Solving a problem often requires running many different instances of an application. This poster discusses some workflows that have been implemented to coordinate these instances. These may serve as a pattern for running other applications facing similar challenges. LSST USCMS (1day) NEES EIC D0 (1day) DES (1day) Pheno Acknowledgments: Fermilab is operated by the Fermi Research Alliance, LLC under Contract No. DE-AC02-07CH11359 with the United States Department of Energy. The Open Science Grid ( is funded by the National Science Foundation and the Office of Science, Department of Energy (DOE). Science Goals 8 meter wide-field ground- based telescope providing time-lapse digital imaging of faint astronomical objects across the entire visible sky every few nights for 10 years. This computing project focuses on image simulation. LSST Workflow 1.(Only once) Pre-stage star catalog and focal plane configuration files (15 GB) 2.Submit 1 job to trim the pre-staged catalog file into 189 files, one per CCD chip in the camera. 3.Submit 2 x 189 jobs: simulate 1 image pair (same image with 2 exposures). Transfer “instance catalog” (telescope position, wind speed, etc.) with each job 4.Gather output, perform bookkeeping, etc. Features Needs input pre-processing  complex workflow. Large catalog file pre-staged at all sites. Workflow simplified at Purdue since then. Science Goals To better understand the properties of dark energy, the Dark Energy Survey (DES) is planning to collect and process images of the night sky. DES Features Large amount of data to transfer. Complex requirements and end-to-end Grid system. NASA/ESA via Wikipedia Workflow For prototyping, we are using a workflow similar to the one for the Pheno group at SLAC. Final workflow would be complicated by large data requirements and end-to-end system complexity. NEES Features Large amount of output (3GB / job). This caused multiple challenges, including jobs occupying computing resources after completion, waiting on the output data transfer queue. This community would benefit from using the Pheno workflow, instead. Workflow 1.Submit the OpenSees simulation application to OSG sites. Condor transfers input data 2.Return the data using Condor to the submit host. 3.Use Globus Online to transfer the data to the user’s archive. A. R. Barbosa, J. P. Conte, J. I. Restrepo, UCSD Science Goals The Network for Earthquake Engineering Simulation studies the response of buildings and other structures to earthquakes. [4] Completion time of 340 jobs at 16 sites. Jobs wait after completioon to transfer their output. Graph and tables [2]: The numbers in the “Data Transferred” column, and a few others, are estimates. Aim of the graph is to give a sense of the scale of workflows. Data for US CMS and D0 on OSG are shown for comparison. The data is for only one day of their operations so that they’re on the same scale as the OSG projects. The OSG projects, however, accumulate their time over the course of months. Among the OSG sites, D0 runs preferentially on FermiGrid (250 TB/day transferred). Science Goals The Electron Ion Collider (EIC) is a proposed facility for studying the structure of the nuclei. Engineers need a large amount of computations to define its design.[3] EIC Workflow 1.(Only once) Pre-stage a 1GB read-only file to each site. This way, that file does not need to be repeatedly transferred over the wide area network. 2.Submit the jobs to the sites with the pre- staged files. Transfer the application. 3.Jobs run, and read the pre-staged files. 4.Condor transfers output data to submit host. 5.User takes possession of the results Features Lot of internal site transfers, as compared to the data transferred over the WAN. Tobias Toll, Thomas Ulrich BNL Worker Nodes Distributed File System mount Site A Worker Nodes Distributed File System mount Site B User Archive 5 Staging Server Submit Host Worker Nodes Site A Worker Nodes Site B Submit Host User Archive Pheno Science Goals The Phenomenology group at SLAC ran an application called Sherpa that does multi- particle Quantum Chromo Dynamics calculations using Monte Carlo methods. Workflow 1.(Only once) Stage software to each site. 2.Submit the jobs to the sites and transfer input data to local storage. 3.Transferring output data to local storage at site. 4.Users retrieves output data directly from the storage at each site. Features Fair amount of output data. After the experience with NEES on large output, using a distributed data workflow. Worker Nodes Distributed File System mount Site A Worker Nodes Distributed File System mount Site B User Archive Staging Server Submit Host References Derek Weitzel, UNL Fermilab/DES collaboration [1] "What is the Open Science Grid?", Documentation/WhatIsOSG [2] The OSG Gratia Server, the PheEDEx monitoring system, and discussion with representatives from experiments. [3] T. Toll, "Electron Ion Collider Simulations on OSG”, Management/NovDec2011 [4] A. R. Barbosa, J. P. Conte, J. I. Restrepo, “Running OpenSees Production for NEES on OSG”, Management/Oct2011Newsletter Work in collaboration with DES Data Management group. This project is being evaluated for running on OSG. LSST Image Simulation team Project Workflow Steps Job CountWall time (h) Data (TB) Hours per job Pheno at SLAC EIC LSST Simulation NEES OpenSees DES (1 day) US CMS (1 day) D0 (non-local) (1 day)