April 19, 2005Scott Koranda1 The LIGO Scientific Collaboration Data Grid Scott Koranda UW-Milwaukee On behalf of the LIGO Scientific Collaboration.

Slides:



Advertisements
Similar presentations
Case Study 1: Data Replication for LIGO Scott Koranda Ann Chervenak.
Advertisements

11/12/2003LIGO Document G Z1 Data reduction for S3 I Leonor (UOregon), P Charlton (CIT), S Anderson (CIT), K Bayer (MIT), M Foster (PSU), S Grunewald.
MTA SZTAKI Hungarian Academy of Sciences Grid Computing Course Porto, January Introduction to Grid portals Gergely Sipos
The Open Science Grid: Bringing the power of the Grid to scientific research
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
Building on the BIRN Workshop BIRN Systems Architecture Overview Philip Papadopoulos – BIRN CC, Systems Architect.
Workload Management Workpackage Massimo Sgaravatto INFN Padova.
Experience with ATLAS Data Challenge Production on the U.S. Grid Testbed Kaushik De University of Texas at Arlington CHEP03 March 27, 2003.
Intermediate HTCondor: Workflows Monday pm Greg Thain Center For High Throughput Computing University of Wisconsin-Madison.
Design considerations for the Indigo Data Analysis Centre. Anand Sengupta, University of Delhi Thanks to -Maria Alessandra Papa (AEI) -Stuart Anderson.
QCDgrid Technology James Perry, George Beckett, Lorna Smith EPCC, The University Of Edinburgh.
CONDOR DAGMan and Pegasus Selim Kalayci Florida International University 07/28/2009 Note: Slides are compiled from various TeraGrid Documentations.
The Data Replication Service Ann Chervenak Robert Schuler USC Information Sciences Institute.
Vladimir Litvin, Harvey Newman Caltech CMS Scott Koranda, Bruce Loftis, John Towns NCSA Miron Livny, Peter Couvares, Todd Tannenbaum, Jamie Frey Wisconsin.
CANS Meeting (December 1, 2004)Paul Avery1 University of Florida UltraLight U.S. Grid Projects and Open Science Grid Chinese American.
G D LSC ITR2003, ANL, LIGO Scientific Collaboration ITR 2003 Advisory Committee Meeting K. Blackburn 1, P. Brady 2, S. Finn 3, E.
The Grid is a complex, distributed and heterogeneous execution environment. Running applications requires the knowledge of many grid services: users need.
LSC Segment Database Duncan Brown Caltech LIGO-G Z.
LIGO-G9900XX-00-M LIGO and GriPhyN Carl Kesselman Ewa Deelman Sachin Chouksey Roy Williams Peter Shawhan Albert Lazzarini Kent Blackburn CaltechUSC/ISI.
Data Management Kelly Clynes Caitlin Minteer. Agenda Globus Toolkit Basic Data Management Systems Overview of Data Management Data Movement Grid FTP Reliable.
Don Quijote Data Management for the ATLAS Automatic Production System Miguel Branco – CERN ATC
10/20/05 LIGO Scientific Collaboration 1 LIGO Data Grid: Making it Go Scott Koranda University of Wisconsin-Milwaukee.
LIGO- G E LIGO Grid Applications Development Within GriPhyN Albert Lazzarini LIGO Laboratory, Caltech GriPhyN All Hands.
Patrick R Brady University of Wisconsin-Milwaukee
Profiling Grid Data Transfer Protocols and Servers George Kola, Tevfik Kosar and Miron Livny University of Wisconsin-Madison USA.
ARGONNE  CHICAGO Ian Foster Discussion Points l Maintaining the right balance between research and development l Maintaining focus vs. accepting broader.
G Z LIGO Scientific Collaboration Grid Patrick Brady University of Wisconsin-Milwaukee LIGO Scientific Collaboration.
QCDGrid Progress James Perry, Andrew Jackson, Stephen Booth, Lorna Smith EPCC, The University Of Edinburgh.
 The workflow description modified to output a VDS DAX.  The workflow description toolkit developed allows any concrete workflow description to be migrated.
ILDG Middleware Status Chip Watson ILDG-6 Workshop May 12, 2005.
10/24/2015OSG at CANS1 Open Science Grid Ruth Pordes Fermilab
Virtual Data Grid Architecture Ewa Deelman, Ian Foster, Carl Kesselman, Miron Livny.
Resource Brokering in the PROGRESS Project Juliusz Pukacki Grid Resource Management Workshop, October 2003.
Part Four: The LSC DataGrid Part Four: LSC DataGrid A: Data Replication B: What is the LSC DataGrid? C: The LSCDataFind tool.
Quick Introduction to NorduGrid Oxana Smirnova 4 th Nordic LHC Workshop November 23, 2001, Stockholm.
Status of Standalone Inspiral Code Duncan Brown University of Wisconsin-Milwaukee LIGO Scientific Collaboration Inspiral Working Group LIGO-G Z.
NA-MIC National Alliance for Medical Image Computing UCSD: Engineering Core 2 Portal and Grid Infrastructure.
Intermediate Condor: Workflows Rob Quick Open Science Grid Indiana University.
What is SAM-Grid? Job Handling Data Handling Monitoring and Information.
Some Grid Science California Institute of Technology Roy Williams Paul Messina Grids and Virtual Observatory Grids and and LIGO.
The LIGO Scientific Collaboration Data Grid Client/Server Environment Junwei Cao MIT LIGO Laboratory For the LIGO Scientific Collaboration.
The GriPhyN Planning Process All-Hands Meeting ISI 15 October 2001.
Ruth Pordes November 2004TeraGrid GIG Site Review1 TeraGrid and Open Science Grid Ruth Pordes, Fermilab representing the Open Science.
LIGO-G E LIGO Scientific Collaboration Data Grid Status Albert Lazzarini Caltech LIGO Laboratory Trillium Steering Committee Meeting 20 May 2004.
Junwei Cao March LIGO Document ID: LIGO-G Computing Infrastructure for Gravitational Wave Data Analysis Junwei Cao.
Alain Roy Computer Sciences Department University of Wisconsin-Madison Condor & Middleware: NMI & VDT.
December 26, 2015 RHIC/USATLAS Grid Computing Facility Overview Dantong Yu Brookhaven National Lab.
Contact: Junwei Cao SC2005, Seattle, WA, November 12-18, 2005 The authors gratefully acknowledge the support of the United States National.
LIGO Plans for OSG J. Kent Blackburn LIGO Laboratory California Institute of Technology Open Science Grid Technical Meeting UCSD December 15-17, 2004.
State of LSC Data Analysis and Software LSC Meeting LIGO Hanford Observatory November 11 th, 2003 Kent Blackburn, Stuart Anderson, Albert Lazzarini LIGO.
Scott Koranda, UWM & NCSA 14 January 2016www.griphyn.org Lightweight Data Replicator Scott Koranda University of Wisconsin-Milwaukee & National Center.
11/12/2003LIGO-G Z1 Data reduction for S3 P Charlton (CIT), I Leonor (UOregon), S Anderson (CIT), K Bayer (MIT), M Foster (PSU), S Grunewald (AEI),
AERG 2007Grid Data Management1 Grid Data Management GridFTP Carolina León Carri Ben Clifford (OSG)
GriPhyN Project Paul Avery, University of Florida, Ian Foster, University of Chicago NSF Grant ITR Research Objectives Significant Results Approach.
Biomedical Informatics Research Network The Storage Resource Broker & Integration with NMI Middleware Arcot Rajasekar, BIRN-CC SDSC October 9th 2002 BIRN.
Super Computing 2000 DOE SCIENCE ON THE GRID Storage Resource Management For the Earth Science Grid Scientific Data Management Research Group NERSC, LBNL.
LIGO-G W Use of Condor by the LIGO Scientific Collaboration Gregory Mendell, LIGO Hanford Observatory On behalf of the LIGO Scientific Collaboration.
LIGO-G Z1 Using Condor for Large Scale Data Analysis within the LIGO Scientific Collaboration Duncan Brown California Institute of Technology.
LIGO: The Laser Interferometer Gravitational-wave Observatory Sarah Caudill Louisiana State University Physics Graduate Student LIGO Hanford LIGO Livingston.
1 CMS Virtual Data Overview Koen Holtman Caltech/CMS GriPhyN all-hands meeting, Marina del Rey April 9, 2001.
Managing LIGO Workflows on OSG with Pegasus Karan Vahi USC Information Sciences Institute
VO Experiences with Open Science Grid Storage OSG Storage Forum | Wednesday September 22, 2010 (10:30am)
A Web Based Job Submission System for a Physics Computing Cluster David Jones IOP Particle Physics 2004 Birmingham 1.
Feb 4, 2005Scott Koranda1 Cataloging, Replicating, and Managing LIGO Data on the Grid Scott Koranda UW-Milwaukee On behalf of the LIGO.
Running LIGO workflows on the OSG
Scott Koranda, UWM & NCSA 20 November 2016www.griphyn.org Lightweight Replication of Heavyweight Data Scott Koranda University of Wisconsin-Milwaukee &
U.S. ATLAS Grid Production Experience
Pegasus and Condor Gaurang Mehta, Ewa Deelman, Carl Kesselman, Karan Vahi Center For Grid Technologies USC/ISI.
Software Implementation
Presentation transcript:

April 19, 2005Scott Koranda1 The LIGO Scientific Collaboration Data Grid Scott Koranda UW-Milwaukee On behalf of the LIGO Scientific Collaboration GridChem Workshop April 19, 2005 Urbana

April 19, 2005 Scott Koranda 2 Laser Interferometer Gravitational-wave Observatory LIGO is opening a new frontier in observational astrophysics  Detect & use gravitational waves (GW) to observe the Universe, provide a more complete picture of the Cosmos.  Complementary to radio/infrared/optical/X-ray/  -ray astronomy EM emitters not likely to be strong GW emitters & vice versa  Detect & observe cataclysmic events leading to death of stars, birth of neutron stars & black holes  Study Einstein’s theory of general relativity in the strong- field regime near massive compact objects, where GW are produced LIGO is now observing, acquiring science data and in full analysis production

April 19, 2005 Scott Koranda 3 Computation & Data Intensive Revealing the full science content of LIGO data is a computationally and data intensive challenge  LIGO interferometers generate ~ 10 MB/s or almost 1 TB/day Several classes of data analysis challenges require large-scale computational resources  In general for analysis 1. FFT data segment 2. Choose template (based on physical parameters) and filter 3. Repeat again and again and again… Sun E450 LIGO data file “frame” builder

April 19, 2005 Scott Koranda 4 Realizing the full science potential of LIGO Search for gravitational wave (GW) analogs of electromagnetic (EM) pulsars  GW sources not likely to have EM counterparts Fast (ms) EM pulsars are stable, old neutron stars  GW emission likely to come shortly after birth of a rapidly rotating (deformed, hot) NS  GW sky is unknown Searches will need to survey a large parameter space  All-sky blind search for periodic sources requires > FLOPS

April 19, 2005 Scott Koranda 5 LIGO Laboratory Hardware for Analysis CPU (GHz) Disk (TB)Tape (TB) Network() LHO OC3 LLO OC3 CIT GigE MIT FastE TOTAL

April 19, 2005 Scott Koranda 6 LIGO Scientific Collaboration (LSC) Tier 2 Sites 936 GHz CPU 34 TB RAID 5 storage OC-12 (622 Mbps) to Abilene

April 19, 2005 Scott Koranda 7 LIGO Scientific Collaboration (LSC) Tier 2 Sites 296 GHz CPU 64 TB storage (commodity IDE) for data OC-12 (622 Mbps) to Abilene NEWSFLASH: awarded $1.4M (NSF) + $0.3M (UW) for 3 TFlop cluster with 365 TB storage

April 19, 2005 Scott Koranda 8 GEO 600 Analysis Sites 272 GHz CPU 18 TB storage (RAID 5 and commodity IDE) GigE to SuperJANET (to GEANT to Abilene) 670 GHz CPU 40 TB storage (commodity IDE) Fast Ethernet to G-WiN (to GEANT to Abilene) 508 GHz CPU 18 TB storage (RAID5 and commodity IDE) 10 Gbps to SuperJANET (to GEANT to Abilene)

April 19, 2005 Scott Koranda 9 LSC DataGrid LSC Data Grid: 6 US sites + 3 EU sites (Birmingham, Cardiff, AEI)  Cardiff AEI/Golm Birmingham ~5400 GHz CPU

April 19, 2005 Scott Koranda 10 LSC: Partner in GriPhyN GriPhyN: Grid Physics Network  “GriPhyN research will enable the development of Petascale Virtual Data Grids (PVDGs)”  US CMS  US ATLAS  SDSS  LIGO Scientific Collaboration  Computer Scientists (Globus, Condor) Virtual Data Toolkit (VDT)  “one stop shopping for DataGrid software”  Virtual Data System (VDS)

April 19, 2005 Scott Koranda 11 LSC: Parnter in iVDGL/Grid3+ Persistent Grid! 4184 CPUs (and growing…) Virtual Organizations (VOs) and applications  US CMS  US ATLAS  SDSS  LSC  BTeV Other applications  SnB: Bio-molecular analysis  GADU/Gnare: Genome analysis

April 19, 2005 Scott Koranda 12 LSC: Partner in Open Science Grid

April 19, 2005 Scott Koranda 13 LSC: Partner in Open Science Grid

April 19, 2005 Scott Koranda 14 LSC Data Grid Focus in 3 areas 1. Development of Data Grid itself 2. LIGO & GEO interferomter data replication 3. Data analysis infrastructure

April 19, 2005 Scott Koranda 15 LSC Data Grid Focus in 3 areas 1. Development of Data Grid itself 2. LIGO & GEO interferomter data replication 3. Data analysis infrastructure

April 19, 2005 Scott Koranda 16 Building the LSC DataGrid Deploy server and client LSC DataGrid packages

April 19, 2005 Scott Koranda 17 Building the LSC DataGrid LSC DataGrid Client/Server packages Built on top of the VDT Packaged using Pacman from BU/ATLAS/iVDGL  cross platform packaging solution Fedora Core 3,4 Debian 3.1 (Sarge) Solaris 9,10  well suited for “Grid” middleware more flexible post- and pre-installation configuration VO level “control” over software stack

April 19, 2005 Scott Koranda 18 Building the LSC DataGrid description = 'LSC DataGrid Client 3.0.0' url = ' depends = [ 'LSC-DataGrid-Client-Environment', ' ' ' ' ' ' ' ' 'LSCdataFind', 'LSCcertUtil', 'LSC-VDS' ]

April 19, 2005 Scott Koranda 19 DOEGrids Certificate Authority LSC, through its partnership in large Grid projects, uses DOEGrids Certificate Authority for details of handing out digital certificates

April 19, 2005 Scott Koranda 20 LSC Data Grid Focus in 3 areas 1. Development of Data Grid itself 2. LIGO & GEO interferomter data replication 3. Data analysis infrastructure

April 19, 2005 Scott Koranda 21 LIGO/GEO Data Replication Replicate or copy data sets to analysis sites  Quickly  Securely  Robustly But also provide an infrastructure for answering  What data exists?  Where does it exists?  How do I get it? Answers needed for people and for analysis codes

April 19, 2005 Scott Koranda 22 Lightweight Data Replicator (LDR) Tie together 3 basic Grid services 1. Metadata Service info about files such as size, md5, GPS time,... sets of interesting metadata propagate answers question “What files or data are available?” 2. Globus Replica Location Service (RLS) catalog service maps filenames to URLs also maps filenames to sites answers question “Where are the files?” 3. GridFTP Service server and customized, tightly integrated client use to actually replicate files from site to site answers question “How do we move the files?”

TO Z LDR LHOCITMITLLOUWMPSUAEI“Publish” data Metadata Catalog H-RDS_R_L gwf bytes Frame type RDS_R_L3 Run tag S3 Locked … “Publish” data Metadata Catalog L-RDS_R_L gwf bytes Frame type RDS_R_L3 Run tag S3 Locked … What data do we want? Ask metadata catalog Collection: Instrument = ‘H’ AND frameType = ‘RDS_R_L3’ AND runTag = ‘S3’ Where can we get it? Ask URL catalog H-RDS_R_L gwf is available at LHO Local Replica Catalog H-RDS_R_L gwf → gsiftp://ldas.ligo-wa.caltech. edu:15000/samrds/S3/L3/LHO/H- RDS_R_L3-7526/H-RDS_R_L gwf Local Replica Catalog L-RDS_R_L gwf → gsiftp://ldas.ligo-la.caltech. edu:15000/samrds/S3/L3/LLO/L- RDS_R_L3-7526/L-RDS_R_L gwf “I have URLs for files…” URL Catalog What is URL for H-RDS_R_L gwf? gsiftp://ldas.ligo-wa.caltech.edu:15000/samrds/S3/L3/LHO/H-RDS_R_L3-7526/H-RDS_R_L gwf

April 19, 2005 Scott Koranda 24 Lightweight Data Replicator (LDR) Python – glue to hold it all together and make it robust PyGlobus Python API for Globus toolkit (3.2.x) Keith Jackson’s group at LBL LIGO/GEO S4 science run completed March  replicated over 30 TB in 30 days  mean time between failure now one month (outside of CIT)  over 30 million files in LDR network  pushed into Python performance limits now

April 19, 2005 Scott Koranda 25 Lightweight Data Replicator (LDR) Also provide infrastructure for data discovery Clients query for URLs based on metadata LSCdataFind --observatory H --type R -- gps-start-time gps-end- time url-type gsiftp gsiftp://dataserver.phys.uwm.edu:15000/data/gsiftp_root/c luster_storage/data/s001/S1/R/H/ /H- R gwf gsiftp://dataserver.phys.uwm.edu:15000/data/gsiftp_root/c luster_storage/data/s001/S1/R/H/ /H- R gwf

April 19, 2005 Scott Koranda 26 LSC Data Grid Focus in 3 areas 1. Development of Data Grid itself 2. LIGO & GEO interferomter data replication 3. Data analysis infrastructure

April 19, 2005 Scott Koranda 27 Growing LSC Grid Users We have a “recipe” for growing LSC DataGrid users and applications Start with developing applications on Linux destops  all current algorithms “pleasantly parallel”  most applications written in ANSI C  some Matlab (compiled)

April 19, 2005 Scott Koranda 28 Growing LSC Grid Users String together smaller, simple tools to create complex pipelines or workflows  ala the UNIX small tools model  output of one application is input of the next Eventually manage these complex workflows using Condor DAGman  DAG = Directed Acyclic Graph  describe parent-child relationships  DAGman runs child jobs after succesful parents jobs

April 19, 2005 Scott Koranda 29 Growing LSC Grid Users Workflow for inspiral search (neutron star, black hole, other compact objects) for a single 2048 sec stretch of data

April 19, 2005 Scott Koranda 30 Growing LSC Grid Users Develop tools to allow LSC scientists to easily create these workflows pipeline.py  python module with useful classes and object orientied data structures no GUI tools yet...  applications evolve too quickly the “application” becomes a DAG generator that ties together many smaller applications

April 19, 2005 Scott Koranda 31 Growing LSC Grid Users Users graduate to running on N of 8 Linux clusters  start on a single Linux cluster  debug any scaling issues (usually I/O)  Condor DAGman still used to manage workflows Still...too much effort lost managing jobs at N different sites

April 19, 2005 Scott Koranda 32 Growing LSC Grid Users Users complain to Scott and Duncan, “Why can’t we run across multiple clusters/sites all at once?” Scott and Duncan reply with a wry smile... “Add the flag --dax to your DAG generator script”

April 19, 2005 Scott Koranda 33 Growing LSC Grid Users DAX is a more generic description of workflow only applications and necessary files no hard-coded paths no job manager or service addresses XML (ready for machine processing) No change to applications or DAG/DAX generation script! Just add --dax flag

April 19, 2005 Scott Koranda 34 Growing LSC Grid Users Pegasus planner converts DAX to workflow for Grid search catalogs for locations of data search catalogs for locations of executables search catalogs for locations of resources plan the workflow for the LSC DataGrid  stage data if necessary  stage executables if necessary  output is a “concrete DAG”  run under DAGman as usual

April 19, 2005 Scott Koranda 35 Growing LSC Grid Users Linux desktop to the TeraFLOP...

April 19, 2005 Scott Koranda 36 Not that we don’t have growing pains... DataGrid infrastructure issues  Globus GRAM 2.x, 3.x limitations  Are they gone in GT 4.0? Data replication problems  Scaling issues for metadata service  Stability issues for Globus RLS recently solved a major stability problem Pegasus still a research product  only just realized application staging  “smart” scheduling of data transfers

April 19, 2005 Scott Koranda 37