LIGO- G030537-01-E LIGO Grid Applications Development Within GriPhyN Albert Lazzarini LIGO Laboratory, Caltech GriPhyN All Hands.

Slides:



Advertisements
Similar presentations
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
Advertisements

Moving towards a hierarchical search. We now expand the coherent search to inspect a larger parameter space. (At the same time the incoherent stage is.
DataGrid Kimmo Soikkeli Ilkka Sormunen. What is DataGrid? DataGrid is a project that aims to enable access to geographically distributed computing power.
David Adams ATLAS DIAL Distributed Interactive Analysis of Large datasets David Adams BNL June 23, 2003 GAE workshop Caltech.
TeraGrid Gateway User Concept – Supporting Users V. E. Lynch, M. L. Chen, J. W. Cobb, J. A. Kohl, S. D. Miller, S. S. Vazhkudai Oak Ridge National Laboratory.
CONDOR DAGMan and Pegasus Selim Kalayci Florida International University 07/28/2009 Note: Slides are compiled from various TeraGrid Documentations.
LIGO Meeting LIGO Scientific Collaboration - University of Wisconsin - Milwaukee LIGO-G Z SuperComputing 2003 grid-distributed wide-area.
LIGO-G E ITR 2003 DMT Sub-Project John G. Zweizig LIGO/Caltech Argonne, May 10, 2004.
GEO-LIGO data analysis M.Alessandra Papa Max Planck Inst. f. Gravitationsphysik, Potsdam, Germany G Z VESF foundation meeting, Pisa, Dec
April 19, 2005Scott Koranda1 The LIGO Scientific Collaboration Data Grid Scott Koranda UW-Milwaukee On behalf of the LIGO Scientific Collaboration.
The Grid is a complex, distributed and heterogeneous execution environment. Running applications requires the knowledge of many grid services: users need.
LIGO-G9900XX-00-M LIGO and GriPhyN Carl Kesselman Ewa Deelman Sachin Chouksey Roy Williams Peter Shawhan Albert Lazzarini Kent Blackburn CaltechUSC/ISI.
Long Term Ecological Research Network Information System LTER Grid Pilot Study LTER Information Manager’s Meeting Montreal, Canada 4-7 August 2005 Mark.
Patrick R Brady University of Wisconsin-Milwaukee
Grid Computing - AAU 14/ Grid Computing Josva Kleist Danish Center for Grid Computing
G Z LIGO Scientific Collaboration Grid Patrick Brady University of Wisconsin-Milwaukee LIGO Scientific Collaboration.
OSG Area Coordinators’ Meeting LIGO Applications (NEW) Kent Blackburn Caltech / LIGO October 29 th,
3rd June 2004 CDF Grid SAM:Metadata and Middleware Components Mòrag Burgon-Lyon University of Glasgow.
4th EGEE user forum / OGF 25 Catania, TheoSSA on AstroGrid-D Iliya Nickelt (AIP / GAVO), Thomas Rauch (IAAT / GAVO), Harry Enke (AIP.
 The workflow description modified to output a VDS DAX.  The workflow description toolkit developed allows any concrete workflow description to be migrated.
Searching for gravitational radiation from Scorpius X-1: Limits from the second LIGO science run Alberto Vecchio on behalf of the LIGO Scientific Collaboration.
LIGO-G9900XX-00-M ITR 2003 DMT Sub-Project John G. Zweizig LIGO/Caltech.
Development Timelines Ken Kennedy Andrew Chien Keith Cooper Ian Foster John Mellor-Curmmey Dan Reed.
Grid Workload Management Massimo Sgaravatto INFN Padova.
10/24/2015OSG at CANS1 Open Science Grid Ruth Pordes Fermilab
Virtual Data Grid Architecture Ewa Deelman, Ian Foster, Carl Kesselman, Miron Livny.
José D. Zamora, Sean R. Morriss and Manuela Campanelli.
LIGO Z LIGO Scientific Collaboration -- UWM 1 LSC Data Analysis Alan G. Wiseman (LSC Software Coordinator) LIGO Scientific Collaboration.
Status of Standalone Inspiral Code Duncan Brown University of Wisconsin-Milwaukee LIGO Scientific Collaboration Inspiral Working Group LIGO-G Z.
Ames Research CenterDivision 1 Information Power Grid (IPG) Overview Anthony Lisotta Computer Sciences Corporation NASA Ames May 2,
Pegasus: Running Large-Scale Scientific Workflows on the TeraGrid Ewa Deelman USC Information Sciences Institute
What is Triana?. GAPGAP Triana Distributed Work-flow Network Action Commands Workflow, e.g. BPEL4WS Triana Engine Triana Controlling Service (TCS) Triana.
The GriPhyN Planning Process All-Hands Meeting ISI 15 October 2001.
LIGO-G E Generic LabelLIGO Laboratory at Caltech 1 Distributed Computing for LIGO Data Analysis The Aspen Winter Conference on Gravitational Waves,
Ruth Pordes November 2004TeraGrid GIG Site Review1 TeraGrid and Open Science Grid Ruth Pordes, Fermilab representing the Open Science.
High Energy Physics and Grids at UF (Dec. 13, 2002)Paul Avery1 University of Florida High Energy Physics.
May 6, 2002Earth System Grid - Williams The Earth System Grid Presented by Dean N. Williams PI’s: Ian Foster (ANL); Don Middleton (NCAR); and Dean Williams.
LIGO-G E LIGO Scientific Collaboration Data Grid Status Albert Lazzarini Caltech LIGO Laboratory Trillium Steering Committee Meeting 20 May 2004.
GWDAW-8 December 18, 2003LIGO Scientific Collaboration, UW - Milwaukee 1LIGO-G Z Broad-band CW searches (for isolated pulsars) in LIGO and GEO.
LIGO-G Z LIGO at the start of continuous observation Prospects and Challenges Albert Lazzarini LIGO Scientific Collaboration Presentation at NSF.
David Adams ATLAS DIAL: Distributed Interactive Analysis of Large datasets David Adams BNL August 5, 2002 BNL OMEGA talk.
TeraGrid Gateway User Concept – Supporting Users V. E. Lynch, M. L. Chen, J. W. Cobb, J. A. Kohl, S. D. Miller, S. S. Vazhkudai Oak Ridge National Laboratory.
Alain Roy Computer Sciences Department University of Wisconsin-Madison Condor & Middleware: NMI & VDT.
Pegasus: Planning for Execution in Grids Ewa Deelman, Carl Kesselman, Gaurang Mehta, Gurmeet Singh, Karan Vahi Information Sciences Institute University.
LIGO Plans for OSG J. Kent Blackburn LIGO Laboratory California Institute of Technology Open Science Grid Technical Meeting UCSD December 15-17, 2004.
State of LSC Data Analysis and Software LSC Meeting LIGO Hanford Observatory November 11 th, 2003 Kent Blackburn, Stuart Anderson, Albert Lazzarini LIGO.
Scott Koranda, UWM & NCSA 14 January 2016www.griphyn.org Lightweight Data Replicator Scott Koranda University of Wisconsin-Milwaukee & National Center.
LSC Meeting LIGO Scientific Collaboration - University of Wisconsin - Milwaukee 1 Software Coordinator Report Alan Wiseman LIGO-G Z.
Super Computing 2000 DOE SCIENCE ON THE GRID Storage Resource Management For the Earth Science Grid Scientific Data Management Research Group NERSC, LBNL.
LIGO-G Z LSC ASIS Meeting LIGO Scientific Collaboration - University of Wisconsin - Milwaukee 1 Update on November 2003 grid-distributed.
LIGO-G W Use of Condor by the LIGO Scientific Collaboration Gregory Mendell, LIGO Hanford Observatory On behalf of the LIGO Scientific Collaboration.
LIGO-G Z1 Using Condor for Large Scale Data Analysis within the LIGO Scientific Collaboration Duncan Brown California Institute of Technology.
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
LIGO: The Laser Interferometer Gravitational-wave Observatory Sarah Caudill Louisiana State University Physics Graduate Student LIGO Hanford LIGO Livingston.
Collaborative Tools for the Grid V.N Alexandrov S. Mehmood Hasan.
PARALLEL AND DISTRIBUTED PROGRAMMING MODELS U. Jhashuva 1 Asst. Prof Dept. of CSE om.
1 USC Information Sciences InstituteYolanda Gil AAAI-08 Tutorial July 13, 2008 Part IV Workflow Mapping and Execution in Pegasus (Thanks.
Summary of OSG Activities by LIGO and LSC LIGO NSF Review November 9-11, 2005 Kent Blackburn LIGO Laboratory California Institute of Technology LIGO DCC:
Managing LIGO Workflows on OSG with Pegasus Karan Vahi USC Information Sciences Institute
VO Experiences with Open Science Grid Storage OSG Storage Forum | Wednesday September 22, 2010 (10:30am)
INTRODUCTION TO XSEDE. INTRODUCTION  Extreme Science and Engineering Discovery Environment (XSEDE)  “most advanced, powerful, and robust collection.
Pegasus WMS Extends DAGMan to the grid world
U.S. ATLAS Grid Production Experience
on behalf of the LIGO Scientific Collaboration
Pegasus and Condor Gaurang Mehta, Ewa Deelman, Carl Kesselman, Karan Vahi Center For Grid Technologies USC/ISI.
Broad-band CW searches in LIGO and GEO S2 and S3 data
Broad-band CW searches in LIGO and GEO S2 and S3 data
LIGO Scientific Collaboration, UW - Milwaukee
Broad-band CW searches in LIGO and GEO S2 and S3 data
Frieda meets Pegasus-WMS
Presentation transcript:

LIGO- G E LIGO Grid Applications Development Within GriPhyN Albert Lazzarini LIGO Laboratory, Caltech GriPhyN All Hands Meeting October 2003 ANL

LIGO- G E Oct 2003 Albert Lazzarini, LIGO Laboratory, Caltech LIGO Applications Development Team (GriPhyN) l LIGO Scientific Collaboration –LIGO Laboratory –UWM –AEI (Germany) l ISI

LIGO- G E Oct 2003 Albert Lazzarini, LIGO Laboratory, Caltech LIGO and Grid Computing l Revealing the full science content of LIGO data is a computationally and data intense challenge –Several classes of data analysis challenges require large-scale computational resources u Search for gravitational wave (GW) analogs of electromagnetic (EM) pulsars u GW sources not likely to have EM counterparts u Fast (millisecond) EM pulsars are stable, old neutron stars (NS) u GW emission likely to come shortly after birth of a rapidly rotating (deformed, hot) NS u GW sky is unknown u Searches will need to survey a large parameter space u All-sky search for previously unidentified periodic sources requires > floating point operations per second (FLOPS) u These analyses are ideally suited for distributed (grid-based) computing u Coalescence of compact binary systems (“inspiral chirps”) which include spin-spin interactions will cover a huge parameter space (~ 10 6 greater than spinless systems) u Important for more massive systems u Massive systems have greater GW luminosities u Likely to be the first detected

LIGO- G E Oct 2003 Albert Lazzarini, LIGO Laboratory, Caltech Efficient search of entire sky for unknown GW pulsars (with no EM counterpart) requires > FLOPS # pixels on sky vs. T observation #FLOPS vs. T observation Figures taken from Brady et al., Phys  Rev  D 

LIGO- G E Oct 2003 Albert Lazzarini, LIGO Laboratory, Caltech LIGO GriPhyN Application: Grid-deployed production-level GW pulsar search The pulsar search conducted at SC TOY DEMO Used LIGO data subset collected during the first scientific run of the instrument Targeted a set of 1000 locations including both known EM pulsar as well as random locations in the sky Results of the analysis are available via LDAS (LIGO Data Analysis System) DB queries Performed using LDAS and compute and storage resources at Caltech, University of Southern California, University of Wisconsin Milwaukee. SC 2002 Over 58 directions in the sky searched Total of 330 tasks 469 data transfers 330 output files produced. The total runtime:11.4 hr

LIGO- G E Oct 2003 Albert Lazzarini, LIGO Laboratory, Caltech Goal: Use SC2003 to initiate a production analysis run for GW pulsars l Stand-alone search codes to be used for the SC2003 –Perform frequency-time transformations of data –Parameters of transform depend on {,,( f, f ‘, f “,…)} l Code originally developed by GEO project as part of LIGO collaboration l The code has been modified to enable a user to specify input and output file names and parameters on the command line –Stage data to/from compute resources as needed l Auto generator will approximately search the galactic core for pulsars –Specifies the range of parameters to be search l Results will be shown on a 3D Visualization

LIGO- G E Oct 2003 Albert Lazzarini, LIGO Laboratory, Caltech Grid-deployed production-level GW pulsar search l Use SC2003 to implement production-level scalable version of SC2002 demonstration Original goal: Implement a production GW pulsar search over a patch (less than 4 of sky) search running for ~30 days on O[10x] more resources than LIGO has => use the grid (10000 CPUs for 1 month) –Subscale implementation using ONLY LIGO-owned resources in development –Joint development with ISI CS team. l As additional resources become available, will scale production run to expand beyond LIGO resources.

LIGO- G E Oct 2003 Albert Lazzarini, LIGO Laboratory, Caltech LIGO Grid Sites LIGO Grid: 6 US sites + 2 EU sites (Cardiff/UK, AEI/Germany) Collaboration: 35+ institutions world wide; 400 members  Cardiff  AEI/Golm

LIGO- G E Oct 2003 Albert Lazzarini, LIGO Laboratory, Caltech SC2003 production run l Sites to be used ISI Condor and LSF Clusters (035 CPUs) LIGO CALTECH condor pool (220 CPUs) UWM condor pool (300 CPUs) AEI (Germany) condor pool (300 CPUs)  Cardiff condor pool(046 CPUs)  UW Madison condor pool(800 CPUs) ?Teragrid resources(~1024 CPUs) ?Grid3 resources

LIGO- G E Oct 2003 Albert Lazzarini, LIGO Laboratory, Caltech Implementation details l Web portal based authentication using MyProxy and request submission. l Resource discovery and information using Globus MDS l S2 science run data from 3 interferometers to be used l Each input file contains data for 1800s approximately 25Mb in size l Executable staging for sites which don’t have geo-code installed l Site monitoring being done using GANGLIA l Globus Replica Location Service used for registering raw and materialized data products

LIGO- G E Oct 2003 Albert Lazzarini, LIGO Laboratory, Caltech GW Pulsar Search - Dag Structure Input transfer Search Execution nodes Interpool transfer Output transfer RLS Registration ComputeFstatistic ExtractSFT Virtual Data Threshold & Detection

LIGO- G E Oct 2003 Albert Lazzarini, LIGO Laboratory, Caltech Status l First dry run week of for testing out the search code and the SC demo/production infrastructure. l The jobs were run at the ISI condor pool, ISI lsf cluster, CALTECH condor pool and UWM condor pool l The outputs transferred to the ISI condor pool l The jobs breakdown is as follows –11 input transfer jobs (bringing in 1002 input files) ( 1000 raw files and 2 calibration files earth.dat and sun.dat) –10 Extract SFT jobs each crunching on 100 input files –1 ComputeFstat job which takes the 1000 extracted SFT files and computes fstatistic. –1 interpool transfer job which transfers data from runs at CALTECH, UWM and ISI LSF cluster to the ISI condor pool (The computeFstat was running at ISI LSF cluster) –11 output transfers (All the output data was stored in the storage location associated with the ISI condor-pool) –11 RLS registration jobs ( All data was registered at rls://smarty.isi.edu associated with the ISI condor-pool)

LIGO- G E Oct 2003 Albert Lazzarini, LIGO Laboratory, Caltech GW Pulsar Search - Dag Over Time

LIGO- G E Oct 2003 Albert Lazzarini, LIGO Laboratory, Caltech Run time Dag Visualization Visualization code by Gregor von Laszewski, and Michael Hategan (ANL)

LIGO- G E Oct 2003 Albert Lazzarini, LIGO Laboratory, Caltech LIGO Science Has Started l First Science run (“S1”): –First LIGO scientific results submitted for publication l Second science run (“S2”): –Sensitivity was ~10x better than first run –Duration was ~ 4x longer l Third science run (“S3”): l LIGO is analyzing real data today with emerging grid technologies –Need to balance priorities: >data analysis challenges vs. grid R&D challenges –Opportunity to provide “case study” feedback to middleware development activities >robustness, QA >timeliness -- tracking of opensource software releases (e.g., linux kernels, …) >what works, what does not work, what needs improvement, …

LIGO- G E Oct 2003 Albert Lazzarini, LIGO Laboratory, Caltech GW Pulsar Search - Dag Over Time by Hosts