A Data Handling System for Modern and Future Fermilab Experiments Robert Illingworth Fermilab Scientific Computing Division.

Slides:



Advertisements
Similar presentations
GridPP July 2003Stefan StonjekSlide 1 SAM middleware components Stefan Stonjek University of Oxford 7 th GridPP Meeting 02 nd July 2003 Oxford.
Advertisements

Spark: Cluster Computing with Working Sets
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
6/4/20151 Introduction LHCb experiment. LHCb experiment. Common schema of the LHCb computing organisation. Common schema of the LHCb computing organisation.
23/04/2008VLVnT08, Toulon, FR, April 2008, M. Stavrianakou, NESTOR-NOA 1 First thoughts for KM3Net on-shore data storage and distribution Facilities VLV.
Data Grids: Globus vs SRB. Maturity SRB  Older code base  Widely accepted across multiple communities  Core components are tightly integrated Globus.
Oxford Jan 2005 RAL Computing 1 RAL Computing Implementing the computing model: SAM and the Grid Nick West.
Slide 1 of 9 Presenting 24x7 Scheduler The art of computer automation Press PageDown key or click to advance.
TIBCO Designer TIBCO BusinessWorks is a scalable, extensible, and easy to use integration platform that allows you to develop, deploy, and run integration.
The SAM-Grid Fabric Services Gabriele Garzoglio (for the SAM-Grid team) Computing Division Fermilab.
Hall D Online Data Acquisition CEBAF provides us with a tremendous scientific opportunity for understanding one of the fundamental forces of nature. 75.
Experiences Deploying Xrootd at RAL Chris Brew (RAL)
CERN - IT Department CH-1211 Genève 23 Switzerland t Monitoring the ATLAS Distributed Data Management System Ricardo Rocha (CERN) on behalf.
Connecting OurGrid & GridSAM A Short Overview. Content Goals OurGrid: architecture overview OurGrid: short overview GridSAM: short overview GridSAM: example.
Test Of Distributed Data Quality Monitoring Of CMS Tracker Dataset H->ZZ->2e2mu with PileUp - 10,000 events ( ~ 50,000 hits for events) The monitoring.
Workload Management WP Status and next steps Massimo Sgaravatto INFN Padova.
The SAMGrid Data Handling System Outline:  What Is SAMGrid?  Use Cases for SAMGrid in Run II Experiments  Current Operational Load  Stress Testing.
LcgCAF:CDF submission portal to LCG Federica Fanzago for CDF-Italian Computing Group Gabriele Compostella, Francesco Delli Paoli, Donatella Lucchesi, Daniel.
Flexibility and user-friendliness of grid portals: the PROGRESS approach Michal Kosiedowski
MapReduce: Hadoop Implementation. Outline MapReduce overview Applications of MapReduce Hadoop overview.
03/27/2003CHEP20031 Remote Operation of a Monte Carlo Production Farm Using Globus Dirk Hufnagel, Teela Pulliam, Thomas Allmendinger, Klaus Honscheid (Ohio.
November 7, 2001Dutch Datagrid SARA 1 DØ Monte Carlo Challenge A HEP Application.
Building a distributed software environment for CDF within the ESLEA framework V. Bartsch, M. Lancaster University College London.
CDF Grid Status Stefan Stonjek 05-Jul th GridPP meeting / Durham.
3rd June 2004 CDF Grid SAM:Metadata and Middleware Components Mòrag Burgon-Lyon University of Glasgow.
QCDGrid Progress James Perry, Andrew Jackson, Stephen Booth, Lorna Smith EPCC, The University Of Edinburgh.
CHEP'07 September D0 data reprocessing on OSG Authors Andrew Baranovski (Fermilab) for B. Abbot, M. Diesburg, G. Garzoglio, T. Kurca, P. Mhashilkar.
1 st December 2003 JIM for CDF 1 JIM and SAMGrid for CDF Mòrag Burgon-Lyon University of Glasgow.
Interactive Job Monitor: CafMon kill CafMon tail CafMon dir CafMon log CafMon top CafMon ps LcgCAF: CDF submission portal to LCG resources Francesco Delli.
LCG Middleware Testing in 2005 and Future Plans E.Slabospitskaya, IHEP, Russia CERN-Russia Joint Working Group on LHC Computing March, 6, 2006.
GridPP18 Glasgow Mar 07 DØ – SAMGrid Where’ve we come from, and where are we going? Evolution of a ‘long’ established plan Gavin Davies Imperial College.
Frontiers in Massive Data Analysis Chapter 3.  Difficult to include data from multiple sources  Each organization develops a unique way of representing.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
Datasets on the GRID David Adams PPDG All Hands Meeting Catalogs and Datasets session June 11, 2003 BNL.
November SC06 Tampa F.Fanzago CRAB a user-friendly tool for CMS distributed analysis Federica Fanzago INFN-PADOVA for CRAB team.
Tier-2  Data Analysis  MC simulation  Import data from Tier-1 and export MC data CMS GRID COMPUTING AT THE SPANISH TIER-1 AND TIER-2 SITES P. Garcia-Abia.
22 nd September 2003 JIM for CDF 1 JIM and SAMGrid for CDF Mòrag Burgon-Lyon University of Glasgow.
And Tier 3 monitoring Tier 3 Ivan Kadochnikov LIT JINR
EGEE is a project funded by the European Union under contract IST HEP Use Cases for Grid Computing J. A. Templon Undecided (NIKHEF) Grid Tutorial,
What is SAM-Grid? Job Handling Data Handling Monitoring and Information.
Data reprocessing for DZero on the SAM-Grid Gabriele Garzoglio for the SAM-Grid Team Fermilab, Computing Division.
Metadata Mòrag Burgon-Lyon University of Glasgow.
CASTOR evolution Presentation to HEPiX 2003, Vancouver 20/10/2003 Jean-Damien Durand, CERN-IT.
1 Andrea Sciabà CERN Critical Services and Monitoring - CMS Andrea Sciabà WLCG Service Reliability Workshop 26 – 30 November, 2007.
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
David Adams ATLAS DIAL: Distributed Interactive Analysis of Large datasets David Adams BNL August 5, 2002 BNL OMEGA talk.
Integration of the ATLAS Tag Database with Data Management and Analysis Components Caitriana Nicholson University of Glasgow 3 rd September 2007 CHEP,
UTA MC Production Farm & Grid Computing Activities Jae Yu UT Arlington DØRACE Workshop Feb. 12, 2002 UTA DØMC Farm MCFARM Job control and packaging software.
Super Computing 2000 DOE SCIENCE ON THE GRID Storage Resource Management For the Earth Science Grid Scientific Data Management Research Group NERSC, LBNL.
Data Transfer Service Challenge Infrastructure Ian Bird GDB 12 th January 2005.
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI Monitoring of the LHC Computing Activities Key Results from the Services.
Chapter 9: Web Services and Databases Title: NiagaraCQ: A Scalable Continuous Query System for Internet Databases Authors: Jianjun Chen, David J. DeWitt,
T3g software services Outline of the T3g Components R. Yoshida (ANL)
Distributed Physics Analysis Past, Present, and Future Kaushik De University of Texas at Arlington (ATLAS & D0 Collaborations) ICHEP’06, Moscow July 29,
STAR Scheduler Gabriele Carcassi STAR Collaboration.
CMS: T1 Disk/Tape separation Nicolò Magini, CERN IT/SDC Oliver Gutsche, FNAL November 11 th 2013.
Simulation Production System Science Advisory Committee Meeting UW-Madison March 1 st -2 nd 2007 Juan Carlos Díaz Vélez.
SAM architecture EGEE 07 Service Availability Monitor for the LHC experiments Simone Campana, Alessandro Di Girolamo, Nicolò Magini, Patricia Mendez Lorenzo,
Data Infrastructure in the TeraGrid Chris Jordan Campus Champions Presentation May 6, 2009.
Fabric for Frontier Experiments at Fermilab Gabriele Garzoglio Grid and Cloud Services Department, Scientific Computing Division, Fermilab ISGC – Thu,
DØ Grid Computing Gavin Davies, Frédéric Villeneuve-Séguier Imperial College London On behalf of the DØ Collaboration and the SAMGrid team The 2007 Europhysics.
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI EGI solution for high throughput data analysis Peter Solagna EGI.eu Operations.
Simulation Production System
Platform Overview Provide your marketing and sales groups with a single, integrated, web based on-demand platform that allows them to easily automate and.
Data Bridge Solving diverse data access in scientific applications
Self Healing and Dynamic Construction Framework:
David Adams Brookhaven National Laboratory September 28, 2006
Survey on User’s Computing Experience
DØ MC and Data Processing on the Grid
MapReduce: Simplified Data Processing on Large Clusters
Presentation transcript:

A Data Handling System for Modern and Future Fermilab Experiments Robert Illingworth Fermilab Scientific Computing Division

Why is data movement difficult? Your files don’t reside where your jobs run 2 Tape RobotDisk Storage Your files reside hereNot here, where you need them Compute Farm of worker nodes Files must be moved to where a worker node can access them with efficiency and scalability Output files must be returned from worker node NOT a simple process, especially the efficiency and scalability part! CHEP Robert Illingworth

Supporting many experiments at Fermilab Fermilab supports multiple experiments which require data handling services – scale ranging from 10s TB/year up to multiple PB/year To minimize support requirements, need a common solution – It’s not practical for every experiment to go off on its own way Evaluated various options, including adopting existing system or implementing entirely new one – Some existing systems are tightly coupled to experiment computing model. Others are too general. Decided to adopt and modernize the existing Fermilab SAM system from Run II of the Tevatron CHEP Robert Illingworth

The FIFE project Physics task Grid Jobs Databases (calib/conditi ons,…) Files DB Applications Data handling Output Files 4 FIFE is Fermilab’s overarching program of common tools and systems for scientific data processing (see yesterday’s presentation by Mike Kirby) Job submission Rolling out to Minerνa, NOνA, Muon g- 2, Mu2e, LBNE, MicroBoone, Argoneut, Darkside CHEP Robert Illingworth

Analysis dataflow (user view) Take raw data Reconstruct it Analyze reconstructed data Publish Reconstruction Analysis Published paper Experiment Raw data Reconstructed data CHEP Robert Illingworth

Analysis dataflow (more realistic view) Not usually a nice linear progression Even a small HEP experiment has millions of data files to keep track of and feed the correct ones into the next stage of processing Reconstruction pass 1 Reconstruction pass 1 Analysis (repeat many times) Analysis (repeat many times) Published paper Experiment Raw data Reconstructed data Disk storage Reconstruction pass n Reconstruction pass n Tape storage Tape storage 6 Simulation Monte- Carlo CHEP Robert Illingworth

User requirements of a data management system What is in my data? Give me my data What did I do with my data? Store my output data safely CHEP Robert Illingworth

User requirements of a data management system What is in my data? – Metadata catalogue Give me my data – File access service that turns a dataset into concrete locations What did I do with my data? – Extensive tracking of job activity; what files were provided to a job; what did the job report was successfully processed Store my output data safely – Reliably add output files to catalogue and archive them on disk or tape CHEP Robert Illingworth

SAM data management SAM (Sequential Access via Metadata) was originally begun as a data handling system for Run II of the Tevatron Used very successfully by CDF and D0 (metadata for over 207 million files; over 600 physics papers published) Now modernizing the architecture for the next decade of operations – build on extensive past experience – make use of modern technologies CHEP Robert Illingworth

Metadata catalogue Metadata catalogue stores information about the files – Physical data (size) and physics data – Experiments can define their own fields; not restricted to a predefined set Example raw file from NOνA on right CHEP Robert Illingworth

Making datasets Individual files are combined into datasets Datasets are defined by a simple query language, for example “run_number and file_format raw” Users can create their own datasets; they aren’t confined to predefined ones Datasets can be dynamic; new files that match the criteria are included each time they are evaluated CHEP Robert Illingworth

Running jobs Jobs only need the dataset name as input; no need for users to know where the files are physically located The files will be staged from tape if needed Multiple jobs can be run on a single dataset. The files will be dynamically distributed across all the jobs CHEP Robert Illingworth

Tracking and recovery Each job updates the database as it opens and closes each file – Provides a history of the entire job This allows simple recovery of failures; make a new data set that contains: – Files not provided due to storage system problems – Files which were processed but the output was lost (job crashed, was evicted, transferring the output failed, etc.) CHEP Robert Illingworth

Monitoring of processing activity CHEP Robert Illingworth

Operational requirements Designed as a highly-automated, “hands-off” system requiring minimal routine intervention – Ideal for the “smaller” (by LHC standards) collaborations, which can’t provide dedicated expertise for operating data management systems Data staging from tape and from storage element to storage element is automatic – Philosophy is “bring the data to the jobs”, not “bring the jobs to the data” CHEP Robert Illingworth

SAM as part of FIFE Part of an integrated system – FIFE – with other components – Jobsub glideinWMS based job submission – art framework – dCache disk cache But loosely coupled (via http REST interfaces), so can easily substitute other services CHEP Robert Illingworth

Job submission integration The FIFE job submission system (jobsub) takes the SAM dataset name when the job is submitted It notifies SAM that the specified dataset is needed, and passes contact information to the individual jobs CHEP Robert Illingworth

Storage integration SAM acts as a “high level” manager; leaves the low level details of handling storage to other systems (dCache, SRM, etc) – can command prestaging from tape and transfers from system to system – because SAM works with entire datasets it knows which files will be used before they are accessed; allows efficient automated prestaging Can work with multiple file access methods: – gridftp – xrootd – direct local access – …anything that can be expressed as a URL CHEP Robert Illingworth

Framework integration art framework modules provide integration with SAM Input module provides next file from SAM – SAM provided files are just another source of events to process Metadata handling stores metadata within the output file – Makes cataloguing the file easy; just read the metadata out and add it to the database CHEP Robert Illingworth

Putting it all together Tape store Disk cache Storage element SAM Database art executable Analysis job on worker node dCache Job file access (gridftp, xrootd) File transfer to SE (gridftp, SRM, iRODS); Federated xrootd Request staging from tape Request transfer to SE Request for next file Jobsu b Specify dataset Start jobs CHEP Robert Illingworth

Current and future deployment In full use – Minos (~100 TB/yr) – Minerνa (~10 TB/yr) – NOνA (~PB/yr) Deploying – MicroBoone (~100 TB/yr) – Darkside (~100 TB/yr) – LBNE (many PB/yr) Planned – g-2 (~1 PB total) CHEP Robert Illingworth

Conclusion SAM is the data handling system used by multiple current and future Fermilab experiments Builds on long experience from the Tevatron; updated to use modern technologies Experiment agnostic design provides flexibility High level of automation requires low level of administrative effort from experiments CHEP Robert Illingworth