LHCb The LHCb Data Management System Philippe Charpentier CERN On behalf of the LHCb Collaboration.

Slides:



Advertisements
Similar presentations
Andrew C. Smith, 11 th May 2007 EGEE User Forum 2 - DIRAC Data Management System User Forum 2 Data Management System “Tell me and I forget. Show me and.
Advertisements

Grid and CDB Janusz Martyniak, Imperial College London MICE CM37 Analysis, Software and Reconstruction.
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) gLite Grid Services Abderrahman El Kharrim
23/04/2008VLVnT08, Toulon, FR, April 2008, M. Stavrianakou, NESTOR-NOA 1 First thoughts for KM3Net on-shore data storage and distribution Facilities VLV.
Data Grid Web Services Chip Watson Jie Chen, Ying Chen, Bryan Hess, Walt Akers.
LHCb Quarterly Report October Core Software (Gaudi) m Stable version was ready for 2008 data taking o Gaudi based on latest LCG 55a o Applications.
1 Chapter 12 File Management Systems. 2 Systems Architecture Chapter 12.
ATLAS DQ2 Deletion Service D.A. Oleynik, A.S. Petrosyan, V. Garonne, S. Campana (on behalf of the ATLAS Collaboration)
Computing Infrastructure Status. LHCb Computing Status LHCb LHCC mini-review, February The LHCb Computing Model: a reminder m Simulation is using.
LHCb input to DM and SM TEGs. Remarks to DM and SM TEGS Introduction m We have already provided some input during our dedicated session of the TEG m Here.
Bookkeeping Tutorial. Bookkeeping & Monitoring Tutorial2 Bookkeeping content  Contains records of all “jobs” and all “files” that are created by production.
Event Data History David Adams BNL Atlas Software Week December 2001.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
MAGDA Roger Jones UCL 16 th December RWL Jones, Lancaster University MAGDA  Main authors: Wensheng Deng, Torre Wenaus Wensheng DengTorre WenausWensheng.
9 February 2000CHEP2000 Paper 3681 CDF Data Handling: Resource Management and Tests E.Buckley-Geer, S.Lammel, F.Ratnikov, T.Watts Hardware and Resources.
CCRC’08 Weekly Update Jamie Shiers ~~~ LCG MB, 1 st April 2008.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE middleware: gLite Data Management EGEE Tutorial 23rd APAN Meeting, Manila Jan.
Enabling Grids for E-sciencE Introduction Data Management Jan Just Keijser Nikhef Grid Tutorial, November 2008.
1 LCG-France sites contribution to the LHC activities in 2007 A.Tsaregorodtsev, CPPM, Marseille 14 January 2008, LCG-France Direction.
SRM workshop – September’05 1 SRM: Expt Reqts Nick Brook Revisit LCG baseline services working group Priorities & timescales Use case (from LHCb)
T3 analysis Facility V. Bucard, F.Furano, A.Maier, R.Santana, R. Santinelli T3 Analysis Facility The LHCb Computing Model divides collaboration affiliated.
1 LHCb on the Grid Raja Nandakumar (with contributions from Greig Cowan) ‏ GridPP21 3 rd September 2008.
1 LHCb File Transfer framework N. Brook, Ph. Charpentier, A.Tsaregorodtsev LCG Storage Management Workshop, 6 April 2005, CERN.
CHEP 2006, February 2006, Mumbai 1 LHCb use of batch systems A.Tsaregorodtsev, CPPM, Marseille HEPiX 2006, 4 April 2006, Rome.
Managing Data DIRAC Project. Outline  Data management components  Storage Elements  File Catalogs  DIRAC conventions for user data  Data operation.
INFSO-RI Enabling Grids for E-sciencE Introduction Data Management Ron Trompert SARA Grid Tutorial, September 2007.
Bookkeeping Tutorial. 2 Bookkeeping content  Contains records of all “jobs” and all “files” that are produced by production jobs  Job:  In fact technically.
David Adams ATLAS ATLAS distributed data management David Adams BNL February 22, 2005 Database working group ATLAS software workshop.
DIRAC Data Management: consistency, integrity and coherence of data Marianne Bargiotti CERN.
The new FTS – proposal FTS status. EMI INFSO-RI /05/ FTS /05/ /05/ Bugs fixed – Support an SE publishing more than.
Super Computing 2000 DOE SCIENCE ON THE GRID Storage Resource Management For the Earth Science Grid Scientific Data Management Research Group NERSC, LBNL.
The GridPP DIRAC project DIRAC for non-LHC communities.
David Adams ATLAS ATLAS-ARDA strategy and priorities David Adams BNL October 21, 2004 ARDA Workshop.
LHCb report to LHCC and C-RSG Philippe Charpentier CERN on behalf of LHCb.
DIRAC Project A.Tsaregorodtsev (CPPM) on behalf of the LHCb DIRAC team A Community Grid Solution The DIRAC (Distributed Infrastructure with Remote Agent.
Handling of T1D0 in CCRC’08 Tier-0 data handling Tier-1 data handling Experiment data handling Reprocessing Recalling files from tape Tier-0 data handling,
LHCbDirac and Core Software. LHCbDirac and Core SW Core Software workshop, PhC2 Running Gaudi Applications on the Grid m Application deployment o CVMFS.
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
LHCb status and plans Ph.Charpentier CERN. LHCb status and plans WLCG Workshop 1-2 Sept 2007, Victoria, BC 2 Ph.C. Status of DC06  Reminder:  Two-fold.
Dynamic staging to a CAF cluster Jan Fiete Grosse-Oetringhaus, CERN PH/ALICE CAF / PROOF Workshop,
CMS: T1 Disk/Tape separation Nicolò Magini, CERN IT/SDC Oliver Gutsche, FNAL November 11 th 2013.
Finding Data in ATLAS. May 22, 2009Jack Cranshaw (ANL)2 Starting Point Questions What is the latest reprocessing of cosmics? Are there are any AOD produced.
Markus Frank (CERN) & Albert Puig (UB).  An opportunity (Motivation)  Adopted approach  Implementation specifics  Status  Conclusions 2.
The GridPP DIRAC project DIRAC for non-LHC communities.
1 DIRAC Data Management Components A.Tsaregorodtsev, CPPM, Marseille DIRAC review panel meeting, 15 November 2005, CERN.
LHCb Computing activities Philippe Charpentier CERN – LHCb On behalf of the LHCb Computing Group.
A Data Handling System for Modern and Future Fermilab Experiments Robert Illingworth Fermilab Scientific Computing Division.
Breaking the frontiers of the Grid R. Graciani EGI TF 2012.
LHCb 2009-Q4 report Q4 report LHCb 2009-Q4 report, PhC2 Activities in 2009-Q4 m Core Software o Stable versions of Gaudi and LCG-AA m Applications.
EGEE-II INFSO-RI Enabling Grids for E-sciencE Architecture of LHC File Catalog Valeria Ardizzone INFN Catania – EGEE-II NA3/NA4.
Experiment Support CERN IT Department CH-1211 Geneva 23 Switzerland t DBES The Common Solutions Strategy of the Experiment Support group.
Joe Foster 1 Two questions about datasets: –How do you find datasets with the processes, cuts, conditions you need for your analysis? –How do.
LHCb LHCb GRID SOLUTION TM Recent and planned changes to the LHCb computing model Marco Cattaneo, Philippe Charpentier, Peter Clarke, Stefan Roiser.
CERN IT Department CH-1211 Genève 23 Switzerland t EGEE09 Barcelona ATLAS Distributed Data Management Fernando H. Barreiro Megino on behalf.
LHCb distributed computing during the LHC Runs 1,2 and 3
L’analisi in LHCb Angelo Carbone INFN Bologna
ATLAS Use and Experience of FTS
Introduction to Data Management in EGI
LHCb Computing Model and Data Handling Angelo Carbone 5° workshop italiano sulla fisica p-p ad LHC 31st January 2008.
Philippe Charpentier CERN – LHCb On behalf of the LHCb Computing Group
Evolution of the distributed computing model The case of CMS
Simulation use cases for T2 in ALICE
The INFN Tier-1 Storage Implementation
New developments on the LHCb Bookkeeping
R. Graciani for LHCb Mumbay, Feb 2006
LHCb Computing Philippe Charpentier CERN
Data Management in LHCb: consistency, integrity and coherence of data
DIRAC Data Management: consistency, integrity and coherence of data
Production Manager Tools (New Architecture)
The LHCb Computing Data Challenge DC06
Presentation transcript:

LHCb The LHCb Data Management System Philippe Charpentier CERN On behalf of the LHCb Collaboration

LHCb Data Management LHCb Computing Model in a nutshell m RAW files (3GB) are transferred to Tier0 o Verify migration and checksum m Transfer to Tier1s o Each file replicated once (whole run at a single site) m Reconstruction o At CERN and Tier1s (up to 300 HS06.hours) P If needed Tier2s can also be used as “Tier1 co-processor” m Stripping o On Tier1 where SDST is available m MC simulation o Complex workflow: simulation, digitization, reconstruction, trigger, filtering o Running everywhere with low priority P Tier2, Tier1, CERN and unpledged resources (some non-Grid) m User analysis o Running at Tier1s for data access, anywhere for MC studies m Grid activities under control of LHCbDirac o LHCb extensions of the DIRAC framework (cf Poster o Many presentations at CHEP (again this time, talks and posters) CHEP 2012, New York City Poster #272, S.Roiser Poster #145, F.Stagni

LHCb Data Management Real data applications m Reconstruction: o 50,000 events (3GB), =20 HS06.s/evt m Stripping: o 50,000 events (2.5 GB), =2 HS06.s/evt o 14 streams m Merge: o 40,000 events (5 GB) m Input files are downloaded on local disk CHEP 2012, New York City SDST Brunel (recons) Brunel (recons) DaVinci (stripping and streaming) DaVinci (stripping and streaming) RAW DST1 Merge DST1

LHCb Data Management Data replication m Replication performed using FTS o Active data placement m Split disk (for analysis) and archive o No disk replica for RAW and SDST (read few times) m Using SRM spaces (T1D0 and T0D1) CHEP 2012, New York City SDST Brunel (recons) Brunel (recons) DaVinci (stripping and streaming) DaVinci (stripping and streaming) RAW DST1 Merge DST1 1 Tier1 (scratch) CERN + 1 Tier1 CERN + 1 Tier1 CERN + 3 Tier1s

LHCb Data Management Datasets m Granularity at the file level o Data Management operations (replicate, remove replica, delete file) o Workload Management: input/output files of jobs m LHCbDirac perspective o DMS and WMS use LFNs to reference files o LFN namespace refers to the origin of the file P Constructed by the jobs (uses production and job number) P Hierarchical namespace for convenience P Used to define file class (tape-sets) for RAW, SDST, DST P GUID used for internal navigation between files (Gaudi) m User perspective o File is part of a dataset (consistent for physics analysis) o Dataset: specific conditions of data, processing version and processing level P Files in a dataset should be exclusive and consistent in quality and content CHEP 2012, New York City

LHCb Data Management Replica Catalog (1) m Logical namespace o Reflects somewhat the origin of the file (run number for RAW, production number for output files of jobs) o File type also explicit in the directory tree m Storage Elements o Essential component in the DIRAC DMS o Logical SEs: several DIRAC SEs can physically use the same hardware SE (same instance, same SRM space) o Described in the DIRAC configuration P Protocol, endpoint, port, SAPath, Web Service URL P Allows autonomous construction of the SURL  SURL = srm: : o SRM spaces at Tier1s P Used to have as many SRM spaces as DIRAC SEs, now only 3 P LHCb-Tape (T1D0) custodial storage P LHCb-Disk (T0D1) fast disk access P LHCb-User (T0D1) fast disk access for user data CHEP 2012, New York City

LHCb Data Management Replica Catalog (2) m Currently using the LFC o Master write service at CERN o Replication using Oracle streams to Tier1s o Read-only instances at CERN and Tier1s P Mostly for redundancy, no need for scaling m LFC information: o Metadata of the file o Replicas P Use “host name” field for the DIRAC SE name P Store SURL of creation for convenience (not used) d Allows lcg-util commands to work o Quality flag P One character comment used to set temporarily a replica as unavailable m Testing scalability of the DIRAC file catalog o Built-in storage usage capabilities (per directory) CHEP 2012, New York City Poster: A.Tsaregorodtsev

LHCb Data Management Bookkeeping Catalog (1) m User selection criteria o Origin of the data (real or MC, year of reference) P LHCb/Collision12 o Conditions for data taking of simulation (energy, magnetic field, detector configuration… P Beam4000GeV-VeloClosed-MagDown o Processing Pass is the level of processing (reconstruction, stripping…) including compatibility version P Reco13/Stripping19 o Event Type is mostly useful for simulation, single value for real data P 8 digit numeric code ( , ) o File Type defines which type of output files the user wants to get for a given processing pass (e.g. which stream) P RAW, SDST, BHADRON.DST (for a streamed file) m Bookkeeping search o Using a path  / / / / / CHEP 2012, New York City

LHCb Data Management Bookkeeping Catalog (2) m Much more than a dataset catalog! m Full provenance of files and jobs o Files are input of processing steps (“jobs”) that produce files o All files ever created are recorded, each processing step as well P Full information on the “job” (location, CPU, wall clock time…) m BK relational database o Two main tables: “files” and “jobs” o Jobs belong to a “production” o “Productions” belong to a “processing pass”, with a given “origin” and “condition” o Highly optimized search for files, as well as summaries m Quality flags o Files are immutable, but can have a mutable quality flag o Files have a flag indicating whether they have a replica or not CHEP 2012, New York City

LHCb Data Management Bookkeeping browsing m Allows to save datasets o Filter, selection o Plain list of files o Gaudi configuration file m Can return files with only replica at a given location CHEP 2012, New York City

LHCb Data Management Dataset based transformation m Same mechanism used for jobs (workload management tasks) and data management m Input datasets based on a bookkeeping query o Tasks are data driven, used by both WMS and DMS CHEP 2012, New York City

LHCb Data Management Data management transformations m Replication transformations o Uses a policy implementing the Computing Model o Creates transfer tasks depending on the original location and space availability o Replication using a tree (not all from the initial source) P Optimized w.r.t. channel recent bandwidth and SE locality o Transfers are whenever possible performed using FTS P If not possible, 3 rd party gridftp transfer d If no FTS channel or for user files (special credentials) o Replicas are automatically registered in the LFC m Removal transformations o Used for retiring datasets or reducing the number of replicas o Used exceptionally to completely remove files (tests, bugs…) o Replica removal protected against last replica removal! o Transfers and removal use the Data Manager credentials CHEP 2012, New York City

LHCb Data Management Staging: using files from tape m If jobs use files that are not online (on disk) o Before submitting the job o Stage the file from tape, and pin it on cache m Stager agent o Performs also cache management o Throttle staging requests depending on the cache size and the amount of pinned data o Requires fine tuning (pinning and cache size) P Caching architecture highly site dependent P No publication of cache sizes (except Castor and StoRM) m Jobs using staged files o Check first the file is still staged P If not reschedule the job o Copies the file locally on the WN whenever possible P Space is released faster P More reliable access for very long jobs (reconstruction) or jobs using many files (merging) CHEP 2012, New York City

LHCb Data Management Data Management Operational experience m Transfer accounting o Per site, channel, user… m Storage accounting o Per dataset, SE, user… o User quotas P Not strictly enforced… CHEP 2012, New York City

LHCb Data Management Outlook m Improvements on staging o Improve the tuning of cache settings P Depends on how caches are used by sites o Pinning/unpinning P Difficult if files are used by more than one job m Popularity o Record dataset usage P Reported by jobs: number of files used in a given dataset P Account number of files used per dataset per day/week o Assess dataset popularity P Relate usage to dataset size o Take decisions on the number of online replicas P Taking into account available space P Taking into account expected need in the coming weeks o First rely on Data Manager receiving an advice P Possibly move to automated dynamic management CHEP 2012, New York City

LHCb Data Management Conclusions m LHCb uses two views for data management: o Dataset view as seen by users and productions o Files view as seen by Data Management tools (replication, removal) m Datasets are handled by the LHCb Bookkeeping system (part of LHCbDirac) m File view is generic and handled by the DIRAC DMS m The LHCb DMS gives full flexibility for managing data on the Grid m In the future LHCb expect to use popularity criteria for deciding on the number of replicas for each dataset o Should give more flexibility to the Computing Model CHEP 2012, New York City