LHCb Computing Philippe Charpentier CERN

Slides:



Advertisements
Similar presentations
31/03/00 CMS(UK)Glenn Patrick What is the CMS(UK) Data Model? Assume that CMS software is available at every UK institute connected by some infrastructure.
Advertisements

Status GridKa & ALICE T2 in Germany Kilian Schwarz GSI Darmstadt.
23/04/2008VLVnT08, Toulon, FR, April 2008, M. Stavrianakou, NESTOR-NOA 1 First thoughts for KM3Net on-shore data storage and distribution Facilities VLV.
Large scale data flow in local and GRID environment V.Kolosov, I.Korolko, S.Makarychev ITEP Moscow.
1 Managing distributed computing resources with DIRAC A.Tsaregorodtsev, CPPM-IN2P3-CNRS, Marseille September 2011, NEC’11, Varna.
LHCb Quarterly Report October Core Software (Gaudi) m Stable version was ready for 2008 data taking o Gaudi based on latest LCG 55a o Applications.
Petabyte-scale computing challenges of the LHCb experiment UK e-Science All Hands Meeting 2008, Edinburgh, 9 th September 2008 Y. Y. Li on behalf of the.
Computing Infrastructure Status. LHCb Computing Status LHCb LHCC mini-review, February The LHCb Computing Model: a reminder m Simulation is using.
LHCb computing in Russia Ivan Korolko (ITEP Moscow) Russia-CERN JWGC, October 2005.
ATLAS and GridPP GridPP Collaboration Meeting, Edinburgh, 5 th November 2001 RWL Jones, Lancaster University.
Bookkeeping Tutorial. Bookkeeping & Monitoring Tutorial2 Bookkeeping content  Contains records of all “jobs” and all “files” that are created by production.
F. Fassi, S. Cabrera, R. Vives, S. González de la Hoz, Á. Fernández, J. Sánchez, L. March, J. Salt, A. Lamas IFIC-CSIC-UV, Valencia, Spain Third EELA conference,
Status of the LHCb MC production system Andrei Tsaregorodtsev, CPPM, Marseille DataGRID France workshop, Marseille, 24 September 2002.
November SC06 Tampa F.Fanzago CRAB a user-friendly tool for CMS distributed analysis Federica Fanzago INFN-PADOVA for CRAB team.
4/5/2007Data handling and transfer in the LHCb experiment1 Data handling and transfer in the LHCb experiment RT NPSS Real Time 2007 FNAL - 4 th May 2007.
1 LCG-France sites contribution to the LHC activities in 2007 A.Tsaregorodtsev, CPPM, Marseille 14 January 2008, LCG-France Direction.
LHCb The LHCb Data Management System Philippe Charpentier CERN On behalf of the LHCb Collaboration.
T3 analysis Facility V. Bucard, F.Furano, A.Maier, R.Santana, R. Santinelli T3 Analysis Facility The LHCb Computing Model divides collaboration affiliated.
The LHCb CERN R. Graciani (U. de Barcelona, Spain) for the LHCb Collaboration International ICFA Workshop on Digital Divide Mexico City, October.
1 LHCb on the Grid Raja Nandakumar (with contributions from Greig Cowan) ‏ GridPP21 3 rd September 2008.
Grid User Interface for ATLAS & LHCb A more recent UK mini production used input data stored on RAL’s tape server, the requirements in JDL and the IC Resource.
1 LHCb File Transfer framework N. Brook, Ph. Charpentier, A.Tsaregorodtsev LCG Storage Management Workshop, 6 April 2005, CERN.
CHEP 2006, February 2006, Mumbai 1 LHCb use of batch systems A.Tsaregorodtsev, CPPM, Marseille HEPiX 2006, 4 April 2006, Rome.
The ATLAS Cloud Model Simone Campana. LCG sites and ATLAS sites LCG counts almost 200 sites. –Almost all of them support the ATLAS VO. –The ATLAS production.
Bookkeeping Tutorial. 2 Bookkeeping content  Contains records of all “jobs” and all “files” that are produced by production jobs  Job:  In fact technically.
Testing and integrating the WLCG/EGEE middleware in the LHC computing Simone Campana, Alessandro Di Girolamo, Elisa Lanciotti, Nicolò Magini, Patricia.
Outline: The LHCb Computing Model Philippe Charpentier, CERN ICFA workshop on Grid activities, Sinaia, Romania, October
Large scale data flow in local and GRID environment Viktor Kolosov (ITEP Moscow) Ivan Korolko (ITEP Moscow)
LHCb report to LHCC and C-RSG Philippe Charpentier CERN on behalf of LHCb.
DIRAC Project A.Tsaregorodtsev (CPPM) on behalf of the LHCb DIRAC team A Community Grid Solution The DIRAC (Distributed Infrastructure with Remote Agent.
1 LHCb computing for the analysis : a naive user point of view Workshop analyse cc-in2p3 17 avril 2008 Marie-Hélène Schune, LAL-Orsay for LHCb-France Framework,
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
LHCb status and plans Ph.Charpentier CERN. LHCb status and plans WLCG Workshop 1-2 Sept 2007, Victoria, BC 2 Ph.C. Status of DC06  Reminder:  Two-fold.
Meeting with University of Malta| CERN, May 18, 2015 | Predrag Buncic ALICE Computing in Run 2+ P. Buncic 1.
LHCb Computing activities Philippe Charpentier CERN – LHCb On behalf of the LHCb Computing Group.
Breaking the frontiers of the Grid R. Graciani EGI TF 2012.
LHCb 2009-Q4 report Q4 report LHCb 2009-Q4 report, PhC2 Activities in 2009-Q4 m Core Software o Stable versions of Gaudi and LCG-AA m Applications.
1 LHCb Computing A.Tsaregorodtsev, CPPM, Marseille 14 March 2007, Clermont-Ferrand.
DIRAC for Grid and Cloud Dr. Víctor Méndez Muñoz (for DIRAC Project) LHCb Tier 1 Liaison at PIC EGI User Community Board, October 31st, 2013.
1 S. JEZEQUEL- First chinese-french workshop 13 December 2006 Grid: An LHC user point of vue S. Jézéquel (LAPP-CNRS/Université de Savoie)
LHCb LHCb GRID SOLUTION TM Recent and planned changes to the LHCb computing model Marco Cattaneo, Philippe Charpentier, Peter Clarke, Stefan Roiser.
CERN IT Department CH-1211 Genève 23 Switzerland t EGEE09 Barcelona ATLAS Distributed Data Management Fernando H. Barreiro Megino on behalf.
LHCb Computing 2015 Q3 Report Stefan Roiser LHCC Referees Meeting 1 December 2015.
1-2 March 2006 P. Capiluppi INFN Tier1 for the LHC Experiments: ALICE, ATLAS, CMS, LHCb.
ATLAS – statements of interest (1) A degree of hierarchy between the different computing facilities, with distinct roles at each level –Event filter Online.
Lessons learned administering a larger setup for LHCb
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI EGI solution for high throughput data analysis Peter Solagna EGI.eu Operations.
LHCb D ata P rocessing S oftware J. Blouw, A. Zhelezov Physikalisches Institut, Universitaet Heidelberg DESY Computing Seminar, Nov. 29th, 2010.
LHCb distributed computing during the LHC Runs 1,2 and 3
WLCG Tier-2 Asia Workshop TIFR, Mumbai 1-3 December 2006
L’analisi in LHCb Angelo Carbone INFN Bologna
INFN GRID Workshop Bari, 26th October 2004
Data Challenge with the Grid in ATLAS
Database Readiness Workshop Intro & Goals
The LHCb Software and Computing NSS/IEEE workshop Ph. Charpentier, CERN B00le.
ALICE Physics Data Challenge 3
Status and Prospects of The LHC Experiments Computing
LHCb Computing Model and Data Handling Angelo Carbone 5° workshop italiano sulla fisica p-p ad LHC 31st January 2008.
Philippe Charpentier CERN – LHCb On behalf of the LHCb Computing Group
Artem Trunov and EKP team EPK – Uni Karlsruhe
Simulation use cases for T2 in ALICE
R. Graciani for LHCb Mumbay, Feb 2006
LHC Data Analysis using a worldwide computing grid
Pierre Girard ATLAS Visit
LHCb status and plans Ph.Charpentier CERN.
CC and LQCD dimanche 13 janvier 2019dimanche 13 janvier 2019
Gridifying the LHCb Monte Carlo production system
ATLAS DC2 & Continuous production
LHCb thinking on Regional Centres and Related activities (GRIDs)
The LHCb Computing Data Challenge DC06
Presentation transcript:

LHCb Computing Philippe Charpentier CERN

LHCb in brief Experiment dedicated to studying CP-violation Responsible for the dominance of matter on antimatter Matter-antimatter difference studied using the b-quark (beauty) High precision physics (tiny difference…) Single arm spectrometer Looks like a fixed-target experiment Smallest of the 4 big LHC experiments ~500 physicists Nevertheless, computing is also a challenge….

LHCb data processing software Simul. Gauss Analysis DaVinci MCHits DST Raw Data (r)DST MCParts GenParts Event model / Physics event model µDST Conditions Database Gaudi Digit. Boole Trigger Moore Recons. Brunel LHCb Computing, PhC

LHCb Basic Computing principles Raw data shipped in real time to Tier-0 Registered in the Grid File Catalog (LFC) Raw data provenance in a Bookkeeping database (query-enabled) Resilience enforced by a second copy at Tier-1’s Rate: ~2000 evts/s (35 kB)  70 MB/s All data processing up to final µDST or Tuple production distributed Not possible to perform first pass reconstruction of all data at Tier0 Consider Tier0 also is distributed First pass reconstruction at all Tier1s like re-processing Analysis performed at Analysis Facilities In Computing Model: AF at Tier1s Part of the analysis is not data-related Extracting physics parameters on CP violation (toy-MC, complex fitting procedures…) Also using distributed computing resources LHCb Computing, PhC

Basic principles (cont’d) LHCb runs jobs where data are All data are placed explicitly Analysis made possible by reduction of datasets many different channels of interest very few events in each channel (from 102 to 106 events / year) physicist dealing with maximum 107 events small and simple events final dataset manageable on physicist’s desktop (100’s of GBytes) Calibration and alignment performed on a selected part of the data stream (at CERN) Alignment and tracking calibration using dimuons (~5/s) Used also for validation of new calibration PID calibration using Ks, D* LHCb Computing, PhC

LHCb dataflow Tier0 Tier2 Tier1 Online MSS-SE Recons. Tier1 Analysis Simulation. Online Tier0 Tier2 Raw MSS-SE Tier1 Digi Recons. Raw/Digi Tier1 MSS-SE rDST Analysis Stripping rDST+Raw DST DST LHCb Computing, PhC

Comments on the LHCb Distributed Computing Only last part of the analysis is foreseen to be “interactive” Either analysing ROOT trees or using GaudiPython/pyRoot User analysis at Tier1’s - why? Analysis is very delicate, needs careful file placement Tier1’s are easier to check, less prone (in principle) to outages CPU requirements are very modest What is LHCb’s concept of the Grid? It is a set of computing resources working in a collaborative way Provides computing resources for the collaboration as a whole Recognition of contributions is independent on what type of jobs are run at a site There are no noble and less noble tasks. All are needed to make the experiment a success Resources are not made available for nationals Resource high availability is the key issue LHCb Computing, PhC

Further comments on Analysis Preliminary: currently being discussed Local analysis (Tier3) Non-pledged resources, reserved to local users (no CE access, local batch queues, no central accounting) Storage may be a Grid-SE (i.e. SRM-enabled) or not Copy or Replication performed by Dirac DMS tools Grid-SE: replication, can use third-party transfers Replica should be registered in LFC Non Grid-SE: copy from a local node LFC registration more problematic (no SRM), but possible Analysis on Tier2 Pledged resources, therefore available to the whole collaboration Resources should be additional (dedicated to analysis) We have just enough with Tier2 for simulation… Storage and data access handled by local team (no central manpower available) Data fully replicated in Grid-SE (LFC) CE centrally banned in case of failures (as for Tier1s) LHCb Computing, PhC

How to best achieve Distributed Computing? Data Management is primordial Availability of Storage Elements at Tier1’s Reliability of SRM and transfers Efficiency of data access protocols (rfio, (gsi)dcap, xrootd…) Infrastructure is vital Resource management 24x7 support coverage Reliable and powerful networks (OPN) Resource sharing is a must Less support needed Best resource usage (less idle CPUs, empty tapes, unused networks…) Shares must be on long term, no hard limit on number of slots …. but opportunistic resources should not be neglected… EGEE, EGI? LHCb Computing, PhC

LHCb Distributed Computing software Integrated WMS and DMS : DIRAC Distributed analysis portal: GANGA Uses DIRAC W&DMS as back-end DIRAC’s main characteristics Implements late job scheduling Overlay network (pilot jobs, central task queue) Pull paradigm Generic pilot jobs: allows to run multiple payload Allows LHCb policy to be enforced Alleviates the level of support required from sites LHCb services designed to be redundant and hence highly available (multiple instances with failover, VO-BOXes) LHCb Computing, PhC

WMS with pilot jobs Jobs are submitted with credentials of their owner (VOMS proxy) The proxy is renewed automatically inside the WMS repository The Pilot Job fetches the User Job and proxy The User Job is executed with its owner’s proxy used to access SE, catalogs, etc

The LHCb Tier1s 6 Tier1s Contribute to Keeps copies on MSS of CNAF (IT, Bologna) GridKa (DE, Karlsruhe) IN2P3 (FR, Lyon) NIKHEF (NL, Amsterdam) PIC (ES, Barcelona) RAL (UK, Didcot) Contribute to Reconstruction Stripping Analysis Keeps copies on MSS of Raw (2 copies shared) Locally produced rDST DST (2 copies) MC data (2 copies) Keeps copies on disk of DST (7 copies)

LHCb Computing: a few numbers Event sizes on persistent medium (not in memory) Processing time Best estimates as of today Requirements for 2009-10 6 106 seconds of beam ~109 MC b-events ~3. 109 MC non-b events TDR estimate Current estimate Event Size kB RAW 25 35 rDST 20 DST 100 110 Evt processing kSI2k.s Simulation 75 90 Reconstruction 2.4 3 Stripping 0.2 Analysis 0.3

Resource requirements for 2009-10 Site Fraction (%) CERN 14 FZK 11 IN2P3 25 CNAF 9 NL-T1 26 PIC 4 RAL CPU requirements Disk requirements Tape requirements IN2P3-CC represents 25% of the LHCb Tier1 pledges LHCb Computing, PhC

Statistiques depuis le 1er janvier 2009 LHCb and LCG-France Statistiques depuis le 1er janvier 2009 LHCb Computing, PhC

CPU (in days) from January 2009 LHCb Computing, PhC

Jobs in France LHCb Computing, PhC

Tests for analysis LHCb Computing, PhC

Analysis data access (cont’d) LHCb Computing, PhC

Conclusions LHCb has proposed a Computing Model adapted at its specific needs (number of events, event size, low number of physics candidates) Reconstruction, stripping and analysis resources located at Tier1s (and possibly some Tier2s with enough storage and CPU capacities) CPU requirements dominated by Monte-Carlo, assigned to Tier2s and opportunistic sites With DIRAC, even idle desktops / laptops could be used ;-) LHCb@home ? Requirements are modest compared to other experiments DIRAC is well suited and adapted to this computing model Integrated WMS and DMS LHCb Computing, PhC