LHCb Computing Model and Data Handling Angelo Carbone 5° workshop italiano sulla fisica p-p ad LHC 31st January 2008.

Slides:



Advertisements
Similar presentations
LHCb on the Grid A Tale of many Migrations
Advertisements

Exporting Raw/ESD data from Tier-0 Tier-1s Wrap-up.
CHEP 2012 – New York City 1.  LHC Delivers bunch crossing at 40MHz  LHCb reduces the rate with a two level trigger system: ◦ First Level (L0) – Hardware.
Stuart K. PatersonCHEP 2006 (13 th –17 th February 2006) Mumbai, India 1 from DIRAC.Client.Dirac import * dirac = Dirac() job = Job() job.setApplication('DaVinci',
Large scale data flow in local and GRID environment V.Kolosov, I.Korolko, S.Makarychev ITEP Moscow.
December 17th 2008RAL PPD Computing Christmas Lectures 11 ATLAS Distributed Computing Stephen Burke RAL.
Computing Infrastructure Status. LHCb Computing Status LHCb LHCC mini-review, February The LHCb Computing Model: a reminder m Simulation is using.
Nick Brook Current status Future Collaboration Plans Future UK plans.
Dan Tovey, University of Sheffield GridPP: Experiment Status & User Feedback Dan Tovey University Of Sheffield.
F. Fassi, S. Cabrera, R. Vives, S. González de la Hoz, Á. Fernández, J. Sánchez, L. March, J. Salt, A. Lamas IFIC-CSIC-UV, Valencia, Spain Third EELA conference,
BESIII Production with Distributed Computing Xiaomei Zhang, Tian Yan, Xianghu Zhao Institute of High Energy Physics, Chinese Academy of Sciences, Beijing.
Status of the LHCb MC production system Andrei Tsaregorodtsev, CPPM, Marseille DataGRID France workshop, Marseille, 24 September 2002.
November SC06 Tampa F.Fanzago CRAB a user-friendly tool for CMS distributed analysis Federica Fanzago INFN-PADOVA for CRAB team.
CCRC’08 Weekly Update Jamie Shiers ~~~ LCG MB, 1 st April 2008.
1 LCG-France sites contribution to the LHC activities in 2007 A.Tsaregorodtsev, CPPM, Marseille 14 January 2008, LCG-France Direction.
LHCb The LHCb Data Management System Philippe Charpentier CERN On behalf of the LHCb Collaboration.
T3 analysis Facility V. Bucard, F.Furano, A.Maier, R.Santana, R. Santinelli T3 Analysis Facility The LHCb Computing Model divides collaboration affiliated.
The LHCb CERN R. Graciani (U. de Barcelona, Spain) for the LHCb Collaboration International ICFA Workshop on Digital Divide Mexico City, October.
Author: Andrew C. Smith Abstract: LHCb's participation in LCG's Service Challenge 3 involves testing the bulk data transfer infrastructure developed to.
1 LHCb on the Grid Raja Nandakumar (with contributions from Greig Cowan) ‏ GridPP21 3 rd September 2008.
Outline: The LHCb Computing Model Philippe Charpentier, CERN ICFA workshop on Grid activities, Sinaia, Romania, October
Large scale data flow in local and GRID environment Viktor Kolosov (ITEP Moscow) Ivan Korolko (ITEP Moscow)
LHCb report to LHCC and C-RSG Philippe Charpentier CERN on behalf of LHCb.
David Stickland CMS Core Software and Computing
DIRAC Project A.Tsaregorodtsev (CPPM) on behalf of the LHCb DIRAC team A Community Grid Solution The DIRAC (Distributed Infrastructure with Remote Agent.
1 LHCb computing for the analysis : a naive user point of view Workshop analyse cc-in2p3 17 avril 2008 Marie-Hélène Schune, LAL-Orsay for LHCb-France Framework,
LHCb status and plans Ph.Charpentier CERN. LHCb status and plans WLCG Workshop 1-2 Sept 2007, Victoria, BC 2 Ph.C. Status of DC06  Reminder:  Two-fold.
The DataGrid Project NIKHEF, Wetenschappelijke Jaarvergadering, 19 December 2002
Distributed Physics Analysis Past, Present, and Future Kaushik De University of Texas at Arlington (ATLAS & D0 Collaborations) ICHEP’06, Moscow July 29,
GDB meeting - July’06 1 LHCb Activity oProblems in production oDC06 plans & resource requirements oPreparation for DC06 oLCG communications.
Distributed Analysis Tutorial Dietrich Liko. Overview  Three grid flavors in ATLAS EGEE OSG Nordugrid  Distributed Analysis Activities GANGA/LCG PANDA/OSG.
LHCb 2009-Q4 report Q4 report LHCb 2009-Q4 report, PhC2 Activities in 2009-Q4 m Core Software o Stable versions of Gaudi and LCG-AA m Applications.
1 LHCb Computing A.Tsaregorodtsev, CPPM, Marseille 14 March 2007, Clermont-Ferrand.
The Worldwide LHC Computing Grid WLCG Milestones for 2007 Focus on Q1 / Q2 Collaboration Workshop, January 2007.
The CMS Beijing Tier 2: Status and Application Xiaomei Zhang CMS IHEP Group Meeting December 28, 2007.
LHCb Computing 2015 Q3 Report Stefan Roiser LHCC Referees Meeting 1 December 2015.
1-2 March 2006 P. Capiluppi INFN Tier1 for the LHC Experiments: ALICE, ATLAS, CMS, LHCb.
LHCb Status report June 08. LHCb Computing Report Activities since February  Applications and Core Software  Preparation of applications for real data.
ATLAS – statements of interest (1) A degree of hierarchy between the different computing facilities, with distinct roles at each level –Event filter Online.
Lessons learned administering a larger setup for LHCb
THE ATLAS COMPUTING MODEL Sahal Yacoob UKZN On behalf of the ATLAS collaboration.
ATLAS Distributed Analysis S. González de la Hoz 1, D. Liko 2, L. March 1 1 IFIC – Valencia 2 CERN.
LHCb D ata P rocessing S oftware J. Blouw, A. Zhelezov Physikalisches Institut, Universitaet Heidelberg DESY Computing Seminar, Nov. 29th, 2010.
L’analisi in LHCb Angelo Carbone INFN Bologna
LCG Service Challenge: Planning and Milestones
ATLAS Use and Experience of FTS
Overview of the Belle II computing
INFN GRID Workshop Bari, 26th October 2004
DIRAC Production Manager Tools
Akiya Miyamoto KEK 1 June 2016
Added value of new features of the ATLAS computing model and a shared Tier-2 and Tier-3 facilities from the community point of view Gabriel Amorós on behalf.
Data Challenge with the Grid in ATLAS
INFN-GRID Workshop Bari, October, 26, 2004
ATLAS activities in the IT cloud in April 2008
The LHCb Software and Computing NSS/IEEE workshop Ph. Charpentier, CERN B00le.
LHCb Software & Computing Status
LHCb computing in Russia
Philippe Charpentier CERN – LHCb On behalf of the LHCb Computing Group
Readiness of ATLAS Computing - A personal view
Evolution of the distributed computing model The case of CMS
MC data production, reconstruction and analysis - lessons from PDC’04
Simulation use cases for T2 in ALICE
WLCG Service Report 5th – 18th July
LCG middleware and LHC experiments ARDA project
R. Graciani for LHCb Mumbay, Feb 2006
LHCb Computing Philippe Charpentier CERN
LHC Data Analysis using a worldwide computing grid
LHCb status and plans Ph.Charpentier CERN.
ATLAS DC2 & Continuous production
The LHCb Computing Data Challenge DC06
Presentation transcript:

LHCb Computing Model and Data Handling Angelo Carbone 5° workshop italiano sulla fisica p-p ad LHC 31st January 2008

LHCb Computing Model (I) DAQ parameters ~2000 evts/s 35 kB /evts ~70 MB/s recording rate at T0 (CERN tape) RAW DATA CERN RAW DATA CNAF PIC RAL IN2P3 GRIDKA NIKHEF Reconstruction output rDST Performed at the 7 sites (Tier0 + 6 Tier1s) Local storage of rDST (purely reconstruction output) 5° workshop italiano sulla fisica p-p ad LHC Angelo Carbone Perugia, 30 Gennaio-2 Febbraio

LHCb Computing Model (II) CERN PIC IN2P3 DST replicated in each T1 NIKHEF CNAF RAL GRIDKA Stripping (pre-selection) output  DST In order to reduce the data sample with specific pre-selection algorithms Performed at site where rDST are produced Output DST replicated to the 7 sites Analysis Performed at 7 sites (CNAF + Tier1s) MC Production Performed at Tier2, Italian collaboration has 1 Tier2 at CNAF 5° workshop italiano sulla fisica p-p ad LHC Angelo Carbone Perugia, 30 Gennaio-2 Febbraio

Workload, production and data management tools DIRAC Distributed Infrastructure with Remote Agent Control Workload management System Central task queue Use of Pilot agents Data Management System Data transfers with full error recovery Automatic data distribution (Tier0 to Tier1s) Production tools Complex job workflows (multi-applications) Automatic job submission (data driven) for processing Jobs are submitted where data are (no forced transfers) 5° workshop italiano sulla fisica p-p ad LHC Angelo Carbone Perugia, 30 Gennaio-2 Febbraio

DIRAC WMS Realizes the PULL scheduling paradigm Offers support for both centrally managed productions and individual user jobs Pilot Agent paradigm LCG jobs are Pilot jobs for the DIRAC WMS Allows checking of the environment before job scheduling On WN: matches job in the central task queue Terminates gracefully if no work is available 5° workshop italiano sulla fisica p-p ad LHC Angelo Carbone Perugia, 30 Gennaio-2 Febbraio

Data management system Different tools are used for data transfer FTS gLite File Transfer Service It is a multi-experiment service, used to balance usage of site resources DIRAC has a transfer DB service to manage the transfer request (i.e RAW, DST distribution to TIER-1) Bulk transfers submitted to FTS lcg-utils Directly copy the file to the SE (i.e. MC production) DIRAC provides a distributed failover in case the select SE is not available 5° workshop italiano sulla fisica p-p ad LHC Angelo Carbone Perugia, 30 Gennaio-2 Febbraio

LHCb MC production DC06 700 million events were simulated since May 2006 1.5 M events / day through the DC06 period 1.5 million jobs successfully executed Up to 10000 jobs running simultaneously (figure below) Over 120 sites used worldwide 5° workshop italiano sulla fisica p-p ad LHC Angelo Carbone Perugia, 30 Gennaio-2 Febbraio

Reconstruction and stripping DC06 100 million events have been reconstructed 200,000 files recalled from tape 10 million events stripped 10,000 jobs submitted Up to 1200 jobs running simultaneously All the Tier-1 sites used CNAF Computing model successfully tested Data management is the big issue 5° workshop italiano sulla fisica p-p ad LHC Angelo Carbone Perugia, 30 Gennaio-2 Febbraio

Computing centre contribution LHCb CPU usage 2007 LHCb data proccessing 2007 Data from http://lhcb.pic.es/DIRAC/Accounting/ 5° workshop italiano sulla fisica p-p ad LHC Angelo Carbone Perugia, 30 Gennaio-2 Febbraio

Distributed Data Analysis: Ganga Ganga (Gaudi /Athena and Grid Alliance) is an easy-to-use frontend for job definition and management Application: specify the software to be run (Gaudi, ROOT) Backend: specify processing system (for LHCb: DIRAC) Input dataset: use only Logical File Name (LFN) Output dataset Splitters: possibility to divide a job into subjobs that can be processed in parallel Merging: possibility to combine the resultant outputs Used by LHCb and ATLAS 5° workshop italiano sulla fisica p-p ad LHC Angelo Carbone Perugia, 30 Gennaio-2 Febbraio

Job Splitting GANGA provides, if requested, the automatic job parallelization Need only to specify the splitting rule Automatic data driven submission of sub-jobs main Job List of LFN GANGA List of PFN catalog LFC sub-jobs CNAF RAL CERN IN2P3 GRIDKA PIC NL-T1 5° workshop italiano sulla fisica p-p ad LHC Angelo Carbone Perugia, 30 Gennaio-2 Febbraio

Ganga usage LHCb 65 unique users in the last month total of 2500 sessions in the last month 5° workshop italiano sulla fisica p-p ad LHC Angelo Carbone Perugia, 30 Gennaio-2 Febbraio

BACKUP 5° workshop italiano sulla fisica p-p ad LHC Angelo Carbone Perugia, 30 Gennaio-2 Febbraio

CCRC’08 February Activities Maintain equivalent of 2 weeks of 2008 data taking Assuming a 50% machine cycle efficiency Raw data distribution from pit → T0 centre Raw data distribution from T0 → T1 centres Recons of raw data at CERN & T1 centres RAW rDST Stripping of data at CERN & T1 centres rDST DST Distribution of DST data to all other centres 1 copy of DST data replicated at each TIER-1 5° workshop italiano sulla fisica p-p ad LHC Angelo Carbone Perugia, 30 Gennaio-2 Febbraio

CCRC’08 numbers at CNAF (February) A total of ~15 TB of data to be written T1D0, 7.2TB, CASTOR-2, srm 2.2 T0D1, 6.8TB, StoRM, srm 2.2 T1D1, 1.2 TB, StoRM, srm 2.2 Overall bandwidth for each “way” < 10 MB/s 330 simultaneous jobs/day Job type Duration (hrs) input files Recons 24 2 GB ~ 20KB/s Strip 6 8 GB ~ 100 KB/s 5° workshop italiano sulla fisica p-p ad LHC Angelo Carbone Perugia, 30 Gennaio-2 Febbraio

CCRC’08 activities across the sites Breakdown of processing activities (CPU needs) Site Fraction (%) CERN 14 FZK IN2P3 22 CNAF 15 NIKHEF/SARA 17 PIC 8 RAL 10 No other activities, except user analysis 5° workshop italiano sulla fisica p-p ad LHC Angelo Carbone Perugia, 30 Gennaio-2 Febbraio

Reconstruction 12.6 MB/s RAW 2.7 MB/s rDST 12.6 MB/s RAW 5.4 MB/s RAW 5° workshop italiano sulla fisica p-p ad LHC Angelo Carbone Perugia, 30 Gennaio-2 Febbraio

Stripping 2.7 MB/s rDST 6.3 MB/s DST 1.1 MB/s DST 5.4 MB/s RAW 5° workshop italiano sulla fisica p-p ad LHC Angelo Carbone Perugia, 30 Gennaio-2 Febbraio