LHCb D ata P rocessing S oftware J. Blouw, A. Zhelezov Physikalisches Institut, Universitaet Heidelberg DESY Computing Seminar, Nov. 29th, 2010.

Slides:



Advertisements
Similar presentations
31/03/00 CMS(UK)Glenn Patrick What is the CMS(UK) Data Model? Assume that CMS software is available at every UK institute connected by some infrastructure.
Advertisements

CHEP 2012 – New York City 1.  LHC Delivers bunch crossing at 40MHz  LHCb reduces the rate with a two level trigger system: ◦ First Level (L0) – Hardware.
Grid and CDB Janusz Martyniak, Imperial College London MICE CM37 Analysis, Software and Reconstruction.
Batch Production and Monte Carlo + CDB work status Janusz Martyniak, Imperial College London MICE CM37 Analysis, Software and Reconstruction.
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
1 Grid services based architectures Growing consensus that Grid services is the right concept for building the computing grids; Recent ARDA work has provoked.
6/4/20151 Introduction LHCb experiment. LHCb experiment. Common schema of the LHCb computing organisation. Common schema of the LHCb computing organisation.
23/04/2008VLVnT08, Toulon, FR, April 2008, M. Stavrianakou, NESTOR-NOA 1 First thoughts for KM3Net on-shore data storage and distribution Facilities VLV.
5 November 2001F Harris GridPP Edinburgh 1 WP8 status for validating Testbed1 and middleware F Harris(LHCb/Oxford)
K.Harrison CERN, 21st November 2002 GANGA: GAUDI/ATHENA AND GRID ALLIANCE - Background and scope - Project organisation - Technology survey - Design -
Computing and LHCb Raja Nandakumar. The LHCb experiment  Universe is made of matter  Still not clear why  Andrei Sakharov’s theory of cp-violation.
3 Sept 2001F HARRIS CHEP, Beijing 1 Moving the LHCb Monte Carlo production system to the GRID D.Galli,U.Marconi,V.Vagnoni INFN Bologna N Brook Bristol.
Computing Infrastructure Status. LHCb Computing Status LHCb LHCC mini-review, February The LHCb Computing Model: a reminder m Simulation is using.
Cosener’s House – 30 th Jan’031 LHCb Progress & Plans Nick Brook University of Bristol News & User Plans Technical Progress Review of deliverables.
Nick Brook Current status Future Collaboration Plans Future UK plans.
1 DIRAC – LHCb MC production system A.Tsaregorodtsev, CPPM, Marseille For the LHCb Data Management team CHEP, La Jolla 25 March 2003.
Bookkeeping Tutorial. Bookkeeping & Monitoring Tutorial2 Bookkeeping content  Contains records of all “jobs” and all “files” that are created by production.
Databases E. Leonardi, P. Valente. Conditions DB Conditions=Dynamic parameters non-event time-varying Conditions database (CondDB) General definition:
Results of the LHCb experiment Data Challenge 2004 Joël Closier CERN / LHCb CHEP’ 04.
Status of the LHCb MC production system Andrei Tsaregorodtsev, CPPM, Marseille DataGRID France workshop, Marseille, 24 September 2002.
November SC06 Tampa F.Fanzago CRAB a user-friendly tool for CMS distributed analysis Federica Fanzago INFN-PADOVA for CRAB team.
The ILC And the Grid Andreas Gellrich DESY LCWS2007 DESY, Hamburg, Germany
4/5/2007Data handling and transfer in the LHCb experiment1 Data handling and transfer in the LHCb experiment RT NPSS Real Time 2007 FNAL - 4 th May 2007.
And Tier 3 monitoring Tier 3 Ivan Kadochnikov LIT JINR
1 LCG-France sites contribution to the LHC activities in 2007 A.Tsaregorodtsev, CPPM, Marseille 14 January 2008, LCG-France Direction.
EGEE is a project funded by the European Union under contract IST HEP Use Cases for Grid Computing J. A. Templon Undecided (NIKHEF) Grid Tutorial,
Computing Division Requests The following is a list of tasks about to be officially submitted to the Computing Division for requested support. D0 personnel.
Introduction CMS database workshop 23 rd to 25 th of February 2004 Frank Glege.
LHCb The LHCb Data Management System Philippe Charpentier CERN On behalf of the LHCb Collaboration.
Author: Andrew C. Smith Abstract: LHCb's participation in LCG's Service Challenge 3 involves testing the bulk data transfer infrastructure developed to.
1 LHCb on the Grid Raja Nandakumar (with contributions from Greig Cowan) ‏ GridPP21 3 rd September 2008.
Bookkeeping Tutorial. 2 Bookkeeping content  Contains records of all “jobs” and all “files” that are produced by production jobs  Job:  In fact technically.
The JANA Reconstruction Framework David Lawrence - JLab May 25, /25/101JANA - Lawrence - CLAS12 Software Workshop.
Testing and integrating the WLCG/EGEE middleware in the LHC computing Simone Campana, Alessandro Di Girolamo, Elisa Lanciotti, Nicolò Magini, Patricia.
Tier3 monitoring. Initial issues. Danila Oleynik. Artem Petrosyan. JINR.
Large scale data flow in local and GRID environment Viktor Kolosov (ITEP Moscow) Ivan Korolko (ITEP Moscow)
General requirements for BES III offline & EF selection software Weidong Li.
ATLAS-specific functionality in Ganga - Requirements for distributed analysis - ATLAS considerations - DIAL submission from Ganga - Graphical interfaces.
DIRAC Project A.Tsaregorodtsev (CPPM) on behalf of the LHCb DIRAC team A Community Grid Solution The DIRAC (Distributed Infrastructure with Remote Agent.
INFSO-RI Enabling Grids for E-sciencE Using of GANGA interface for Athena applications A. Zalite / PNPI.
1 LHCb computing for the analysis : a naive user point of view Workshop analyse cc-in2p3 17 avril 2008 Marie-Hélène Schune, LAL-Orsay for LHCb-France Framework,
LHCbDirac and Core Software. LHCbDirac and Core SW Core Software workshop, PhC2 Running Gaudi Applications on the Grid m Application deployment o CVMFS.
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
Markus Frank (CERN) & Albert Puig (UB).  An opportunity (Motivation)  Adopted approach  Implementation specifics  Status  Conclusions 2.
M.Frank, CERN/LHCb Persistency Workshop, Dec, 2004 Distributed Databases in LHCb  Main databases in LHCb Online / Offline and their clients  The cross.
A Data Handling System for Modern and Future Fermilab Experiments Robert Illingworth Fermilab Scientific Computing Division.
Breaking the frontiers of the Grid R. Graciani EGI TF 2012.
ATLAS Physics Analysis Framework James R. Catmore Lancaster University.
DIRAC for Grid and Cloud Dr. Víctor Méndez Muñoz (for DIRAC Project) LHCb Tier 1 Liaison at PIC EGI User Community Board, October 31st, 2013.
Scientific Data Processing Portal and Heterogeneous Computing Resources at NRC “Kurchatov Institute” V. Aulov, D. Drizhuk, A. Klimentov, R. Mashinistov,
ATLAS – statements of interest (1) A degree of hierarchy between the different computing facilities, with distinct roles at each level –Event filter Online.
LHCb distributed computing during the LHC Runs 1,2 and 3
Domenico Galli, Bologna
L’analisi in LHCb Angelo Carbone INFN Bologna
Database Replication and Monitoring
Report from WLCG Workshop 2017: WLCG Network Requirements GDB - CERN 12th of July 2017
INFN GRID Workshop Bari, 26th October 2004
CMS High Level Trigger Configuration Management
Moving the LHCb Monte Carlo production system to the GRID
The LHCb Software and Computing NSS/IEEE workshop Ph. Charpentier, CERN B00le.
LHCb Computing Model and Data Handling Angelo Carbone 5° workshop italiano sulla fisica p-p ad LHC 31st January 2008.
TriggerDB copy in TriggerTool
The Ganga User Interface for Physics Analysis on Distributed Resources
Monitoring of the infrastructure from the VO perspective
R. Graciani for LHCb Mumbay, Feb 2006
LHCb Computing Philippe Charpentier CERN
Gridifying the LHCb Monte Carlo production system
Status and plans for bookkeeping system and production tools
The LHCb Computing Data Challenge DC06
Presentation transcript:

LHCb D ata P rocessing S oftware J. Blouw, A. Zhelezov Physikalisches Institut, Universitaet Heidelberg DESY Computing Seminar, Nov. 29th, 2010

2/2/Contents The LHCb experiment Context Framework Design Online Offline Results

3/3/ The LHCb experiment & performance

4/4/ Excellent performance of LHC! Excellent start of data taking! Collected Integrated & and recorded luminosity:Data taking efficiency vs. run number:

5/5/ Context: data volume/flow Use of the framework during operations: Data rate O(100 MB)/s ⇒ online requirements Onlin e Offlin e

6/6/ Context: data volume/flow Use of the framework during operations: Data rate O(100 MB)/s ⇒ online requirements This year: O(20 TB) of raw data Including dst’s and physics streams: O(700) TB Onlin e Offlin e

7/7/ Context: data volume/flow Use of the framework during operations: Data rate O(100 MB)/s ⇒ online requirements This year: O(20 TB) of raw data Including dst’s and physics streams: O(700) TB ⇒ offline requirements Onlin e Offlin e

8/8/ The LHCb Software Framework implemented in C++ Separation between data and algorithms Separation between persistent and transient data Useful for developing applications in HEP Run in different environments (trigger applications, track reconstruction) Useful both online & offline The ’baby’ is called: Gaudi.

9/9/Gaudi

10/Gaudi transient data: lifetime for duration of process algorithms read/write data objects from transient data stores buffer for data conversion has unix, tree-like structure

11/ Gaudi transient data: lifetime for duration of process algorithms read/write data objects from transient data stores buffer for data conversion has unix, tree-like structure persistent data: “Converters” convert transient data in persistent data and vice versa different persistency technologies depending on access pattern

12/Gaudi transient data: lifetime for duration of process algorithms read/write data objects from transient data stores buffer for data conversion has unix, tree-like structure persistent data: “Converters” convert transient data in persistent data and vice versa different persistency technologies depending on access pattern services MessageSvc: verbosity level of output Job option service: configuration of user algorithms persistency services: input/output of data other services...

13/ Online: Trigger Reminder: dataflow

14/ Online: Trigger Scheme

15/ Online: Trigger Performance Efficiency: ε = 98% from Offline reconstruction Resolution: ∼ 3%

16/ Online: Trigger

17/ Online: Track-based Monitoring Schematic

18/ Online: Track-based Monitoring Online Track Reconstruction

19/ Online: Track-based Monitoring Monitoring

20/ Online: Track-based Monitoring Presenter

21/ Online: Conclusions Gaudi succesfully used: in software trigger (HLT) in monitoring software for development of LHCb software (simulation, reconstruction) by the Atlas experiment...

22/ Offline processing tasks Steered by dedicated Production Team Data distribution (T0->T1) Reconstruction + Stripping (T0/1) RAW -> rDST rDST -> DST (streams) Monte Carlo production (T2) Data replication (T0/1->T0/1) Steered by individual users Analysis (T0/T1) DST -> Tuples/Histograms

23/ Storage Offline Data Flow Ta pe CERN (T0) CERN (T0) LHC b Disc RAW data (T2) Monte Carlo DST Histograms/Tuples rDST GridK a (T1) CNAF (T1) NIKH EF (T1) PIC (T1) IN2P3 (T1) RAL (T1)

24/ Production Management System Workload Management System Data Management System

25/DIRAC ● Distributed Infrastructure with Remote Agent Control ● Python based daemons written within common framework ● Services: Configuration, Bookkeeping, Accounting, etc. ● Agents: Monitoring, Transformation, Production tracking, etc. ● Common web interface ● 200+ command line utilities ● No longer LHCb specific

26/ Production Request System

27/ Production Monitoring Separate request and production monitoring Easy navigation down to the job level

28/Accounting Jobs, pilots and data transfers history with different granularity and selection criteria

29/ User jobs Unpredictable nature Less tested applications Strict deadline (conference) Several approaches DIRAC API GANGA + DIRAC backend GANGA + batch system backend

30/ GANGA for LHCb G PI GUI CLI P Scri pt Application plugins Gaudi Brunel DaVinci... GANGA Core Backends DIRAC Batch Local LHCBDataset

31/ LHCb Bookkeeping Part of DIRAC Oracle DB based 3 user interfaces DIRAC Web Portal GUI Command line

32/ DIRAC job monitoring Progress Parameters Failure analysis

33/ Not in GRID Reasons Development Not standard software Reliability Problems Software model: lxplus/CERN batch only How to get the data? How to install the software?

34/ Localized resources Local farm (PI HD) NAF (DESY ZN)

35/ Website to support users

36/ Data and software manager The same interface for local farm and NAF

37/ Personal “Computer Center” Reason LHCb core software “prefer” SL(C) 4/5/(6) Modern desktops/notebooks “prefer” Ubuntu/SUSE Solution: “chroot” virtualization No resource penalties Several systems in parallel without RAM multiplication Zero processing speed degradation Easy to deploy 3 years without any problem (Debian/Ubuntu hosts)

38/ Some results & Conclusions CP violation in Bs system is the main program for LHCb Prerequisites: oscillation frequency of Bs mixing known Use Bs → KK and Bd → ππ to resolve ambiguity

39/ Some results & Conclusions CP violation in Bs system is the main program for LHCb Prerequisites: oscillation frequency of Bs mixing known Use Bs → KK and Bd → ππ to resolve ambiguity Bs → J/Ψφ signal with 1/20 of this years data

40/ Some results & Conclusions CP violation in Bs system is the main program for LHCb Prerequisites: oscillation frequency of Bs mixing known Use Bs → KK and Bd → ππ to resolve ambiguity Bs → J/Ψφ signal with 1/20 of this years data Oscillations in Bd → D ∗ lν already seen! Expect competitive measurement with 2010 data!

41/ Some results & Conclusions CP violation in Bs system is the main program for LHCb Prerequisites: oscillation frequency of Bs mixing known Use Bs → KK and Bd → ππ to resolve ambiguity Bs → J/Ψφ signal with 1/20 of this years data Oscillations in Bd → D ∗ lν already seen! Expect competitive measurement with 2010 data! Glimpse of CP in Bd → Kπ

42/Conclusion Already published physics papers and numerous talks have proven that our data processing software and tools are capable of doing the job. We are looking forward for many exciting discoveries in the next years.