LHCb 2009-Q4 report. 2009 Q4 report LHCb 2009-Q4 report, PhC2 Activities in 2009-Q4 m Core Software o Stable versions of Gaudi and LCG-AA m Applications.

Slides:



Advertisements
Similar presentations
Clara Gaspar on behalf of the LHCb Collaboration, “Physics at the LHC and Beyond”, Quy Nhon, Vietnam, August 2014 Challenges and lessons learnt LHCb Operations.
Advertisements

K.Harrison CERN, 23rd October 2002 HOW TO COMMISSION A NEW CENTRE FOR LHCb PRODUCTION - Overview of LHCb distributed production system - Configuration.
Stuart K. PatersonCHEP 2006 (13 th –17 th February 2006) Mumbai, India 1 from DIRAC.Client.Dirac import * dirac = Dirac() job = Job() job.setApplication('DaVinci',
Large scale data flow in local and GRID environment V.Kolosov, I.Korolko, S.Makarychev ITEP Moscow.
LHCb Quarterly Report October Core Software (Gaudi) m Stable version was ready for 2008 data taking o Gaudi based on latest LCG 55a o Applications.
Computing and LHCb Raja Nandakumar. The LHCb experiment  Universe is made of matter  Still not clear why  Andrei Sakharov’s theory of cp-violation.
03/27/2003CHEP20031 Remote Operation of a Monte Carlo Production Farm Using Globus Dirk Hufnagel, Teela Pulliam, Thomas Allmendinger, Klaus Honscheid (Ohio.
BaBar Grid Computing Eleonora Luppi INFN and University of Ferrara - Italy.
Computing Infrastructure Status. LHCb Computing Status LHCb LHCC mini-review, February The LHCb Computing Model: a reminder m Simulation is using.
CERN IT Department CH-1211 Genève 23 Switzerland t Internet Services Job Monitoring for the LHC experiments Irina Sidorova (CERN, JINR) on.
1 DIRAC – LHCb MC production system A.Tsaregorodtsev, CPPM, Marseille For the LHCb Data Management team CHEP, La Jolla 25 March 2003.
WLCG Service Report ~~~ WLCG Management Board, 1 st September
Status of the LHCb MC production system Andrei Tsaregorodtsev, CPPM, Marseille DataGRID France workshop, Marseille, 24 September 2002.
4/5/2007Data handling and transfer in the LHCb experiment1 Data handling and transfer in the LHCb experiment RT NPSS Real Time 2007 FNAL - 4 th May 2007.
Enabling Grids for E-sciencE System Analysis Working Group and Experiment Dashboard Julia Andreeva CERN Grid Operations Workshop – June, Stockholm.
1 LCG-France sites contribution to the LHC activities in 2007 A.Tsaregorodtsev, CPPM, Marseille 14 January 2008, LCG-France Direction.
Pilot Jobs John Gordon Management Board 23/10/2007.
LCG Pilot Jobs + glexec John Gordon, STFC-RAL GDB 7 November 2007.
1 LHCb on the Grid Raja Nandakumar (with contributions from Greig Cowan) ‏ GridPP21 3 rd September 2008.
1 Andrea Sciabà CERN Critical Services and Monitoring - CMS Andrea Sciabà WLCG Service Reliability Workshop 26 – 30 November, 2007.
Refitting Tracks from DST E. Rodrigues, NIKHEF LHCb Tracking and Alignment Workshop, Lausanne, 8-9th November 2006  Motivations  Step-by-step …  Current.
CHEP 2006, February 2006, Mumbai 1 LHCb use of batch systems A.Tsaregorodtsev, CPPM, Marseille HEPiX 2006, 4 April 2006, Rome.
CERN IT Department t LHCb Software Distribution Roberto Santinelli CERN IT/GS.
Status of the Bologna Computing Farm and GRID related activities Vincenzo M. Vagnoni Thursday, 7 March 2002.
Liverpool Experience of MDC 1 MAP (and in our belief any system which attempts to be scaleable to 1000s of nodes) broadcasts the code to all the nodes.
LCG Support for Pilot Jobs John Gordon, STFC GDB December 2 nd 2009.
Workload management, virtualisation, clouds & multicore Andrew Lahiff.
The GridPP DIRAC project DIRAC for non-LHC communities.
DIRAC Pilot Jobs A. Casajus, R. Graciani, A. Tsaregorodtsev for the LHCb DIRAC team Pilot Framework and the DIRAC WMS DIRAC Workload Management System.
WLCG LHCC mini-review LHCb Summary. Outline m Activities in 2008: summary m Status of DIRAC m Activities in 2009: outlook m Resources in PhC2.
Large scale data flow in local and GRID environment Viktor Kolosov (ITEP Moscow) Ivan Korolko (ITEP Moscow)
LHCb report to LHCC and C-RSG Philippe Charpentier CERN on behalf of LHCb.
Computing Issues for the ATLAS SWT2. What is SWT2? SWT2 is the U.S. ATLAS Southwestern Tier 2 Consortium UTA is lead institution, along with University.
OPERATIONS REPORT JUNE – SEPTEMBER 2015 Stefan Roiser CERN.
1 Andrea Sciabà CERN The commissioning of CMS computing centres in the WLCG Grid ACAT November 2008 Erice, Italy Andrea Sciabà S. Belforte, A.
Alignment in real-time in current detector and upgrade 6th LHCb Computing Workshop 18 November 2015 Beat Jost / Cern.
1 LHCb computing for the analysis : a naive user point of view Workshop analyse cc-in2p3 17 avril 2008 Marie-Hélène Schune, LAL-Orsay for LHCb-France Framework,
LHCbDirac and Core Software. LHCbDirac and Core SW Core Software workshop, PhC2 Running Gaudi Applications on the Grid m Application deployment o CVMFS.
LHCb Pilot Job Tests. We have not started this test due to several reasons: our analysis job submission (ganga) was not interfaced in time to DIRAC3,
LHCb status and plans Ph.Charpentier CERN. LHCb status and plans WLCG Workshop 1-2 Sept 2007, Victoria, BC 2 Ph.C. Status of DC06  Reminder:  Two-fold.
LHCbComputing Computing for the LHCb Upgrade. 2 LHCb Upgrade: goal and timescale m LHCb upgrade will be operational after LS2 (~2020) m Increase significantly.
Enabling Grids for E-sciencE CMS/ARDA activity within the CMS distributed system Julia Andreeva, CERN On behalf of ARDA group CHEP06.
8 August 2006MB Report on Status and Progress of SC4 activities 1 MB (Snapshot) Report on Status and Progress of SC4 activities A weekly report is gathered.
Markus Frank (CERN) & Albert Puig (UB).  An opportunity (Motivation)  Adopted approach  Implementation specifics  Status  Conclusions 2.
The GridPP DIRAC project DIRAC for non-LHC communities.
Placeholder ES 1 CERN IT EGI Technical Forum, Experiment Support group AAI usage, issues and wishes for WLCG Maarten Litmaath CERN.
LCG Pilot Jobs + glexec John Gordon, STFC-RAL GDB 7 December 2007.
Monitoring the Readiness and Utilization of the Distributed CMS Computing Facilities XVIII International Conference on Computing in High Energy and Nuclear.
LHCb Computing activities Philippe Charpentier CERN – LHCb On behalf of the LHCb Computing Group.
StoRM + Lustre Proposal YAN Tian On behalf of Distributed Computing Group
Geant4 GRID production Sangwan Kim, Vu Trong Hieu, AD At KISTI.
LHCb LHCb GRID SOLUTION TM Recent and planned changes to the LHCb computing model Marco Cattaneo, Philippe Charpentier, Peter Clarke, Stefan Roiser.
WLCG Operations Coordination Andrea Sciabà IT/SDC GDB 11 th September 2013.
LHCb Computing 2015 Q3 Report Stefan Roiser LHCC Referees Meeting 1 December 2015.
ATLAS – statements of interest (1) A degree of hierarchy between the different computing facilities, with distinct roles at each level –Event filter Online.
Real Time Fake Analysis at PIC
L’analisi in LHCb Angelo Carbone INFN Bologna
Eleonora Luppi INFN and University of Ferrara - Italy
Multi User Pilot Jobs update
Progress on NA61/NA49 software virtualisation Dag Toppe Larsen Wrocław
Work report Xianghu Zhao Nov 11, 2014.
LHCb Computing Model and Data Handling Angelo Carbone 5° workshop italiano sulla fisica p-p ad LHC 31st January 2008.
Philippe Charpentier CERN – LHCb On behalf of the LHCb Computing Group
1 VO User Team Alarm Total ALICE ATLAS CMS
LHCb Conditions Database TEG Workshop 7 November 2011 Marco Clemencic
LHCb Computing Philippe Charpentier CERN
LHCb status and plans Ph.Charpentier CERN.
Simulation and Physics
Production Manager Tools (New Architecture)
The LHCb Computing Data Challenge DC06
Presentation transcript:

LHCb 2009-Q4 report

2009 Q4 report LHCb 2009-Q4 report, PhC2 Activities in 2009-Q4 m Core Software o Stable versions of Gaudi and LCG-AA m Applications o Stable as of September for real data o Fast minor releases to cope with reality of life… m Monte-Carlo o Some MC09 channel simulation o Few events in foreseen 2009 configuration o Minimum bias MC09 stripping m Real data reconstruction o As of November 20 th

2009 Q4 report Jobs in 2009-Q4 LHCb 2009-Q4 report, PhC3

2009 Q4 report First experience with real data m Very low crossing rate o Maximum 8 bunches colliding (88 kHz crossing) o Very low luminosity o Minimum bias trigger rate: from 0.1 to 10 Hz o Data taken with single beam and with collisions LHCb 2009-Q4 report, PhC4

2009 Q4 report Real data processing m Iterative process o Small changes in reconstruction application o Improved alignment o In total 5 sets of processing conditions P Only last files were all processed twice m Processing submission o Automatic job creation and submission after: P File is successfully migrated in Castor P File is successfully replicated at Tier1 o If job fails for a reason other than application crash P The file is reset as “to be processed” P New job is created / submitted o Processing more efficient at CERN (see later) P Eventually after few trials at Tier1, the file is processed at CERN o No stripping ;-) P DST files distributed to all Tier1s for analysis LHCb 2009-Q4 report, PhC5

2009 Q4 report Reconstruction jobs LHCb 2009-Q4 report, PhC6

2009 Q4 report Issues with real data m Castor migration o Very low rate: had to change the migration algorithm for more frequent migration m Issue with large files (above 2 GB) o Real data files are not ROOT files but open by ROOT o There was an issue with a compatibility library for slc4-32 bit on slc5 nodes P Fixed within a day m Wrong magnetic field sign o Due to different coordinate systems for LHCb and LHC ;-) o Fixed within hours m Data access problem (by protocol, directly from server) o Still dCache issue at IN2P3 and NIKHEF P dCache experts working on it o Moved to copy mode paradigm for reconstruction o Still a problem for user jobs P Sites have been banned for analysis LHCb 2009-Q4 report, PhC7

2009 Q4 report Transfers and job latency m No problem observed during file transfers o Files randomly distributed to Tier1 o Will move to distribution by runs (few 100’s files) o For 2009, runs were not longer than 4-5 files! m Very good Grid latency o Time between submission and jobs starting running LHCb 2009-Q4 report, PhC8

2009 Q4 report Brief digression on LHCb position w.r.t. MUPJs m So-called Multi-User Pilot Jobs (MUPJ) are used by DIRAC on all sites that accept role=Pilot o They are just regular jobs! m MUPJs match any Dirac job in the central queue o Production or User analysis (single queue) o Each PJ can execute sequentially up to 5 jobs P If remaining capabilities allow (e.g. CPU time left) P MUPJ has 5 tokens for matching jobs o role=Pilot proxy can only retrieve jobs, limited to 5 m A limited user proxy is used for DM operations of the payload o Cannot be used for job submission m Proxies can be hidden when not needed m DIRAC is instrumented for using gLexec (in any mode) m First experience o Problems are not with gLexec but with SCAS configuration LHCb 2009-Q4 report, PhC9

2009 Q4 report MUPJs (cont’d) m LHCb is not willing to loose efficiency due to the introduction of badly configured gLexec o Yet another point of failure! o Cannot afford testing individually all sites at once m This topic has been lasting over 3 years now o Where is the emergency? Why did it take so long if so important? m Propose to reconsider pragmatically the policy o They were defined when the frameworks had not been evaluated, and VOs had to swallow the bullet o Re-assessing the risks was not really done in the TF (yet) o Questionnaire leaves decisions to sites P We got no message that sites are unhappy with current situation o MUPJs are just jobs for which their owner is responsible P Move responsibility to MUPJ owner (“the VO”) P Two tier trust relation ( Sites / VO / User ) P Apply a posteriori control and not a priori mistrust o VOs should assess the risk on their side LHCb 2009-Q4 report, PhC10

2009 Q4 report Conclusions m Concentrating on real data o Very few data (200 GB)! o Very important learning exercise o A few improvements identified for the 2010 running P Run distribution (rather than files) P Conditions DB synchronization check d Make sure Online Conditions are up-to-date m Still some MC productions o With feedback from first real data P E.g. final position of the VeLo (15 mm from beam) m Analysis o First analysis of 2009 data made on the Grid o Foresee a stripping phase for V 0 physics publications m LHCb definitely wants to continue using MUPJs! LHCb 2009-Q4 report, PhC11