Status of the LHCb MC production system Andrei Tsaregorodtsev, CPPM, Marseille DataGRID France workshop, Marseille, 24 September 2002.

Slides:



Advertisements
Similar presentations
GridPP July 2003Stefan StonjekSlide 1 SAM middleware components Stefan Stonjek University of Oxford 7 th GridPP Meeting 02 nd July 2003 Oxford.
Advertisements

LHCb Computing Activities in UK Current activities UK GRID activities RICH s/w activities.
Réunion DataGrid France, Lyon, fév CMS test of EDG Testbed Production MC CMS Objectifs Résultats Conclusions et perspectives C. Charlot / LLR-École.
Grid and CDB Janusz Martyniak, Imperial College London MICE CM37 Analysis, Software and Reconstruction.
1 Grid services based architectures Growing consensus that Grid services is the right concept for building the computing grids; Recent ARDA work has provoked.
6/4/20151 Introduction LHCb experiment. LHCb experiment. Common schema of the LHCb computing organisation. Common schema of the LHCb computing organisation.
Production Planning Eric van Herwijnen Thursday, 20 june 2002.
K.Harrison CERN, 23rd October 2002 HOW TO COMMISSION A NEW CENTRE FOR LHCb PRODUCTION - Overview of LHCb distributed production system - Configuration.
11 Dec 2000F Harris Datagrid Testbed meeting at Milan 1 LHCb ‘use-case’ - distributed MC production
Exploiting the Grid to Simulate and Design the LHCb Experiment K Harrison 1, N Brook 2, G Patrick 3, E van Herwijnen 4, on behalf of the LHCb Grid Group.
Bookkeeping data Monitoring info Get jobs Site A Site B Site C Site D Agent Production service Monitoring service Bookkeeping service Agent © Andrei Tsaregorodtsev.
QCDgrid Technology James Perry, George Beckett, Lorna Smith EPCC, The University Of Edinburgh.
CMS Report – GridPP Collaboration Meeting VI Peter Hobson, Brunel University30/1/2003 CMS Status and Plans Progress towards GridPP milestones Workload.
5 November 2001F Harris GridPP Edinburgh 1 WP8 status for validating Testbed1 and middleware F Harris(LHCb/Oxford)
The SAMGrid Data Handling System Outline:  What Is SAMGrid?  Use Cases for SAMGrid in Run II Experiments  Current Operational Load  Stress Testing.
3 Sept 2001F HARRIS CHEP, Beijing 1 Moving the LHCb Monte Carlo production system to the GRID D.Galli,U.Marconi,V.Vagnoni INFN Bologna N Brook Bristol.
LHCb Applications and GRID Integration Domenico Galli Catania, April 9, st INFN-GRID Workshop.
November 7, 2001Dutch Datagrid SARA 1 DØ Monte Carlo Challenge A HEP Application.
Computing Infrastructure Status. LHCb Computing Status LHCb LHCC mini-review, February The LHCb Computing Model: a reminder m Simulation is using.
3rd June 2004 CDF Grid SAM:Metadata and Middleware Components Mòrag Burgon-Lyon University of Glasgow.
SLICE Simulation for LHCb and Integrated Control Environment Gennady Kuznetsov & Glenn Patrick (RAL) Cosener’s House Workshop 23 rd May 2002.
Grid Workload Management & Condor Massimo Sgaravatto INFN Padova.
Cosener’s House – 30 th Jan’031 LHCb Progress & Plans Nick Brook University of Bristol News & User Plans Technical Progress Review of deliverables.
Nick Brook Current status Future Collaboration Plans Future UK plans.
1 st December 2003 JIM for CDF 1 JIM and SAMGrid for CDF Mòrag Burgon-Lyon University of Glasgow.
1 DIRAC – LHCb MC production system A.Tsaregorodtsev, CPPM, Marseille For the LHCb Data Management team CHEP, La Jolla 25 March 2003.
Belle MC Production on Grid 2 nd Open Meeting of the SuperKEKB Collaboration Soft/Comp session 17 March, 2009 Hideyuki Nakazawa National Central University.
LHCb and DataGRID - the workplan for 2001 Eric van Herwijnen Wednesday, 28 march 2001.
11 December 2000 Paolo Capiluppi - DataGrid Testbed Workshop CMS Applications Requirements DataGrid Testbed Workshop Milano, 11 December 2000 Paolo Capiluppi,
ATLAS and GridPP GridPP Collaboration Meeting, Edinburgh, 5 th November 2001 RWL Jones, Lancaster University.
MAGDA Roger Jones UCL 16 th December RWL Jones, Lancaster University MAGDA  Main authors: Wensheng Deng, Torre Wenaus Wensheng DengTorre WenausWensheng.
7April 2000F Harris LHCb Software Workshop 1 LHCb planning on EU GRID activities (for discussion) F Harris.
LHCb planning for DataGRID testbed0 Eric van Herwijnen Thursday, 10 may 2001.
29 May 2002Joint EDG/WP8-EDT/WP4 MeetingClaudio Grandi INFN Bologna LHC Experiments Grid Integration Plans C.Grandi INFN - Bologna.
4/5/2007Data handling and transfer in the LHCb experiment1 Data handling and transfer in the LHCb experiment RT NPSS Real Time 2007 FNAL - 4 th May 2007.
RAL Site Report John Gordon IT Department, CLRC/RAL HEPiX Meeting, JLAB, October 2000.
WP8 Meeting Glenn Patrick1 LHCb Grid Activities in UK Grid WP8 Meeting, 16th November 2000 Glenn Patrick (RAL)
Giuseppe Codispoti INFN - Bologna Egee User ForumMarch 2th BOSS: the CMS interface for job summission, monitoring and bookkeeping W. Bacchi, P.
And Tier 3 monitoring Tier 3 Ivan Kadochnikov LIT JINR
LHCb Software Week November 2003 Gennady Kuznetsov Production Manager Tools (New Architecture)
EGEE is a project funded by the European Union under contract IST “Interfacing to the gLite Prototype” Andrew Maier / CERN LCG-SC2, 13 August.
UK Grid Meeting Glenn Patrick1 LHCb Grid Activities in UK Grid Prototype and Globus Technical Meeting QMW, 22nd November 2000 Glenn Patrick (RAL)
UTA MC Production Farm & Grid Computing Activities Jae Yu UT Arlington DØRACE Workshop Feb. 12, 2002 UTA DØMC Farm MCFARM Job control and packaging software.
2 Sep 2002F Harris EDG/WP6 meeeting at Budapest LHC experiments use of EDG Testbed F Harris (Oxford/CERN)
Large scale data flow in local and GRID environment Viktor Kolosov (ITEP Moscow) Ivan Korolko (ITEP Moscow)
ATLAS-specific functionality in Ganga - Requirements for distributed analysis - ATLAS considerations - DIAL submission from Ganga - Graphical interfaces.
LHCb Data Challenge in 2002 A.Tsaregorodtsev, CPPM, Marseille DataGRID France meeting, Lyon, 18 April 2002.
Pavel Nevski DDM Workshop BNL, September 27, 2006 JOB DEFINITION as a part of Production.
INFSO-RI Enabling Grids for E-sciencE Using of GANGA interface for Athena applications A. Zalite / PNPI.
LHCb GRID Meeting 11/12 Sept Sept LHCb-GRID T. Bowcock 2 AGENDA 9:30 LHCb MC Production –Points SICB Processing Req. Data Storage Data Transfer.
CLRC Grid Team Glenn Patrick LHCb GRID Plans Glenn Patrick LHCb has formed a GRID technical working group to co-ordinate practical Grid.
Markus Frank (CERN) & Albert Puig (UB).  An opportunity (Motivation)  Adopted approach  Implementation specifics  Status  Conclusions 2.
D.Spiga, L.Servoli, L.Faina INFN & University of Perugia CRAB WorkFlow : CRAB: CMS Remote Analysis Builder A CMS specific tool written in python and developed.
The GridPP DIRAC project DIRAC for non-LHC communities.
Joe Foster 1 Two questions about datasets: –How do you find datasets with the processes, cuts, conditions you need for your analysis? –How do.
LHCb computing model and the planned exploitation of the GRID Eric van Herwijnen, Frank Harris Monday, 17 July 2000.
CMS High Level Trigger Configuration Management
Moving the LHCb Monte Carlo production system to the GRID
BOSS: the CMS interface for job summission, monitoring and bookkeeping
LHCb Computing Model and Data Handling Angelo Carbone 5° workshop italiano sulla fisica p-p ad LHC 31st January 2008.
BOSS: the CMS interface for job summission, monitoring and bookkeeping
D. Galli, U. Marconi, V. Vagnoni INFN Bologna N. Brook Bristol
The Ganga User Interface for Physics Analysis on Distributed Resources
LHCb Distributed Computing and the Grid V. Vagnoni (INFN Bologna)
Gridifying the LHCb Monte Carlo production system
Status and plans for bookkeeping system and production tools
Short to middle term GRID deployment plan for LHCb
Production Manager Tools (New Architecture)
Production client status
Presentation transcript:

Status of the LHCb MC production system Andrei Tsaregorodtsev, CPPM, Marseille DataGRID France workshop, Marseille, 24 September 2002

Contents  Introduction  Production status  Data management  Job submission  Data production monitoring  Bookkeeping  DataGrid involvement status and plans  Conclusions

Introduction  LHCb – one of the 4 LHC experiments for studies of the CP violation phenomena in the Beauty system;  The experiment is reconsidering the general setup now: The revised TDR due September 2003; Will be based on the large volume MC data; An adequate MC production system is being set up now.

Summer 2002 production = Data Challenge 1  In 48 days we produced 3.3 M events ( 6 TB of data ) We have shown we can produced ~70 K per day We can expect to produce ~ 100 K per day when all centres are operational. Usual job (sim+reco) of 500 events: 390s/evt (I.e. 55 hours of CPU ! More than T class limit in CC) SICBMCsimulation(s/evt)Brunelreconstruction(s/evt)RAWHGeant(KB/evt)OODSTReco.(KB/evt) Min bias BB incl

Centers productivity (summer 2002 run)  Contributions (events produced): CERN (1452 k) Bologna (1009 k) Lyon (595 k) RAL (130 k) Cambridge (37 k) Moscow (27 k) Amsterdam/VU (25 k) Rio (24 k) Oxford (24 k)

Some Data Challenge 1 lessons  Thorough data quality checks on each step: Formal based on log files analysis; Informal based on a small analysis job for the produced data;  Crash trace back;  Flexible workflows should be possible;  Production centre dependancies should be as limited as possible;  Bookkeeping: Maintaining integrity; Managing distributed replicas.

Next Data Challenge 2  MC production = Physics Data Challenge;  Volume: DC2 = 10 x DC1;  Available capacity seems to match requirements: ~1000 CPU world wide during 5 months;  Planning for the DC2: Production software ready by mid Nov 2002; Preproduction: mid Dec 2002 – mid Jan 2003; Production: Feb – May 2003.

Status remote centres (autumn 2002) CenterNo. of CPU’s (1 GHz)Production tools CERN~ 400new Lyon60 +new Liverpool~ 120new Imperial College~ 100new DataGrid~ 20new RAL~ 300old Bologna~ 200old Nikhef~ 20old Bristol~ 20old Edinburgh~ 120old Cambridge~ 15old Oxford~ 10old Moscow~ 40old Rio~ 20old Total~ 1000 (outside CERN)

Data Management

Job workflow  Workflow consists of several steps with various executables running in a certain order

Job configuration  Workflow definition: Executables; Input data types; Output data types;  Job step parameters: Executable algorithm parameters; Statistics; Monitoring mode.

Algorithms configuration

Job submission Storage Local production demon Production center Bookkeeping DB Data Production DB CERN Monitoring service Castor Job scripts Production service Bookkeeping service

Local production demon (at a production center)  Customized for the particular center;  Checks availability of the local resources;  Gets jobs scripts from the Production service;  Installs the necessary software if needed;  Submits jobs;  Updates job status in the Production DB;  Checks the jobs output;  Initiates data transfer to CERN/Castor;  Updates the Bookkeeping database;  Technology: Python; XML-RPC, can easily migrate to SOAP;

Data Production and Job Monitoring PVSS

Bookkeeping  Bookkeeping DB available via a Web service interface: XML-RPC server; ODBC mediated persistant back-end (ORACLE, MySQL)  Flexible schema: Allow easy addition of new data types, parameters; Handles distributed dataset replicas;  Web based user GUI is in the works.

DataGrid status and plans  Installation operational  Long job problem fixed  Long file transfer problem (~ 1 Gb)  New production tools being installed  Test: Run 500 event MC generation Store on SE Recover logs and histograms to CERN Run reconstruction. Output to SE. Recover log files and histos. Write recon output to mass store (Castor) Read Castor data with an analysis job outside Grid

Conclusions  Data Challenge 1 in summer 2002 – the need to upgrade the production tools;  Deployment of the new Data Management tools: Oct-Dec 2002 to support Data Challenge 2 production;  DC2 in Dec 2002 – May 2002;  DataGRID facilities will be used in DC2.