LHCb Readiness for Run WLCG Workshop Okinawa

Slides:



Advertisements
Similar presentations
31/03/00 CMS(UK)Glenn Patrick What is the CMS(UK) Data Model? Assume that CMS software is available at every UK institute connected by some infrastructure.
Advertisements

WLCG Interaction Stefan Roiser LHCb Computing Operations Workshop 27 Jan ‘15.
Grid and CDB Janusz Martyniak, Imperial College London MICE CM37 Analysis, Software and Reconstruction.
1 Bridging Clouds with CernVM: ATLAS/PanDA example Wenjing Wu
Large scale data flow in local and GRID environment V.Kolosov, I.Korolko, S.Makarychev ITEP Moscow.
Pilots 2.0: DIRAC pilots for all the skies Federico Stagni, A.McNab, C.Luzzi, A.Tsaregorodtsev On behalf of the DIRAC consortium and the LHCb collaboration.
LHCb Quarterly Report October Core Software (Gaudi) m Stable version was ready for 2008 data taking o Gaudi based on latest LCG 55a o Applications.
Ian Fisk and Maria Girone Improvements in the CMS Computing System from Run2 CHEP 2015 Ian Fisk and Maria Girone For CMS Collaboration.
Status of CMS Matthew Nguyen Recontres LCG-France December 1 st, 2014 *Mostly based on information from CMS Offline & Computing Week November 3-7.
Computing and LHCb Raja Nandakumar. The LHCb experiment  Universe is made of matter  Still not clear why  Andrei Sakharov’s theory of cp-violation.
CHEP – Mumbai, February 2006 The LCG Service Challenges Focus on SC3 Re-run; Outlook for 2006 Jamie Shiers, LCG Service Manager.
Building a distributed software environment for CDF within the ESLEA framework V. Bartsch, M. Lancaster University College London.
Computing Infrastructure Status. LHCb Computing Status LHCb LHCC mini-review, February The LHCb Computing Model: a reminder m Simulation is using.
1 Kittikul Kovitanggoon*, Burin Asavapibhop, Narumon Suwonjandee, Gurpreet Singh Chulalongkorn University, Thailand July 23, 2015 Workshop on e-Science.
Cosener’s House – 30 th Jan’031 LHCb Progress & Plans Nick Brook University of Bristol News & User Plans Technical Progress Review of deliverables.
Bookkeeping Tutorial. Bookkeeping & Monitoring Tutorial2 Bookkeeping content  Contains records of all “jobs” and all “files” that are created by production.
November SC06 Tampa F.Fanzago CRAB a user-friendly tool for CMS distributed analysis Federica Fanzago INFN-PADOVA for CRAB team.
Tier-2  Data Analysis  MC simulation  Import data from Tier-1 and export MC data CMS GRID COMPUTING AT THE SPANISH TIER-1 AND TIER-2 SITES P. Garcia-Abia.
CCRC’08 Weekly Update Jamie Shiers ~~~ LCG MB, 1 st April 2008.
1 LCG-France sites contribution to the LHC activities in 2007 A.Tsaregorodtsev, CPPM, Marseille 14 January 2008, LCG-France Direction.
The LHCb Italian Tier-2 Domenico Galli, Bologna INFN CSN1 Roma,
Stefano Belforte INFN Trieste 1 CMS Simulation at Tier2 June 12, 2006 Simulation (Monte Carlo) Production for CMS Stefano Belforte WLCG-Tier2 workshop.
LHCb The LHCb Data Management System Philippe Charpentier CERN On behalf of the LHCb Collaboration.
The LHCb CERN R. Graciani (U. de Barcelona, Spain) for the LHCb Collaboration International ICFA Workshop on Digital Divide Mexico City, October.
Claudio Grandi INFN Bologna CMS Computing Model Evolution Claudio Grandi INFN Bologna On behalf of the CMS Collaboration.
Bookkeeping Tutorial. 2 Bookkeeping content  Contains records of all “jobs” and all “files” that are produced by production jobs  Job:  In fact technically.
Predrag Buncic Future IT challenges for ALICE Technical Workshop November 6, 2015.
LHCbComputing LHCC status report. Operations June 2014 to September m Running jobs by activity o Montecarlo simulation continues as main activity.
LHCbComputing Resources requests : changes since LHCb-PUB (March 2013) m Assume no further reprocessing of Run I data o (In.
The GridPP DIRAC project DIRAC for non-LHC communities.
Data Placement Intro Dirk Duellmann WLCG TEG Workshop Amsterdam 24. Jan 2012.
LHCb report to LHCC and C-RSG Philippe Charpentier CERN on behalf of LHCb.
David Stickland CMS Core Software and Computing
Workflows and Data Management. Workflow and DM Run3 and after: conditions m LHCb major upgrade is for Run3 (2020 horizon)! o Luminosity x 5 ( )
OPERATIONS REPORT JUNE – SEPTEMBER 2015 Stefan Roiser CERN.
1 Andrea Sciabà CERN The commissioning of CMS computing centres in the WLCG Grid ACAT November 2008 Erice, Italy Andrea Sciabà S. Belforte, A.
ATLAS Distributed Computing perspectives for Run-2 Simone Campana CERN-IT/SDC on behalf of ADC.
Handling of T1D0 in CCRC’08 Tier-0 data handling Tier-1 data handling Experiment data handling Reprocessing Recalling files from tape Tier-0 data handling,
LHCbDirac and Core Software. LHCbDirac and Core SW Core Software workshop, PhC2 Running Gaudi Applications on the Grid m Application deployment o CVMFS.
Maria Girone, CERN CMS Experiment Status, Run II Plans, & Federated Requirements Maria Girone, CERN XrootD Workshop, January 27, 2015.
LHCbComputing Computing for the LHCb Upgrade. 2 LHCb Upgrade: goal and timescale m LHCb upgrade will be operational after LS2 (~2020) m Increase significantly.
CMS: T1 Disk/Tape separation Nicolò Magini, CERN IT/SDC Oliver Gutsche, FNAL November 11 th 2013.
Markus Frank (CERN) & Albert Puig (UB).  An opportunity (Motivation)  Adopted approach  Implementation specifics  Status  Conclusions 2.
ATLAS Distributed Computing ATLAS session WLCG pre-CHEP Workshop New York May 19-20, 2012 Alexei Klimentov Stephane Jezequel Ikuo Ueda For ATLAS Distributed.
Computing Model José M. Hernández CIEMAT, Madrid On behalf of the CMS Collaboration XV International Conference on Computing in High Energy and Nuclear.
16 September 2014 Ian Bird; SPC1. General ALICE and LHCb detector upgrades during LS2  Plans for changing computing strategies more advanced CMS and.
Meeting with University of Malta| CERN, May 18, 2015 | Predrag Buncic ALICE Computing in Run 2+ P. Buncic 1.
LHCb Computing activities Philippe Charpentier CERN – LHCb On behalf of the LHCb Computing Group.
Jianming Qian, UM/DØ Software & Computing Where we are now Where we want to go Overview Director’s Review, June 5, 2002.
ALICE Physics Data Challenge ’05 and LCG Service Challenge 3 Latchezar Betev / ALICE Geneva, 6 April 2005 LCG Storage Management Workshop.
Joe Foster 1 Two questions about datasets: –How do you find datasets with the processes, cuts, conditions you need for your analysis? –How do.
Apr. 25, 2002Why DØRAC? DØRAC FTFM, Jae Yu 1 What do we want DØ Regional Analysis Centers (DØRAC) do? Why do we need a DØRAC? What do we want a DØRAC do?
Computing infrastructures for the LHC: current status and challenges of the High Luminosity LHC future Worldwide LHC Computing Grid (WLCG): Distributed.
LHCbComputing LHCb computing model in Run1 & Run2 Concezio Bozzi Bologna, Feb 19 th 2015.
LHCb LHCb GRID SOLUTION TM Recent and planned changes to the LHCb computing model Marco Cattaneo, Philippe Charpentier, Peter Clarke, Stefan Roiser.
LHCb Computing 2015 Q3 Report Stefan Roiser LHCC Referees Meeting 1 December 2015.
ATLAS – statements of interest (1) A degree of hierarchy between the different computing facilities, with distinct roles at each level –Event filter Online.
LHCb distributed computing during the LHC Runs 1,2 and 3
Ian Bird WLCG Workshop San Francisco, 8th October 2016
Report from WLCG Workshop 2017: WLCG Network Requirements GDB - CERN 12th of July 2017
LHCb Software & Computing Status
LHCb Computing Model and Data Handling Angelo Carbone 5° workshop italiano sulla fisica p-p ad LHC 31st January 2008.
Philippe Charpentier CERN – LHCb On behalf of the LHCb Computing Group
Dagmar Adamova (NPI AS CR Prague/Rez) and Maarten Litmaath (CERN)
ALICE Computing Model in Run3
R. Graciani for LHCb Mumbay, Feb 2006
LHCb Computing Philippe Charpentier CERN
ATLAS DC2 & Continuous production
The LHCb Computing Data Challenge DC06
Presentation transcript:

LHCb Readiness for Run 2 2015 WLCG Workshop Okinawa Stefan Roiser / CERN IT-SDC for LHCb Distributed Computing

LHCb Run2 Readiness - StR Content Online changes with impact for Offline Offline Data Processing Offline Data Management Services LHCb relies on 11 April '15 LHCb Run2 Readiness - StR

Preamble – LHC Evolution Run 1 Planned for Run 2 Max beam energy 4 TeV 6.5 TeV Transverse beam emittance 1.8 μm 1.9 μm β* (beam oscillation) 0.6 m / LHCb 3 m 0.4 m / LHCb 3 m Number of bunches 1374 2508 Max protons per bunch 1.7 * 1011 1.15 * 1011 Bunch spacing 50 ns 25 ns LHC Maximum Luminosity 7.7 * 1033 cm-2s-1 1.6 * 1034 cm-2s-1 LHCb Maximum Luminosity 4 * 1032 cm-2s-1 LHCb μ (avg # collisions/crossing) 1.6 1.2 ATLAS & CMS LHCb NB: LHCb uses “luminosity leveling”, ie. the “in time pile up” and so the instantaneous luminosity stays constant for LHCb during an LHC fill 11 April '15 LHCb Run2 Readiness - StR

Pit & Online

LHCb Run2 Readiness - StR Trigger Scheme Hardware trigger reduce event rate to ~ 1 MHz High Level Trigger computing farm split into HLT1 with partial event reconstruction, output will be buffered on local disks HLT 1 output used for detector calibration and alignment (O(hours)). (was done offline in Run 1) HLT2 runs deferred with signal event reconstruction very close to offline reconstruction 12.5 kHz event rate to OFFLINE At ~ 60kB event size this is ~ 750 MB/s Event rate was 4.5 kHz in Run 1 NB: Because of deferred trigger, very little availability of HLT for offline data processing See also Marco’s talk tomorrow on further evolution for future Runs 11 April '15 LHCb Run2 Readiness - StR

HLT Output Stream Splitting 12.5 kHz to Storage 10 kHz Full (+Parked) Stream 2.5 kHz Turbo Stream 10 kHz go to classic Offline reconstruction / stripping on distributed computing resources If needed part of this can be “parked” and processed in LS 2 New concept of “Turbo Stream” in Run 2 for ~ 2.5 kHZ i.e. wherever sufficient, take the HLT output with its event reconstruction directly for physics analysis Initially RAW information included, will be stripped off Offline S. Benson, “The LHCb Turbo Stream”, T1, Village Center, Thu 10am, 11 April '15 LHCb Run2 Readiness - StR

Data Processing

LHCb Run2 Readiness - StR Legend: Offline Processing Workflow Application File Type Storage Element (M)DST Stripping RAW Reconstr. X 5GB, 1x FULL.DST Tape X RAW 24h Reconstr. 5GB, 1x FULL.DST Buffer 3GB, 1x Buffer 6h Stripping The RAW input file is available on Disk Buffer Reconstruction runs ~ 24 h, 1 input RAW, 1 output FULL.DST to Disk Buffer Asynchronous migration of FULL.DST from Disk Buffer to Tape Stripping (DaVinci) runs on 1 or 2 input files (~ 6h/file), output several unmerged (M)DST files (one per “stream”) to Disk Buffer Input FULL.DST removed from Disk Buffer asynchronously Rerun the above workflows for one run Once a stream reaches 5 GB of unmerged (M)DSTs (up to O(100) files), Merging runs ~ 15 – 30 mins, output one merged (M)DST file to Disk Input (M)DST files removed from Disk Buffer asynchronously … unmerged (M)DST O(MB) 1x Buffer X Merging 30m (M)DST … 5GB, 1x Disk 27 Jan '15 11 April '15 Run2 Operations - StR LHCb Run2 Readiness - StR 8

Offline Data Processing Changes What is reconstructed offline is supposed to be the final reconstruction pass Calibration / Alignment from HLT used also offline No reprocessing (reco) foreseen before end of Run 2 Expecting a higher stripping retention because of calibration and alignment done ONLINE Partly damped by moving most physics streams to M(icro)DST format (Note: MDST O(10kB/Event), DST O(120kB/Event)) All files from one “LHCb Run” are forced to reside on the same storage A run is the smallest granularity for physics analysis files E.g. would reduce impact in case a disk breaks Workflow execution is now also possible on Tier 2 sites Needed because of increase of collected data 11 April '15 LHCb Run2 Readiness - StR

Workflow Execution Location GRIDKA RAW Reco Strppg Merge FULL.DST unm. DST DST CNAF RAW Reco FULL.DST Manchester RAW Reco FULL.DST Manchester RAW Reco FULL.DST CNAF RAW Reco Strippg Merge FULL.DST unm. DST DST X X Data Processing workflow executed by default at Tier 0/1 sites (stays the same as in Run 1) For Run 2 in addition we allow A Tier 2 site to participate for a certain Job Type remotely (most useful would be Reconstruction) Any Tier 2 is allowed at any time to participate on any Job Type (no static 1 to 1 “attaching” anymore) In principle the system also allows for ANY site to participate on any Job Type remotely Monarc (technically) 11 April '15 LHCb Run2 Readiness - StR

LHCb Run2 Readiness - StR All Workflows Run 1 Run 2 Data Processing T 0 / 1 T 0 / 1 / 2 Monte Carlo T 2 (can also run on T 0 / 1 sites if resources available) User analysis T 0 / 1 / 2D (without input data can also run on T2) Very flexible computing model allows almost all workflows to be executed on every tier level / resource type Interested in running multicore jobs – especially on VMs – but no pressing need for it “Elastic MC” – knows the work / event, at start of the payload will calculate on the fly how many events to produce until the “end of the queue” User analysis – least amount of work but highest priority in the central task queue F. Stagni, “Jobs masonry with elastic Grid Jobs”, T4, B250, Mo 5pm 11 April '15 LHCb Run2 Readiness - StR

LHCb Run2 Readiness - StR Compute Resources Non virtualized Virtualized “Classic Grid” – CE, batch system, … BOINC – volunteer computing Non pledged – commercial, HPC, … Vac – self managed cloud resources HLT farm – little use during Run 2 Vcycle – interaction via IaaS A. McNab, “Managing virtual machines with Vac and Vcycle”, T7, C210, Mo 5pm Expect rampup of virtualized infrastructures during Run 2 All environments served by the same pilot infrastructure talking to one LHCb/DIRAC central task queue F. Stagni, “Pilots 2.0: DIRAC pilots for all the skies”, T4, B250, Mo 2pm 11 April '15 LHCb Run2 Readiness - StR

Data Management

LHCb Run2 Readiness - StR Data Storage Introduced concept of Tier 2D(isk) sites i.e. Tier 2 sites with disk areas >= 300 TB No more direct processing from “tape caches” foreseen Interact with disk buffer via FTS3 and process from there E.g. pre-staging “Legacy Run 1 Stripping” data Should lead to reduction of disk cache size in front of tape 11 April '15 LHCb Run2 Readiness - StR

LHCb Run2 Readiness - StR Data Storage (ctd) C. Haen, “Federating LHCb datasets using the Dirac File Catalog”, T3, C209, Mo 4.45pm Catalogs File Catalog: provides replica information, recently migrated from LCG File Catalog to the Dirac File Catalog Bookkeeping (unchanged) provides data provenance information Data Popularity Data collected since 2012 M. Hushchyn, “Disk storage mgmt for LHCb on Data Popularity”, T3, C209, Tue 6.15pm 11 April '15 LHCb Run2 Readiness - StR

Data Access Operations Gaudi Federation In use since last fall. LHCb analysis jobs create a local replica catalog for their input data. If the local copy is not available -> fall back to next remote replica. Data access protocols SRM Shall be in use for tape interactions … and for writing to storage (job output upload, data replication) Xroot LHCb will be constructing turls for input data on the fly without SRM interaction for disk resident data access Need single and stable xroot endpoint per storage element HTTP/WEBDAV All storage sites are equipped with Http/Webdav access Could be used as second access protocol 11 April '15 LHCb Run2 Readiness - StR

Underlying Services

LHCb Run2 Readiness - StR Services CVMFS Building block for LHCb distributed computing, distributes all software and conditions data CernVM Vac, vcycle, BOINC are using CernVM 3 FTS 3 Vital for LHCb WAN transfers and tape interaction (pre-staging of input data) Several WLCG monitoring services in use SAM 3, dashboards, network monitoring Working on perfSonar data extraction into LHCbDIRAC HTTP Federation Builds on top of http/webdav access, provides easy access to LHCb data namespace Development on top ongoing for data consistency checks F. Furano, “Seamless access to LHCb HTTP/WebDAV storage”, Mo/Tue, Poster Sess. A 11 April '15 LHCb Run2 Readiness - StR

LHCb Run2 Readiness - StR Summary LHCb is ready for Run 2 Several changes introduced in Run 2 Calibration/Alignment in the HLT farm Closer integration of Tier 2 sites in data processing New Dirac file replica catalog deployed Disk resident data access via direct xroot CVMFS and FTS3 are key “external” services 11 April '15 LHCb Run2 Readiness - StR

LHCb Run2 Readiness - StR Goodie page http://lhcb-web-dirac.cern.ch/DIRAC/LHCb-Production/undefined/grid/SiteStatus/display?name=LCG.RAL.uk 11 April '15 LHCb Run2 Readiness - StR