LHCb Computing Model and updated requirements John Harvey.

Slides:



Advertisements
Similar presentations
31/03/00 CMS(UK)Glenn Patrick What is the CMS(UK) Data Model? Assume that CMS software is available at every UK institute connected by some infrastructure.
Advertisements

Status GridKa & ALICE T2 in Germany Kilian Schwarz GSI Darmstadt.
Resources for the ATLAS Offline Computing Basis for the Estimates ATLAS Distributed Computing Model Cost Estimates Present Status Sharing of Resources.
T1 at LBL/NERSC/OAK RIDGE General principles. RAW data flow T0 disk buffer DAQ & HLT CERN Tape AliEn FC Raw data Condition & Calibration & data DB disk.
23/04/2008VLVnT08, Toulon, FR, April 2008, M. Stavrianakou, NESTOR-NOA 1 First thoughts for KM3Net on-shore data storage and distribution Facilities VLV.
The LHCb DAQ and Trigger Systems: recent updates Ricardo Graciani XXXIV International Meeting on Fundamental Physics.
Title US-CMS User Facilities Vivian O’Dell US CMS Physics Meeting May 18, 2001.
1 Data Storage MICE DAQ Workshop 10 th February 2006 Malcolm Ellis & Paul Kyberd.
DATA PRESERVATION IN ALICE FEDERICO CARMINATI. MOTIVATION ALICE is a 150 M CHF investment by a large scientific community The ALICE data is unique and.
August 98 1 Jürgen Knobloch ATLAS Software Workshop Ann Arbor ATLAS Computing Planning ATLAS Software Workshop August 1998 Jürgen Knobloch Slides also.
CERN/IT/DB Multi-PB Distributed Databases Jamie Shiers IT Division, DB Group, CERN, Geneva, Switzerland February 2001.
Large scale data flow in local and GRID environment V.Kolosov, I.Korolko, S.Makarychev ITEP Moscow.
Hall D Online Data Acquisition CEBAF provides us with a tremendous scientific opportunity for understanding one of the fundamental forces of nature. 75.
Planning LHCb computing infrastructure 22 May 2000 Slide 1 Planning LHCb computing infrastructure at CERN and at regional centres F. Harris.
CMS Report – GridPP Collaboration Meeting VI Peter Hobson, Brunel University30/1/2003 CMS Status and Plans Progress towards GridPP milestones Workload.
December 17th 2008RAL PPD Computing Christmas Lectures 11 ATLAS Distributed Computing Stephen Burke RAL.
Computing Infrastructure Status. LHCb Computing Status LHCb LHCC mini-review, February The LHCb Computing Model: a reminder m Simulation is using.
LHCb computing in Russia Ivan Korolko (ITEP Moscow) Russia-CERN JWGC, October 2005.
Data Import Data Export Mass Storage & Disk Servers Database Servers Tapes Network from CERN Network from Tier 2 and simulation centers Physics Software.
Copyright © 2000 OPNET Technologies, Inc. Title – 1 Distributed Trigger System for the LHC experiments Krzysztof Korcyl ATLAS experiment laboratory H.
Fermilab User Facility US-CMS User Facility and Regional Center at Fermilab Matthias Kasemann FNAL.
LHC: ATLAS Experiment meeting “Conditions” data challenge Elizabeth Gallas - Oxford - August 29, 2009 XLDB3.
LHC Computing Review - Resources ATLAS Resource Issues John Huth Harvard University.
Finnish DataGrid meeting, CSC, Otaniemi, V. Karimäki (HIP) DataGrid meeting, CSC V. Karimäki (HIP) V. Karimäki (HIP) Otaniemi, 28 August, 2000.
ALICE Upgrade for Run3: Computing HL-LHC Trigger, Online and Offline Computing Working Group Topical Workshop Sep 5 th 2014.
Modeling Regional Centers with MONARC Simulation Tools Modeling LHC Regional Centers with the MONARC Simulation Tools Irwin Gaines, FNAL for the MONARC.
LHC Computing Review Recommendations John Harvey CERN/EP March 28 th, th LHCb Software Week.
Computing for LHCb-Italy Domenico Galli, Umberto Marconi and Vincenzo Vagnoni Genève, January 17, 2001.
7April 2000F Harris LHCb Software Workshop 1 LHCb planning on EU GRID activities (for discussion) F Harris.
December 10,1999: MONARC Plenary Meeting Harvey Newman (CIT) Phase 3 Letter of Intent (1/2)  Short: N Pages è May Refer to MONARC Internal Notes to Document.
Status of the LHCb MC production system Andrei Tsaregorodtsev, CPPM, Marseille DataGRID France workshop, Marseille, 24 September 2002.
Meeting, 5/12/06 CMS T1/T2 Estimates à CMS perspective: n Part of a wider process of resource estimation n Top-down Computing.
EGEE is a project funded by the European Union under contract IST HEP Use Cases for Grid Computing J. A. Templon Undecided (NIKHEF) Grid Tutorial,
ATLAS Data Challenges US ATLAS Physics & Computing ANL October 30th 2001 Gilbert Poulard CERN EP-ATC.
The LHCb CERN R. Graciani (U. de Barcelona, Spain) for the LHCb Collaboration International ICFA Workshop on Digital Divide Mexico City, October.
ATLAS WAN Requirements at BNL Slides Extracted From Presentation Given By Bruce G. Gibbard 13 December 2004.
Report to Worldwide Analysis Review Panel J. Harvey 23 March 2000 Slide 1 LHCb Processing requirements Focus on the first year of data-taking Report to.
Integration of the ATLAS Tag Database with Data Management and Analysis Components Caitriana Nicholson University of Glasgow 3 rd September 2007 CHEP,
Computing R&D and Milestones LHCb Plenary June 18th, 1998 These slides are on WWW at:
5 June 2003Alan Norton / Focus / EP Topics1 Other EP Topics Some 2003 Running Experiments - NA48/2 (Flavio Marchetto) - Compass (Benigno Gobbo) - NA60.
Large scale data flow in local and GRID environment Viktor Kolosov (ITEP Moscow) Ivan Korolko (ITEP Moscow)
The ATLAS Computing Model and USATLAS Tier-2/Tier-3 Meeting Shawn McKee University of Michigan Joint Techs, FNAL July 16 th, 2007.
David Stickland CMS Core Software and Computing
Computing Issues for the ATLAS SWT2. What is SWT2? SWT2 is the U.S. ATLAS Southwestern Tier 2 Consortium UTA is lead institution, along with University.
Predrag Buncic ALICE Status Report LHCC Referee Meeting CERN
LHCb datasets and processing stages. 200 kB100 kB 70 kB 0.1 kB 10kB 150 kB 0.1 kB 200 Hz LHCb datasets and processing stages.
The MEG Offline Project General Architecture Offline Organization Responsibilities Milestones PSI 2/7/2004Corrado Gatto INFN.
CLRC Grid Team Glenn Patrick LHCb GRID Plans Glenn Patrick LHCb has formed a GRID technical working group to co-ordinate practical Grid.
LHCbComputing Computing for the LHCb Upgrade. 2 LHCb Upgrade: goal and timescale m LHCb upgrade will be operational after LS2 (~2020) m Increase significantly.
Distributed Physics Analysis Past, Present, and Future Kaushik De University of Texas at Arlington (ATLAS & D0 Collaborations) ICHEP’06, Moscow July 29,
The ATLAS Computing & Analysis Model Roger Jones Lancaster University ATLAS UK 06 IPPP, 20/9/2006.
Markus Frank (CERN) & Albert Puig (UB).  An opportunity (Motivation)  Adopted approach  Implementation specifics  Status  Conclusions 2.
ALICE Computing Model A pictorial guide. ALICE Computing Model External T1 CERN T0 During pp run i (7 months): P2: data taking T0: first reconstruction.
L. Perini DATAGRID WP8 Use-cases 19 Dec ATLAS short term grid use-cases The “production” activities foreseen till mid-2001 and the tools to be used.
LHCb Current Understanding of Italian Tier-n Centres Domenico Galli, Umberto Marconi Roma, January 23, 2001.
16 September 2014 Ian Bird; SPC1. General ALICE and LHCb detector upgrades during LS2  Plans for changing computing strategies more advanced CMS and.
Hall D Computing Facilities Ian Bird 16 March 2001.
ATLAS – statements of interest (1) A degree of hierarchy between the different computing facilities, with distinct roles at each level –Event filter Online.
LHCb computing model and the planned exploitation of the GRID Eric van Herwijnen, Frank Harris Monday, 17 July 2000.
The LHC Computing Grid Visit of Mtro. Enrique Agüera Ibañez
Philippe Charpentier CERN – LHCb On behalf of the LHCb Computing Group
Dagmar Adamova (NPI AS CR Prague/Rez) and Maarten Litmaath (CERN)
New strategies of the LHC experiments to meet
US ATLAS Physics & Computing
Gridifying the LHCb Monte Carlo production system
The ATLAS Computing Model
LHCb thinking on Regional Centres and Related activities (GRIDs)
Short to middle term GRID deployment plan for LHCb
Development of LHCb Computing Model F Harris
The LHC Computing Grid Visit of Professor Andreas Demetriou
Presentation transcript:

LHCb Computing Model and updated requirements John Harvey

LHCb Computing model and updated requirements Slide 2 LHC Computing Review qThree technical review panels ãWorld wide analysis ãSoftware ãManagement and resources qFirst meeting of each panel took place end March qLHCb presentations available on web ã qDraft LHCb Technical Notes being prepared (see web) ãBaseline Computing model ãAnswers to SPP questions qInvite feedback qNext meetings April 14 th (Panel 3), May 12 th (Panels 1,3)

LHCb Computing model and updated requirements Slide 3 Panel 1 : World-wide analysis qGive a description of baseline computing model, indicating how it differs from MONARC generic model. qWhat is the current tentative distribution of the different regional centres? qGive a tentative assessment for data storage and management and for processing power required at the different centres (now and at the different stages towards LHC start-up). qExplain your current understanding of your policy regarding magnetic tapes versus disks. qGiven the current status of the implementation of the above model, is the experiment involved/interested in the current GRID test activities? qDo you plan Mock Data Challenges or equivalent test-beds to validate your model and provide intermediate services between now and LHC start-up? qWhat are the resources available/planned for the implementation of the above model (again staggered between now and LHC start-up)?

LHCb Computing model and updated requirements Slide 4 Panel 3 : Management and Resources qCompare computing needs in terms of hardware, software, organization, and manpower, to produce a plan to be presented to the funding agencies. qProduce an "Interim version of the Memorandum of Understanding “ indicating responsibilities of institutions ãInclude responsibilities for writing and maintenance of the software as has been done for construction of the detector. qFirst meeting focused on first 6 months of data-taking: ãPrepare a description of the computing systems needed to calibrate the subdetectors and the related strategy to get from raw data to first "simple" physics analysis. q"world wide physics analysis" left to a following meeting.

LHCb Computing model and updated requirements Slide 5 RAW Data DAQ system L2/L3 Trigger Calibration Data Reconstruction Event Summary Data (ESD) Reconstruction Tags Detector RAW Tags L3YES, sample L2/L3NO ESD Reconstruction TagsAnalysis Object Data (AOD)Physics Tags First Pass Analysis Physics Analysis Private Data Analysis Workstation Physics results ESD RAW Dataflow Model

LHCb Computing model and updated requirements Slide 6 Real Data Storage Requirements

LHCb Computing model and updated requirements Slide 7 Real Data CPU Requirements q1 SI95 = 40 MIPS qToday 1 PC ~ 10 SI95 qBy SI95/cpu (low cost), 250 SI95 (high end servers)

LHCb Computing model and updated requirements Slide 8 User Analysis Requirements qAssume that physicist performs a production analysis and requires a response time of 4 hours qThe ~10 7 events tagged by first pass analysis are scanned and candidates selected (0.25 SI95 /event) qThe selected candidates are subjected to analysis algorithm (20 SI95 / event) qTotal installed cpu power needed calculated assuming: ã~140 physicists actively doing analysis ãeach submits 1 job / day ( NB. many short jobs as well) ãanalysis distributed over a number of regional centres (~5) and assume ~20 physicists at each Regional Centre, ~40 at CERN ãAssume 0.3 x 10 7 events selected for algorithm on average ã10,000 SI95 at each Regional Centre, 20,000 SI95 at CERN

LHCb Computing model and updated requirements Slide 9 Simulation Requirements - Signal Events  CPU power to simulate 10 7 B -> D*  events in 1 year ãassume need to simulate 10 times real data sample (10 6 ) ãinstalled capacity needed is 100,000 SI95

LHCb Computing model and updated requirements Slide 10 Simulation Requirements - Background q10 5 bb inclusive events in detector every second q~100 Hz are recorded in real data ãtrigger efficiency qIf need as many to be simulated then need to generate, track, digitise and trigger bb inclusive events/yr and 10 9 will have to be reconstructed ãcorresponds to SI95 sec/yr i.e.10,000,000 SI95 qObviously need to study ways of optimising background simulation ãstore and reuse data produced at generator level (storage!) ãoptimise generation step without biasing physics ãfocus on background particularly dangerous for a specific physics channel ãaim to reduce requirements by > 1 order of magnitude qAssume 400,000 SI95 required

LHCb Computing model and updated requirements Slide 11 Simulation Requirements - Summary qFor comparison the experiments quote ãALICE (2,000,000), ATLAS/CMS (1,500,000)

CPU for production Mass Storage for RAW, ESD AOD, and TAG Institute Selected User Analyses Institute Selected User Analyses Regional Centre User analysis Production Centre Generate raw data Reconstruction Production analysis User analysis Regional Centre User analysis Regional Centre User analysis Institute Selected User Analyses Regional Centre User analysis Institute Selected User Analyses CPU for analysis Mass storage for AOD, TAG CPU and data servers AOD,TAG real : 80TB/yr sim: 120TB/yr AOD,TAG 8-12 TB/yr

Production Centre (x1) Regional Centre (~x5) Institute (~x50) Real DataSimulated Data Data collection Triggering Reconstruction Final State Reconstruction CERN Data shipped to each RC: AOD and TAG datasets 20TB x 4 times/yr= 80TB/yr User Analysis Data shipped to each Institute: AOD and TAG for samples 1TB x 10 times/yr= 10TB/yr RAL, Lyon,... Event Generation GEANT tracking Reconstruction Final State Reconstruction Data shipped to each RC: AOD, Generator and TAG datasets 30TB x 4 times/yr= 120TB/yr User Analysis Selected User Analysis Data shipped to each institute: AOD and TAG for samples 3TB x 10 times/yr= 30TB/yr

LHCb Computing model and updated requirements Slide 14 Computing at CERN qRun high level triggers and reconstruction on same cpu farm located at LHCb pit qSend RAW and ESD data over the CDR link (80 Gbps) to computer centre for archive, and first pass analysis qMaintain local storage at pit in case CDR down ãaccumulate 2 TB/day, therefore need >10 TB qDispatch AOD and TAG etc to regional centres qDuring shutdowns, down periods do re-processing of RAW data on the farm in the pit ãread RAW back from computer centre ãsend new ESD from pit to computer centre ãfull farm available so proceeds at twice the rate ãallows 2-3 reprocessings of complete year’s data qFlexible, efficient and maintainable solution

LHCb Computing model and updated requirements Slide 15 Compute Facilities at CERN DAQ / Calibration L2/L3 trigger processing Reconstruction Re-reconstruction (2-3 times per year) CDR 80 Gb/s Experiment - LHC Pit 8 CERN Computer Centre Temporary store (10 TB) for Raw Data and ESD Calibration (5TB) Data Storage and Analysis Physics Analysis MSS Readout Network Data Taking 200 Hz Raw data 20 MB/s ESD data 20 MB/s SHUTDOWN 600 Hz Raw data 60 MB/s ESD Data 60 MB/s disk RAW ESD AOD TAG Data and CPU server CPU farm AOD TAG to Regional Centres

LHCb Computing model and updated requirements Slide 16 Facility at Pit - Requirements

LHCb Computing model and updated requirements Slide 17 CERN Computer Centre Requirements

LHCb Computing model and updated requirements Slide 18 Simulation requirements q ã10 7 simulated events/yr for detector optimisation studies ãprepare TDRs q ã events/yr for high level trigger studies q ãstart to install and commission large scale facilities ãstart to produce large samples of background events with the final detector description ã~10 8 simulated events/yr q>2001 ãuse simulation and MDC to test computing model ãcontribute to HEP Application WP of EU grid proposal

LHCb Computing model and updated requirements Slide 19 Sizing Estimate for Regional Centre

LHCb Computing model and updated requirements Slide 20 Assignment of responsibility qUnderstood in LHCb institutes building subdetectors also take responsibility for development and maintenance of software qThe detector TDRs are in preparation now qMOU after TDRs

LHCb Computing model and updated requirements Slide 21 Discussion qHow will our computing needs evolve between now and 2005? qWhat regional centres will LHCb use for satisfying these needs? (RAL, CCIN2P3/Lyon, ++…) qWhat resources (cpu, storage) will be available for satisfying our simulation needs? qWhat is our attitude towards making an MOU for computing? Including software? What timescale? qWhat level of engagement should we take in Grid Projects?

LHCb Computing model and updated requirements Slide 22 GRID LHCb WP Physics Study  The total sample of B > J  /K s simulated events needed is ~10 times the number produced in the real data. qIn one year of datataking we expect to collect and fully reconstruct 10 5 events, therefore need 10 6 simulated events. qThe number of events that have to be generated, stored and reconstructed to produce this sample is q10% of the ESD data copied for systematic studies (~100 GB). qThe total amount of data generated in this production would be : RAW data 200 kB/event x 10 7 = 2.0 TB Generator data 12 kB/event x 10 7 = 0.12 TB ESD data 100 kB/event x 10 7 = 1.0 TB AOD data 20 kB/event x 10 7 = 0. 2 TB TAG data 1 kB/event x 10 7 = 0.01 TB

LHCb Computing model and updated requirements Slide 23 Grid LHCb WP - Grid Testbed qMAP farm at Liverpool has 300 processors, would take 4 months to generate the full sample of events qAll data generated (~3TB) would be transferred to RAL for archive (UK regional facility). qAll AOD and TAG datasets dispatched from RAL to other regional centres, such as Lyon and CERN. qPhysicists run jobs at the regional centre or ship AOD and TAG data to local institute and run jobs there. Also copy ESD for a fraction (~10%) of events for systematic studies (~100 GB). qThe resulting data volumes to be shipped between facilities over 4 months would be as follows : Liverpool to RAL 3 TB (RAW ESD AOD and TAG) RAL to LYON/CERN/… 0.3 TB (AOD and TAG) LYON to LHCb institute 0.3 TB (AOD and TAG) RAL to LHCb institute 100 GB (ESD for systematic studies)