The ATLAS Computing Model

Slides:



Advertisements
Similar presentations
The ATLAS Computing Model Roger Jones Lancaster University CHEP06 Mumbai 13 Feb
Advertisements

31/03/00 CMS(UK)Glenn Patrick What is the CMS(UK) Data Model? Assume that CMS software is available at every UK institute connected by some infrastructure.
Bernd Panzer-Steindel, CERN/IT WAN RAW/ESD Data Distribution for LHC.
SLUO LHC Workshop, SLACJuly 16-17, Analysis Model, Resources, and Commissioning J. Cochran, ISU Caveat: for the purpose of estimating the needed.
Amber Boehnlein, FNAL D0 Computing Model and Plans Amber Boehnlein D0 Financial Committee November 18, 2002.
T1 at LBL/NERSC/OAK RIDGE General principles. RAW data flow T0 disk buffer DAQ & HLT CERN Tape AliEn FC Raw data Condition & Calibration & data DB disk.
23/04/2008VLVnT08, Toulon, FR, April 2008, M. Stavrianakou, NESTOR-NOA 1 First thoughts for KM3Net on-shore data storage and distribution Facilities VLV.
1 Data Storage MICE DAQ Workshop 10 th February 2006 Malcolm Ellis & Paul Kyberd.
CHEP – Mumbai, February 2006 The LCG Service Challenges Focus on SC3 Re-run; Outlook for 2006 Jamie Shiers, LCG Service Manager.
Data Import Data Export Mass Storage & Disk Servers Database Servers Tapes Network from CERN Network from Tier 2 and simulation centers Physics Software.
LHC Computing Review - Resources ATLAS Resource Issues John Huth Harvard University.
ATLAS and GridPP GridPP Collaboration Meeting, Edinburgh, 5 th November 2001 RWL Jones, Lancaster University.
14 Aug 08DOE Review John Huth ATLAS Computing at Harvard John Huth.
9 February 2000CHEP2000 Paper 3681 CDF Data Handling: Resource Management and Tests E.Buckley-Geer, S.Lammel, F.Ratnikov, T.Watts Hardware and Resources.
ATLAS Computing Model – US Research Program Manpower J. Shank N.A. ATLAS Physics Workshop Tucson, AZ 21 Dec., 2004.
Meeting, 5/12/06 CMS T1/T2 Estimates à CMS perspective: n Part of a wider process of resource estimation n Top-down Computing.
ATLAS: Heavier than Heaven? Roger Jones Lancaster University GridPP19 Ambleside 28 August 2007.
The ATLAS Grid Progress Roger Jones Lancaster University GridPP CM QMUL, 28 June 2006.
ATLAS WAN Requirements at BNL Slides Extracted From Presentation Given By Bruce G. Gibbard 13 December 2004.
Integration of the ATLAS Tag Database with Data Management and Analysis Components Caitriana Nicholson University of Glasgow 3 rd September 2007 CHEP,
LHCbComputing LHCC status report. Operations June 2014 to September m Running jobs by activity o Montecarlo simulation continues as main activity.
NA62 computing resources update 1 Paolo Valente – INFN Roma Liverpool, Aug. 2013NA62 collaboration meeting.
The ATLAS Computing Model and USATLAS Tier-2/Tier-3 Meeting Shawn McKee University of Michigan Joint Techs, FNAL July 16 th, 2007.
The ATLAS Computing Model Roger Jones Lancaster University CHEP07 Victoria B.C. Sept
Predrag Buncic ALICE Status Report LHCC Referee Meeting CERN
The ATLAS Computing & Analysis Model Roger Jones Lancaster University GDB BNL, Long Island, 6/9/2006.
ATLAS Distributed Computing perspectives for Run-2 Simone Campana CERN-IT/SDC on behalf of ADC.
Victoria, Sept WLCG Collaboration Workshop1 ATLAS Dress Rehersals Kors Bos NIKHEF, Amsterdam.
Ian Bird Overview Board; CERN, 8 th March 2013 March 6, 2013
Distributed Physics Analysis Past, Present, and Future Kaushik De University of Texas at Arlington (ATLAS & D0 Collaborations) ICHEP’06, Moscow July 29,
The ATLAS Computing & Analysis Model Roger Jones Lancaster University ATLAS UK 06 IPPP, 20/9/2006.
GDB, 07/06/06 CMS Centre Roles à CMS data hierarchy: n RAW (1.5/2MB) -> RECO (0.2/0.4MB) -> AOD (50kB)-> TAG à Tier-0 role: n First-pass.
ATLAS Computing Model & Service Challenges Roger Jones 12 th October 2004 CERN.
Finding Data in ATLAS. May 22, 2009Jack Cranshaw (ANL)2 Starting Point Questions What is the latest reprocessing of cosmics? Are there are any AOD produced.
LHCb Current Understanding of Italian Tier-n Centres Domenico Galli, Umberto Marconi Roma, January 23, 2001.
Computing Model José M. Hernández CIEMAT, Madrid On behalf of the CMS Collaboration XV International Conference on Computing in High Energy and Nuclear.
Are We Ready for the LHC Data? Roger Jones Lancaster University PPD Xmas Bash RAL, 19/12/06.
Meeting with University of Malta| CERN, May 18, 2015 | Predrag Buncic ALICE Computing in Run 2+ P. Buncic 1.
LHCb Computing activities Philippe Charpentier CERN – LHCb On behalf of the LHCb Computing Group.
1 June 11/Ian Fisk CMS Model and the Network Ian Fisk.
WLCG November Plan for shutdown and 2009 data-taking Kors Bos.
Jianming Qian, UM/DØ Software & Computing Where we are now Where we want to go Overview Director’s Review, June 5, 2002.
PD2P, Caching etc. Kaushik De Univ. of Texas at Arlington ADC Retreat, Naples Feb 4, 2011.
ATLAS Computing Model Ghita Rahal CC-IN2P3 Tutorial Atlas CC, Lyon
LHCb LHCb GRID SOLUTION TM Recent and planned changes to the LHCb computing model Marco Cattaneo, Philippe Charpentier, Peter Clarke, Stefan Roiser.
CERN IT Department CH-1211 Genève 23 Switzerland t EGEE09 Barcelona ATLAS Distributed Data Management Fernando H. Barreiro Megino on behalf.
LHCb Computing 2015 Q3 Report Stefan Roiser LHCC Referees Meeting 1 December 2015.
ATLAS – statements of interest (1) A degree of hierarchy between the different computing facilities, with distinct roles at each level –Event filter Online.
Computing Operations Roadmap
Ian Bird WLCG Workshop San Francisco, 8th October 2016
Database Replication and Monitoring
U.S. ATLAS Tier 2 Computing Center
Data Challenge with the Grid in ATLAS
ATLAS activities in the IT cloud in April 2008
Vanderbilt Tier 2 Project
Bernd Panzer-Steindel, CERN/IT
Status and Prospects of The LHC Experiments Computing
LHCb Computing Model and Data Handling Angelo Carbone 5° workshop italiano sulla fisica p-p ad LHC 31st January 2008.
Philippe Charpentier CERN – LHCb On behalf of the LHCb Computing Group
ALICE Computing Model in Run3
An introduction to the ATLAS Computing Model Alessandro De Salvo
ALICE Computing Upgrade Predrag Buncic
New strategies of the LHC experiments to meet
R. Graciani for LHCb Mumbay, Feb 2006
Heavy Ion Physics Program of CMS Proposal for Offline Computing
ATLAS DC2 & Continuous production
LHCb thinking on Regional Centres and Related activities (GRIDs)
The LHCb Computing Data Challenge DC06
Presentation transcript:

The ATLAS Computing Model BTeV Talk 5/3/2019 The ATLAS Computing Model Roger Jones Lancaster University CHEP06 Mumbai 13 Feb. 2006 For January Collab Meeting

Brief summary ATLAS Facilities and their roles BTeV Talk 5/3/2019 Overview Brief summary ATLAS Facilities and their roles Growth of resources CPU, Disk, Mass Storage Network requirements CERN↔ Tier 1 ↔ Tier 2 Operational Issues and Hot Topics RWL Jones 13 Feb. 2006 Mumbai For January Collab Meeting

BTeV Talk 5/3/2019 Computing Resources Computing Model fairly well evolved, documented in C-TDR Externally reviewed http://doc.cern.ch//archive/electronic/cern/preprints/lhcc/public/lhcc-2005-022.pdf There are (and will remain for some time) many unknowns Calibration and alignment strategy is still evolving Physics data access patterns MAY be exercised from June Unlikely to know the real patterns until 2007/2008! Still uncertainties on the event sizes , reconstruction time Lesson from the previous round of experiments at CERN (LEP, 1989-2000) Reviews in 1988 underestimated the computing requirements by an order of magnitude! RWL Jones 13 Feb. 2006 Mumbai For January Collab Meeting

ATLAS Facilities Event Filter Farm at CERN Tier 0 Center at CERN BTeV Talk 5/3/2019 ATLAS Facilities Event Filter Farm at CERN Located near the Experiment, assembles data into a stream to the Tier 0 Center Tier 0 Center at CERN Raw data  Mass storage at CERN and to Tier 1 centers Swift production of Event Summary Data (ESD) and Analysis Object Data (AOD) Ship ESD, AOD to Tier 1 centers  Mass storage at CERN Tier 1 Centers distributed worldwide (10 centers) Re-reconstruction of raw data, producing new ESD, AOD Scheduled, group access to full ESD and AOD Tier 2 Centers distributed worldwide (approximately 30 centers) Monte Carlo Simulation, producing ESD, AOD, ESD, AOD  Tier 1 centers On demand user physics analysis CERN Analysis Facility Analysis Heightened access to ESD and RAW/calibration data on demand Tier 3 Centers distributed worldwide Physics analysis RWL Jones 13 Feb. 2006 Mumbai For January Collab Meeting

Processing Tier-0: Tier-1: BTeV Talk 5/3/2019 Processing Tier-0: Prompt first pass processing on express/calibration physics stream 24-48 hours later, process full physics data stream with reasonable calibrations Implies large data movement from T0 →T1s Tier-1: Reprocess 1-2 months after arrival with better calibrations Reprocess all resident RAW at year end with improved calibration and software Implies large data movement from T1↔T1 and T1 → T2 RWL Jones 13 Feb. 2006 Mumbai For January Collab Meeting

Analysis model Analysis model broken into two components BTeV Talk 5/3/2019 Analysis model Analysis model broken into two components Scheduled central production of augmented AOD, tuples & TAG collections from ESD Derived files moved to other T1s and to T2s Chaotic user analysis of augmented AOD streams, tuples, new selections etc and individual user simulation and CPU-bound tasks matching the official MC production Modest job traffic between T2s RWL Jones 13 Feb. 2006 Mumbai For January Collab Meeting

Inputs to the ATLAS Computing Model (1) BTeV Talk 5/3/2019 Inputs to the ATLAS Computing Model (1) RWL Jones 13 Feb. 2006 Mumbai For January Collab Meeting

Inputs to the ATLAS Computing Model (2) BTeV Talk 5/3/2019 Inputs to the ATLAS Computing Model (2) RWL Jones 13 Feb. 2006 Mumbai For January Collab Meeting

Data Flow Tier 0 view Tier 2 view EF farm  T0 BTeV Talk 5/3/2019 Data Flow EF farm  T0 320 MB/s continuous T0 Raw data  Mass Storage at CERN T0 Raw data  Tier 1 centers T0 ESD, AOD, TAG  Tier 1 centers 2 copies of ESD distributed worldwide T1  T2 Some RAW/ESD, All AOD, All TAG Some group derived datasets T2  T1 Simulated RAW, ESD, AOD, TAG T0  T2 Calibration processing? Tier 0 view Tier 2 view RWL Jones 13 Feb. 2006 Mumbai For January Collab Meeting

ATLAS partial &“average” T1 Data Flow (2008) RAW ESD (2x) AODm (10x) 1 Hz 85K f/day 720 MB/s Tape RAW ESD2 AODm2 0.044 Hz 3.74K f/day 44 MB/s 3.66 TB/day RAW 1.6 GB/file 0.02 Hz 1.7K f/day 32 MB/s 2.7 TB/day Tier-0 disk buffer AODm1 500 MB/file 0.04 Hz 3.4K f/day 20 MB/s 1.6 TB/day AODm2 500 MB/file 0.04 Hz 3.4K f/day 20 MB/s 1.6 TB/day ESD1 0.5 GB/file 0.02 Hz 1.7K f/day 10 MB/s 0.8 TB/day AODm1 500 MB/file 0.04 Hz 3.4K f/day 20 MB/s 1.6 TB/day RAW 1.6 GB/file 0.02 Hz 1.7K f/day 32 MB/s 2.7 TB/day AOD2 10 MB/file 0.2 Hz 17K f/day 2 MB/s 0.16 TB/day ESD2 0.5 GB/file 0.02 Hz 1.7K f/day 10 MB/s 0.8 TB/day AOD2 10 MB/file 0.2 Hz 17K f/day 2 MB/s 0.16 TB/day AODm2 500 MB/file 0.004 Hz 0.34K f/day 2 MB/s 0.16 TB/day T1 Each Tier-2 Plus simulation and analysis data flow CPU farm ESD2 0.5 GB/file 0.02 Hz 1.7K f/day 10 MB/s 0.8 TB/day AODm2 500 MB/file 0.036 Hz 3.1K f/day 18 MB/s 1.44 TB/day AODm2 500 MB/file 0.004 Hz 0.34K f/day 2 MB/s 0.16 TB/day ESD2 0.5 GB/file 0.02 Hz 1.7K f/day 10 MB/s 0.8 TB/day ESD2 0.5 GB/file 0.02 Hz 1.7K f/day 10 MB/s 0.8 TB/day AODm2 500 MB/file 0.036 Hz 3.1K f/day 18 MB/s 1.44 TB/day T1 Other Tier-1s disk storage T1 Other Tier-1s RWL Jones 13 Feb. 2006 Mumbai

Total ATLAS Requirements in for 2008 BTeV Talk 5/3/2019 Total ATLAS Requirements in for 2008 RWL Jones 13 Feb. 2006 Mumbai For January Collab Meeting

Important points: Discussion on disk vs tape storage at Tier-1’s BTeV Talk 5/3/2019 Important points: Discussion on disk vs tape storage at Tier-1’s Tape in this discussion means low-access secure storage No ‘disk buffers’ included except input to Tier 0 Storage of Simulation data from Tier 2’s Assumed to be at T1s Need partnerships to plan networking Must have fail-over to other sites Commissioning Requirement of flexibility in the early stages Simulation is a tunable parameter in T2 numbers! Heavy Ion running still under discussion. RWL Jones 13 Feb. 2006 Mumbai For January Collab Meeting

ATLAS T0 Resources RWL Jones 13 Feb. 2006 Mumbai BTeV Talk 5/3/2019 For January Collab Meeting

ATLAS T1 Resources RWL Jones 13 Feb. 2006 Mumbai BTeV Talk 5/3/2019 For January Collab Meeting

ATLAS T2 Resources RWL Jones 13 Feb. 2006 Mumbai BTeV Talk 5/3/2019 For January Collab Meeting

Required Network Bandwidth BTeV Talk 5/3/2019 Required Network Bandwidth Caveats No safety factors No headroom Just sustained average numbers Assumes no years/datasets are ‘junked’ Physics analysis pattern still under study… RWL Jones 13 Feb. 2006 Mumbai For January Collab Meeting

Mainly outward data movement BTeV Talk 5/3/2019 T1 ↔ CERN Bandwidth I+O The projected time profile of the nominal bandwidth required between CERN and the Tier-1 cloud. Mainly outward data movement RWL Jones 13 Feb. 2006 Mumbai For January Collab Meeting

About half is scheduled analysis BTeV Talk 5/3/2019 T1 ↔ T1 Bandwidth I+O About half is scheduled analysis The projected time profile of the nominal bandwidth required T1 and T1 cloud. RWL Jones 13 Feb. 2006 Mumbai For January Collab Meeting

T1↔ T2 Bandwidth I+O Dominated by AOD BTeV Talk 5/3/2019 T1↔ T2 Bandwidth I+O Dominated by AOD The projected time profile of the nominal aggregate bandwidth expected for an average ATLAS Tier- 1 and its three associated Tier-2s. RWL Jones 13 Feb. 2006 Mumbai For January Collab Meeting

Issues 1: T1 Reprocessing Reprocessing at Tier 1s is understood in principle In practice, requires efficient recall of data from archive and processing Pinning, pre-staging, DAGs all required? Requires the different storage roles to be well understood RWL Jones 13 Feb. 2006 Mumbai

Issues 2: Streaming This is *not* a theological issue All discussions are about optimisation of data access TDR has 4 streams from event filter primary physics, calibration, express, problem events Calibration stream has split at least once since! At AOD, envisage ~10 streams ESD streaming? Straw man streaming schemes (trigger based) being agreed Will explore the access improvements in large-scale exercises Will also look at overlaps, bookkeeping etc RWL Jones 13 Feb. 2006 Mumbai

TAG Access TAG is a keyed list of variables/event Two roles Direct access to event in file via pointer Data collection definition function Two formats, file and database Now believe large queries require full database Restricts it to Tier1s and large Tier2s/CAF Ordinary Tier2s hold file-based TAG corresponding to locally-held datasets RWL Jones 13 Feb. 2006 Mumbai

BTeV Talk 5/3/2019 Conclusions Computing Model Data Flow understood for placing Raw, ESD and AOD at Tiered centers Still need to understand data flow implications of Physics Analysis SC4/Computing System Commissioning in 2006 is vital. Some issues will only be resolved with real data in 2007-8 RWL Jones 13 Feb. 2006 Mumbai For January Collab Meeting

Backup Slides RWL Jones 13 Feb. 2006 Mumbai

Heavy Ion Running RWL Jones 13 Feb. 2006 Mumbai BTeV Talk 5/3/2019 For January Collab Meeting