Status of ILC Computing Model and Cost Study Akiya Miyamoto Norman Graf Frank Gaede Andre Sailer Marcel Stanitzki Jan Strube 1.

Slides:



Advertisements
Similar presentations
31/03/00 CMS(UK)Glenn Patrick What is the CMS(UK) Data Model? Assume that CMS software is available at every UK institute connected by some infrastructure.
Advertisements

Resources for the ATLAS Offline Computing Basis for the Estimates ATLAS Distributed Computing Model Cost Estimates Present Status Sharing of Resources.
Amber Boehnlein, FNAL D0 Computing Model and Plans Amber Boehnlein D0 Financial Committee November 18, 2002.
T1 at LBL/NERSC/OAK RIDGE General principles. RAW data flow T0 disk buffer DAQ & HLT CERN Tape AliEn FC Raw data Condition & Calibration & data DB disk.
23/04/2008VLVnT08, Toulon, FR, April 2008, M. Stavrianakou, NESTOR-NOA 1 First thoughts for KM3Net on-shore data storage and distribution Facilities VLV.
1 Data Storage MICE DAQ Workshop 10 th February 2006 Malcolm Ellis & Paul Kyberd.
DATA PRESERVATION IN ALICE FEDERICO CARMINATI. MOTIVATION ALICE is a 150 M CHF investment by a large scientific community The ALICE data is unique and.
October 24, 2000Milestones, Funding of USCMS S&C Matthias Kasemann1 US CMS Software and Computing Milestones and Funding Profiles Matthias Kasemann Fermilab.
Computing plans for the post DBD phase Akiya Miyamoto KEK ILD Session ECFA May 2013.
Ian Fisk and Maria Girone Improvements in the CMS Computing System from Run2 CHEP 2015 Ian Fisk and Maria Girone For CMS Collaboration.
+ discussion in Software WG: Monte Carlo production on the Grid + discussion in TDAQ WG: Dedicated server for online services + experts meeting (Thusday.
Computing Infrastructure Status. LHCb Computing Status LHCb LHCC mini-review, February The LHCb Computing Model: a reminder m Simulation is using.
1 Kittikul Kovitanggoon*, Burin Asavapibhop, Narumon Suwonjandee, Gurpreet Singh Chulalongkorn University, Thailand July 23, 2015 Workshop on e-Science.
Copyright © 2000 OPNET Technologies, Inc. Title – 1 Distributed Trigger System for the LHC experiments Krzysztof Korcyl ATLAS experiment laboratory H.
Fermilab User Facility US-CMS User Facility and Regional Center at Fermilab Matthias Kasemann FNAL.
LHC Computing Review - Resources ATLAS Resource Issues John Huth Harvard University.
Your university or experiment logo here Caitriana Nicholson University of Glasgow Dynamic Data Replication in LCG 2008.
Finnish DataGrid meeting, CSC, Otaniemi, V. Karimäki (HIP) DataGrid meeting, CSC V. Karimäki (HIP) V. Karimäki (HIP) Otaniemi, 28 August, 2000.
ALICE Upgrade for Run3: Computing HL-LHC Trigger, Online and Offline Computing Working Group Topical Workshop Sep 5 th 2014.
14 Aug 08DOE Review John Huth ATLAS Computing at Harvard John Huth.
Computing for ILC experiments Akiya Miyamoto KEK 14 May 2014 AWLC14 Any comments are welcomed.
Computing Resources for ILD Akiya Miyamoto, KEK with a help by Vincent, Mark, Junping, Frank 9 September 2014 ILD Oshu City a report on work.
EGEE is a project funded by the European Union under contract IST HEP Use Cases for Grid Computing J. A. Templon Undecided (NIKHEF) Grid Tutorial,
Tim 18/09/2015 2Tim Bell - Australian Bureau of Meteorology Visit.
Caitriana Nicholson, CHEP 2006, Mumbai Caitriana Nicholson University of Glasgow Grid Data Management: Simulations of LCG 2008.
Les Les Robertson LCG Project Leader High Energy Physics using a worldwide computing grid Torino December 2005.
ESFRI & e-Infrastructure Collaborations, EGEE’09 Krzysztof Wrona September 21 st, 2009 European XFEL.
CERN IT Department CH-1211 Genève 23 Switzerland t Frédéric Hemmer IT Department Head - CERN 23 rd August 2010 Status of LHC Computing from.
CERN – IT Department CH-1211 Genève 23 Switzerland t Working with Large Data Sets Tim Smith CERN/IT Open Access and Research Data Session.
Software Common Task Group Meeting Akiya Miyamoto 6-June-2011.
The KLOE computing environment Nuclear Science Symposium Portland, Oregon, USA 20 October 2003 M. Moulson – INFN/Frascati for the KLOE Collaboration.
Software Overview Akiya Miyamoto KEK JSPS Tokusui Workshop December-2012 Topics MC production Computing reousces GRID Future events Topics MC production.
Predrag Buncic Future IT challenges for ALICE Technical Workshop November 6, 2015.
LHCbComputing LHCC status report. Operations June 2014 to September m Running jobs by activity o Montecarlo simulation continues as main activity.
Computing for LHC Physics 7th March 2014 International Women's Day - CERN- GOOGLE Networking Event Maria Alandes Pradillo CERN IT Department.
LHCbComputing Resources requests : changes since LHCb-PUB (March 2013) m Assume no further reprocessing of Run I data o (In.
LHC Computing, CERN, & Federated Identities
Computing Resources for ILD Akiya Miyamoto, KEK with a help by Vincent, Mark, Junping, Frank 9 September 2014 ILD Oshu City a report on work.
The ATLAS Computing Model and USATLAS Tier-2/Tier-3 Meeting Shawn McKee University of Michigan Joint Techs, FNAL July 16 th, 2007.
Computing Issues for the ATLAS SWT2. What is SWT2? SWT2 is the U.S. ATLAS Southwestern Tier 2 Consortium UTA is lead institution, along with University.
Ian Bird WLCG Networking workshop CERN, 10 th February February 2014
LHCbComputing Computing for the LHCb Upgrade. 2 LHCb Upgrade: goal and timescale m LHCb upgrade will be operational after LS2 (~2020) m Increase significantly.
Distributed Physics Analysis Past, Present, and Future Kaushik De University of Texas at Arlington (ATLAS & D0 Collaborations) ICHEP’06, Moscow July 29,
Markus Frank (CERN) & Albert Puig (UB).  An opportunity (Motivation)  Adopted approach  Implementation specifics  Status  Conclusions 2.
Meeting with University of Malta| CERN, May 18, 2015 | Predrag Buncic ALICE Computing in Run 2+ P. Buncic 1.
Perspectives on LHC Computing José M. Hernández (CIEMAT, Madrid) On behalf of the Spanish LHC Computing community Jornadas CPAN 2013, Santiago de Compostela.
LHCb Computing activities Philippe Charpentier CERN – LHCb On behalf of the LHCb Computing Group.
Belle II Computing Fabrizio Bianchi INFN and University of Torino Meeting Belle2 Italia 17/12/2014.
Jianming Qian, UM/DØ Software & Computing Where we are now Where we want to go Overview Director’s Review, June 5, 2002.
A proposal for the KM3NeT Computing Model Pasquale Migliozzi INFN - Napoli 1.
Grid technologies for large-scale projects N. S. Astakhov, A. S. Baginyan, S. D. Belov, A. G. Dolbilov, A. O. Golunov, I. N. Gorbunov, N. I. Gromova, I.
Computing infrastructures for the LHC: current status and challenges of the High Luminosity LHC future Worldwide LHC Computing Grid (WLCG): Distributed.
ATLAS Computing: Experience from first data processing and analysis Workshop TYL’10.
Hall D Computing Facilities Ian Bird 16 March 2001.
Computing requirements for the experiments Akiya Miyamoto, KEK 31 August 2015 Mini-Workshop on ILC Infrastructure and CFS for Physics and KEK.
ATLAS – statements of interest (1) A degree of hierarchy between the different computing facilities, with distinct roles at each level –Event filter Online.
Ian Bird WLCG Workshop San Francisco, 8th October 2016
Overview of the Belle II computing
The LHC Computing Grid Visit of Mtro. Enrique Agüera Ibañez
evoluzione modello per Run3 LHC
LHC experiments Requirements and Concepts ALICE
Philippe Charpentier CERN – LHCb On behalf of the LHCb Computing Group
Dagmar Adamova (NPI AS CR Prague/Rez) and Maarten Litmaath (CERN)
ALICE Computing Model in Run3
ALICE Computing Upgrade Predrag Buncic
ILD Ichinoseki Meeting
New strategies of the LHC experiments to meet
Heavy Ion Physics Program of CMS Proposal for Offline Computing
The ATLAS Computing Model
Development of LHCb Computing Model F Harris
Presentation transcript:

Status of ILC Computing Model and Cost Study Akiya Miyamoto Norman Graf Frank Gaede Andre Sailer Marcel Stanitzki Jan Strube 1

Introduction ●Cost of computing needs for the detectors was not included in the TDR. ●For the LHC experiments, this was a very large item and (still is) a major source of headaches. o The Grid has been considered “The 5th experiment” at LHC. ●For ILC, I consider computing more like another sub- detector, not least because of the PFA paradigm. o But this needs to be proven with hard numbers 2

Studies in SiD / ILD ILD (Miyamoto) presented first preliminary results of detailed studies at AWLC14 and extended studies over the summer. SiD only had DAQ numbers at 1 TeV. For LCWS14, SiD used ILD extrapolation of background numbers to obtain data rates at lower ILC stages. 3

SiD Data Rate 1060 → 5.3 GB/s Assuming Hz 4

Data Rates From the Detectors: ~1.1 GB raw 500 GB, both for SiD/ILD ●Assuming half of a s year: o ~9 PB raw data / year / detector Slightly less at lower energies: 350 GeV SiD: ~4.9 PB, ILD ~6.3 PB 250 GeV SiD: ~3 PB, ILD ~5.5 PB Large uncertainties from estimation of pair bg. ILD: Detailed simulation study; SiD: extrapolation from background rates and number of BX/train 5

Data Management Where is LHC in Big Data Terms? Business s sent 3000PB/year (Doesn’t count; not managed as a coherent data set) Google search 100PB Facebook uploads 180PB/year Kaiser Permanente 30PB LHC data 15PB/yr YouTube 15PB/yr US Census Lib of Congress Climate DB Nasdaq Wired 4/2013 In 2012: 2800 exabytes created or replicated 1 Exabyte = 1000 PB Current ATLAS data set, all data products: 140 PB Big Data in 2012 ~14x growth expected From: Torre Wenaus, CHEP 2013 SiD 9 PB/yr Computing needs and networking capabilities of society as well as science will increase in the future.

Caveats Data rate looks scary at first sight: > LHC! But: ●This will be in ~2030. HL-LHC will have set a new bar by then. ●There will be many experiments with data rates in the same ball park by then. ●As mentioned, large safety factors on background. Using 5, while 2 is advertized by the machine. ●Data reduction techniques being investigated. (Size is mostly from pair bg) 7

Given Constraints ●Want to avoid driving large investment in networking. o Take advantage of “natural” developments. ●Want to avoid a large computing center at the IP. o Size is constrained, infrastructure limited. ●Want to avoid a large investment in a Tier-0 o Tier-0 at LHC: 45%-90% of all 11 Tier-1 combined o ILC campus computing should be driven by user community on site, not detector data rates. o But we want two copies of the raw data, one on site. 8

Japanese Connections From T. Hara, ALICE WS, Mar 5,

Japanese Connections From T. Hara, ALICE WS, Mar 5,

Strategy 1.Move the data out of the mountain as quickly as possible o Need few days buffer for network outage 2.Store a raw copy on campus 3.Process (filter) events for analysis and ship to grid. o CPU needs for this being investigated. Filtering could happen on the grid. Under study. o A complete copy of the raw data needs to be shipped for redundancy anyway. o Plan to use existing “Tier-1” centers. o Probably want certain “streams” with fast feedback. 11

Event Building and Prompt Reconstruction (PR) 1.Write data 2.sub-detector based preliminary reconstruction 3.identify bunches of interest Fast Physics Data (FPD) Raw Data (RD) Online Processed Data (OPD) JOB-A Re-processing with better constants Offline Reconstructed Data (ORD) DST MC data JOB-C MC-Production JOB-B Produce condensed data sample Calibration Data (CD) 1.Record Train data 2.Send data to Main Campus and Online monitoring 1.Data sample and reconstruction for monitoring 2.Temporary data storage for emergency Temp. Storage ~1GB/s Raw Data Copy ILC Detector Full Offline Reconstruction 1.calibration and alignment 2.background hit rejection 3.full event reconstruction 4.event classification Main Campus Interaction Region Main Capmus / GRID based Offline computing Online Offline GRID PR Storage 12

CPU estimation Currently only have ILD estimation. SiD is unlikely to provide comparable level of detail. Based on detailed studies of computing time for simulation and reconstruction of different physics channels at different energies (Miyamoto): Campus computing (event filter): ~26k HepSPEC06 ILC total: 415k HepSPEC06 This is combined for both experiments. Assumes 2x ILD computing. 13

Comparison with other experiments ATLAS Tier-0+Tier-1 CPU 2015 (667 kHS) (total: 1175 kHS) ATLAS 2015 Disk (total): 108 PB Tape: 98 PB Scaling law (from LHC experiments experience): constant investment buys 20% more CPU, 15% more storage every year → ~13 years to gain order of magnitude with same investment in CPU → ~16.5 years to gain order of magnitude in storage Belle investment in 2016: Tape: PB Disk: PB CPU: kHS kHS = kHepSPEC Value in 2029: Tape: 101 PB Disk: 123 PB CPU: 2924 kHS Example: 14

For comparison: ILC 500 GeV -9 PB / yr / experiment raw data; 220 PB for 5 yr, ILC globally, incl. MC, analysis -415 kHS06 total ILC (SiD+ILD) ILC globally GeV 15

Manpower estimates TaskDetailsFTE Management Operation in general, negotiation among sites, maintain rules for computing, deliver information 6 Network Maintenance, monitor, trouble handling develop faster network 6 Certificate Maintenance, person ID, issue certificate, trouble handling 4 Security Maintenance, Monitor, incident handling, prevention 4 Storage Maintenance, trouble handling, improvements 4 CPU servers Maintenance, trouble handling, improvements 4 Software maintenance Maintenance, trouble handling, improvements 6 Core Software development Design, implementation, maintenance, trouble handling, improvements 6 User contacts User portal, monitor, dispatch problems, trouble handling 12 Misc. server management Operation, maintenance, trouble handling of various servers such as EDMS, Indico, Web, cloud, mail, etc 4 Videoconferencing Support Maintenance & User support 4 Total60 16

Summary ●Input to MEXT process is being prepared ●Large uncertainties on data rates and CPU o data rates conservative, CPU unknown ●Data rates from detectors would compete with LHC today o > 10 years from now, our data rates will not require large additional investments ●Expected continued development of networks lets us take advantage of a distributed infrastructure o Allows granular contributions from FAs Some investment needed to study data distribution strategies. Reconstruction and Analysis Software needs large investments 17