Ian Bird Trigger, Online, Offline Computing Workshop CERN, 5 th September 2014.

Slides:



Advertisements
Similar presentations
Data & Storage Services CERN IT Department CH-1211 Genève 23 Switzerland t DSS Bit preservation cost outlook: Cost for 10, 20, 30 years archive.
Advertisements

SAN DIEGO SUPERCOMPUTER CENTER at the UNIVERSITY OF CALIFORNIA, SAN DIEGO Disk and Tape Storage Cost Models Richard Moore & David Minor San Diego Supercomputer.
SAN DIEGO SUPERCOMPUTER CENTER at the UNIVERSITY OF CALIFORNIA, SAN DIEGO Disk and Tape Storage Cost Models Richard Moore & David Minor San Diego Supercomputer.
12. March 2003Bernd Panzer-Steindel, CERN/IT1 LCG Fabric status
Evolution of Enterprise Services in the Statistics Canada IT Environment Silver Buckler Chief, Managed Storage Section Informatics Technology Services.
DNS Arrow Virtualisation and the business opportunity it presents. Steve Pearce Managing Director.
DATA PRESERVATION IN ALICE FEDERICO CARMINATI. MOTIVATION ALICE is a 150 M CHF investment by a large scientific community The ALICE data is unique and.
Designing Storage Architectures for Preservation Collections Library of Congress, September 17-18, 2007 Preservation and Access Repository Storage Architecture.
Ian M. Fisk Fermilab February 23, Global Schedule External Items ➨ gLite 3.0 is released for pre-production in mid-April ➨ gLite 3.0 is rolled onto.
Experiences and Challenges running CERN's High-Capacity Tape Archive 14/4/2015 CHEP 2015, Okinawa2 Germán Cancio, Vladimír Bahyl
Ian Bird WLCG Workshop Okinawa, 12 th April 2015.
GridPP Steve Lloyd, Chair of the GridPP Collaboration Board.
Ian Bird LHCC Referees’ meeting; CERN, 11 th June 2013 March 6, 2013
Ian Fisk and Maria Girone Improvements in the CMS Computing System from Run2 CHEP 2015 Ian Fisk and Maria Girone For CMS Collaboration.
Evolution of HEP Computing towards FCC
Computing for ILC experiment Computing Research Center, KEK Hiroyuki Matsunaga.
Meeting the Data Protection Demands of a 24x7 Economy Steve Morihiro VP, Programs & Technology Quantum Storage Solutions Group
9/16/2000Ian Bird/JLAB1 Planning for JLAB Computational Resources Ian Bird.
Data Preservation at the Exa-Scale and Beyond Challenges of the Next Decade(s) APARSEN Webinar, November 2014.
Ian Bird LHCC Referee meeting 23 rd September 2014.
Data & Storage Services CERN IT Department CH-1211 Genève 23 Switzerland t DSS From data management to storage services to the next challenges.
Fermilab User Facility US-CMS User Facility and Regional Center at Fermilab Matthias Kasemann FNAL.
HEPiX bit-preservation WG update – Spring 2014 Dmitry Ozerov/DESY Germán Cancio/CERN HEPiX Spring 2014, Annecy.
GStore: GSI Mass Storage ITEE-Palaver GSI Horst Göringer, Matthias Feyerabend, Sergei Sedykh
ALICE Upgrade for Run3: Computing HL-LHC Trigger, Online and Offline Computing Working Group Topical Workshop Sep 5 th 2014.
JLAB Computing Facilities Development Ian Bird Jefferson Lab 2 November 2001.
Ian Bird GDB; CERN, 8 th May 2013 March 6, 2013
Ian Bird LHC Computing Grid Project Leader LHC Grid Fest 3 rd October 2008 A worldwide collaboration.
CERN – IT Department CH-1211 Genève 23 Switzerland t Working with Large Data Sets Tim Smith CERN/IT Open Access and Research Data Session.
Slide David Britton, University of Glasgow IET, Oct 09 1 Prof. David Britton GridPP Project leader University of Glasgow UK-T0 Meeting 21 st Oct 2015 GridPP.
Future computing strategy Some considerations Ian Bird WLCG Overview Board CERN, 28 th September 2012.
LHC Computing, CERN, & Federated Identities
Ian Bird WLCG Workshop, Barcelona, 9 th July July 2014 Ian Bird; Barcelona1.
Advances in Bit Preservation (since DPHEP’2015) 3/2/2016 DPHEP / WLCG Workshop1 Germán Cancio IT Storage Group CERN DPHEP / WLCG Workshop Lisbon, 3/2/2016.
Ian Bird WLCG Networking workshop CERN, 10 th February February 2014
Ian Bird Overview Board; CERN, 8 th March 2013 March 6, 2013
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI EGI strategy and Grand Vision Ludek Matyska EGI Council Chair EGI InSPIRE.
Data & Storage Services CERN IT Department CH-1211 Genève 23 Switzerland t DSS Data architecture challenges for CERN and the High Energy.
01. December 2004Bernd Panzer-Steindel, CERN/IT1 Tape Storage Issues Bernd Panzer-Steindel LCG Fabric Area Manager CERN/IT.
LHCbComputing Computing for the LHCb Upgrade. 2 LHCb Upgrade: goal and timescale m LHCb upgrade will be operational after LS2 (~2020) m Increase significantly.
CERN IT Department CH-1211 Genève 23 Switzerland t The Tape Service at CERN Vladimír Bahyl IT-FIO-TSI June 2009.
GDB, 07/06/06 CMS Centre Roles à CMS data hierarchy: n RAW (1.5/2MB) -> RECO (0.2/0.4MB) -> AOD (50kB)-> TAG à Tier-0 role: n First-pass.
WLCG Status Report Ian Bird Austrian Tier 2 Workshop 22 nd June, 2010.
Tape archive challenges when approaching Exabyte-scale CHEP 2010, Taipei G. Cancio, V. Bahyl, G. Lo Re, S. Murray, E. Cano, G. Lee, V. Kotlyar CERN IT-DSS.
IT-DSS Alberto Pace2 ? Detecting particles (experiments) Accelerating particle beams Large-scale computing (Analysis) Discovery We are here The mission.
Computing infrastructures for the LHC: current status and challenges of the High Luminosity LHC future Worldwide LHC Computing Grid (WLCG): Distributed.
Hall D Computing Facilities Ian Bird 16 March 2001.
LHC collisions rate: Hz New PHYSICS rate: Hz Event selection: 1 in 10,000,000,000,000 Signal/Noise: Raw Data volumes produced.
Ian Bird, CERN WLCG Project Leader Amsterdam, 24 th January 2012.
CERN IT-Storage Strategy Outlook Alberto Pace, Luca Mascetti, Julien Leduc
Evolution of storage and data management
HEP LTDP Use Case & EOSC Pilot
Ian Bird WLCG Workshop San Francisco, 8th October 2016
APARSEN Webinar, November 2014
Scientific Computing Strategy (for HEP)
for the Offline and Computing groups
Experiences and Outlook Data Preservation and Long Term Analysis
Update on Plan for KISTI-GSDC
Luca dell’Agnello INFN-CNAF
Philippe Charpentier CERN – LHCb On behalf of the LHCb Computing Group
Dagmar Adamova (NPI AS CR Prague/Rez) and Maarten Litmaath (CERN)
Thoughts on Computing Upgrade Activities
R&D for HL-LHC from the CWP
Vision for CERN IT Department
Strategy document for LHCC – Run 4
New strategies of the LHC experiments to meet
Cloudstor: Glamming up the ecosystem
Using an Object Oriented Database to Store BaBar's Terabytes
LHCb thinking on Regional Centres and Related activities (GRIDs)
Presentation transcript:

Ian Bird Trigger, Online, Offline Computing Workshop CERN, 5 th September 2014

Scale of challenge … 5 September 2014 Ian Bird; HL-LHC computing workshop2 Data: ~25 PB/yr  400 PB/yr Compute: Growth > x50 10 Year Horizon What we think is affordable unless we do something differently

Following slides on tape from German Cancio (IT-DSS) 5 September 2014 Ian Bird; HL-LHC computing workshop3

Data & Storage Services Reliability and “bit” preservation Systematic verification of archive data ongoing –“Cold” archive: Users only accessed ~20% of the data (2013) –All “historic” data verified between –All new and repacked data being verified as well Data reliability significantly improved over last 5 years –From annual bit loss rates of O( ) (2009) to O( ) (2012) –New drive generations + less strain (HSM mounts, TM “hitchback”) + verification Still, room for improvement –Vendor quoted bit error rates: O( ) –But, these only refer to media failures –Errors (eg bit flips) appearing in complete chain > 17PB verified in 2014 Ian Bir d; HL- LH C co mp utin g wor ksh op

Data & Storage Services Physics tape: cost outlook Capacity/cost planning kept for ~4y time window (currently, up to LS2 start in 2018) –Strategy: Dual-sourced enterprise media/drives; no LTO as not competitive Forecast –Assuming +50PB/year in (+30PB in 2018) –Includes HW, maintenance, media –Cost/year usable TB: 8.2CHF(2014).. 5.4CHF(2017) Ian Bir d; HL- LH C co mp utin g wor ksh op

Data & Storage Services Beyond 2018? –Run 3 ( ): ~150PB/year –Run 4 ( ): ~600PB/year –Peak rates of ~80GB/s Longer term? Ian Bir d; HL- LH C co mp utin g wor ksh op Beyond 2018?

Data & Storage Services Longer term? Beyond 2018? –Run 3 ( ): ~150PB/year –Run 4 ( ): ~600PB/year –Peak rates of ~80GB/s Technology/market forecast (…risky for 15 years!) –INSIC Roadmap: +30% / yr tape capacity per $ (+20%/yr I/O increase) +20% / yr disk capacity per $ Ian Bir d; HL- LH C co mp utin g wor ksh op

Data & Storage Services Longer term? Beyond 2018? –Run 3 ( ): ~150PB/year –Run 4 ( ): ~600PB/year –Peak rates of ~80GB/s Technology/market forecast (…risky for 15 years!) –INSIC Roadmap: +30% / yr tape capacity per $ (+20%/yr I/O increase) +20% / yr disk capacity per $ Ian Bir d; HL- LH C co mp utin g wor ksh op

Data & Storage Services Cost prediction - with many assumptions: –No paradigm change…! –10% disk cache (with 20% redundancy overhead) –3y cycle for disks and tape drives, and 6 years for reusable enterprise tape media (repack every 3y) –Tape libraries upgraded/replaced around –No inflation Longer term? Total tape: ~19M CHF (2.1M CHF / year) Total % disk: ~45M CHF (5M CHF / year) Ian Bir d; HL- LH C co mp utin g wor ksh op

Technology outlook  Effective yearly growth: CPU 20%, Disk 15%, Tape 15%  Assumes:  75% budget additional capacity, 25% replacement  Other factors: infrastructure, network & increasing power costs Ian Bird; HL-LHC computing workshop10 5 September 2014

Storage costs Archives: -Expect that tape costs will be reasonable -Disk costs will increase -Overall – ~ doubling of cost for archiving and accessing the archived data Ian Bird; HL-LHC computing workshop 5 September

Active data - disk Assumes ratio of disk to yearly raw data is as currently requested for 2015 Assumes flat budget annual growth remains at 15-20% In 2025 cost is >x2 too high 5 September 2014 Ian Bird; HL-LHC computing workshop12 Problem is compounded by technology: -Steady decrease in costs (/2 every 18 months) is faltering -New technologies “close”, but not clear how easy to use, or if they need to be used differently (more like tape) -Real growth is likely to slow – our 15% assumption may be too high!

CPU CPU cost – assuming no change of model from today is x3 too expensive 5 September 2014 Ian Bird; HL-LHC computing workshop13 What we think is affordable unless we do something differently

100 PB 10 PB 1 PB 100 TB 10 TB 1 TB 100 GB ESnet traffic growth since 1990 A factor 10 every ~4.3 years ESnet traffic growth since 1990 A factor 10 every ~4.3 years 15.5 PB/mo in April 2013 Exponential fit Bytes/month transferred ESnet March September 2014 Ian Bird; HL- LHC comp uting works hop 14 Networking growth has been dramatic US ESnet as an example

Costs 5 September 2014 Ian Bird; HL-LHC computing workshop15 Main cost driver is active storage – disk

Resource growth - summary Assuming similar computing models as today: Networks -Technology growth will provide what we need; Cost ? Affordable if today’s trends continue Archive storage -Tape (robotics, drives, media) – cost similar to today for full anticipated HL-LHC data growth -Disk buffer cost will be much higher Active storage (data copies, caches, etc) -Costs factor 2-3 higher than flat budgets CPU -Costs factor 3-5 higher than flat budgets  Biggest impact on overall costs is disk storage 5 September 2014 Ian Bird; HL-LHC computing workshop16

Technical measures 1.Need to reduce disk usage significantly 2.Need to be more effective in use of CPU -But not so much that makes point 1 worse 3.Need to be able to use resources:  Simple interfaces are needed 0 config; 0 installed software … 5 September 2014 Ian Bird; HL-LHC computing workshop17 Pledged Opportunistic Commercial Grid Cloud HPC ??? Some of this only helps CPU, storage is not opportunistic (easily)

Funding related… Optimistic view is flat-cash budgets for pledged resources Must supplement with new resources -New countries, new collaborators, other opportunities (partnerships, …) -Unlikely to find factors more resources … Are there better ways to use available funding? -What are the FA’s prepared to do? -These are driven by national issues – environment, scale, decisions, … 5 September 2014 Ian Bird; HL-LHC computing workshop18

Open questions Why do we need to maintain (very) distributed computing -How far can some consolidation happen? Benefit with economies of scale. -Nationally? Internationally? Why do physicists need data locally? (this is a problem that is different for well connected and remote countries) -Better to process at or very near the storage -Reduce the number of copies of data to a minimum Can we imagine a model of (national) large centres forming a logical data processing hub which data does not leave? Outputs would be physics data sets. Access datasets stored in a “physics cloud” Could reduce from 200 centres to O(40) countries ? 5 September 2014 Ian Bird; HL-LHC computing workshop19

Funding agencies Expect to see (NOW!) common efforts towards future economies -They will not accept 4x software stacks for much longer -All 4 experiments are losing computing effort Also expect HEP to collaborate with other large data sciences (at least) to ensure their investments are re-used -Astronomy, astro-particle, cosmology, photon science, life-sciences, etc, etc; Not only do we have to express our own commonalities – we are expected to seek them with other disciplines If we don’t there is a real risk of further reduced funding 5 September 2014 Ian Bird; HL-LHC computing workshop20

Summary Fulfilling LHC resource needs for 2025 is a real challenge Existing computing models will not scale Existing insular attitudes will not be acceptable We need to aggressively address these issues -And show that we are doing so 5 September 2014 Ian Bird; HL-LHC computing workshop21