Evolution of CERN Facilities

Slides:



Advertisements
Similar presentations
Pre-Commercial Procurement proposal - HNSciCloud
Advertisements

Ian Bird LHCC Referees’ meeting; CERN, 11 th June 2013 March 6, 2013
IT Department 29 October 2012 LHC Resources Review Board2 LHC Resources Review Boards Frédéric Hemmer IT Department Head.
Status of WLCG Tier-0 Maite Barroso, CERN-IT With input from T0 service managers Grid Deployment Board 9 April Apr-2014 Maite Barroso Lopez (at)
CERN IT Department CH-1211 Genève 23 Switzerland t The new (remote) Tier 0 What is it, and how will it be used? The new (remote) Tier 0 What.
GStore: GSI Mass Storage ITEE-Palaver GSI Horst Göringer, Matthias Feyerabend, Sergei Sedykh
1 Resource Provisioning Overview Laurence Field 12 April 2015.
This document produced by Members of the Helix Nebula Partners and Consortium is licensed under a Creative Commons Attribution 3.0 Unported License. Permissions.
JLAB Computing Facilities Development Ian Bird Jefferson Lab 2 November 2001.
Ian Bird LHC Computing Grid Project Leader LHC Grid Fest 3 rd October 2008 A worldwide collaboration.
CERN-IT Oracle Database Physics Services Maria Girone, IT-DB 13 December 2004.
Ian Bird, WLCG MB; 27 th October 2015 October 27, 2015
Claudio Grandi INFN Bologna CMS Computing Model Evolution Claudio Grandi INFN Bologna On behalf of the CMS Collaboration.
Ian Bird CMS Computing & Software CERN, 15 th October Oct 2015 Ian Bird; CMS Offline & Computing1.
ATLAS WAN Requirements at BNL Slides Extracted From Presentation Given By Bruce G. Gibbard 13 December 2004.
Slide David Britton, University of Glasgow IET, Oct 09 1 Prof. David Britton GridPP Project leader University of Glasgow UK-T0 Meeting 21 st Oct 2015 GridPP.
A European Open Science Cloud
Possibilities for joint procurement of commercial cloud services for WLCG WLCG Overview Board Bob Jones (CERN) 28 November 2014.
Ian Bird WLCG Networking workshop CERN, 10 th February February 2014
Ian Bird Overview Board; CERN, 8 th March 2013 March 6, 2013
1 Cloud Services Requirements and Challenges of Large International User Groups Laurence Field IT/SDC 2/12/2014.
Ian Bird, CERN WLCG LHCC Referee Meeting 1 st December 2015 LHCC; 1st Dec 2015 Ian Bird; CERN1.
CERN - IT Department CH-1211 Genève 23 Switzerland t Power and Cooling Challenges at CERN IHEPCCC Meeting April 24 th 2007 Tony Cass.
Ian Bird, CERN 1 st February Dec 2015
Farming Andrea Chierici CNAF Review Current situation.
The Helix Nebula marketplace 13 May 2015 Bob Jones, CERN.
Computing infrastructures for the LHC: current status and challenges of the High Luminosity LHC future Worldwide LHC Computing Grid (WLCG): Distributed.
EGI-InSPIRE RI EGI Compute and Data Services for Open Access in H2020 Tiziana Ferrari Technical Director, EGI.eu
Ian Bird LHCC Referees; CERN, 2 nd June 2015 June 2,
LHCb Computing 2015 Q3 Report Stefan Roiser LHCC Referees Meeting 1 December 2015.
Introduction To Cloud Computing By Diptee Chikmurge And Minakshi Vharkate Asst.Professor MIT AOE Alandi(D),Pune.
EGI-InSPIRE RI An Introduction to European Grid Infrastructure (EGI) March An Introduction to the European Grid Infrastructure.
Ian Bird, CERN WLCG Project Leader Amsterdam, 24 th January 2012.
Dynamic Extension of the INFN Tier-1 on external resources
Extending the farm to external sites: the INFN Tier-1 experience
WLCG IPv6 deployment strategy
WLCG Workshop 2017 [Manchester] Operations Session Summary
Understanding The Cloud
Ian Bird WLCG Workshop San Francisco, 8th October 2016
Resource Provisioning Services Introduction and Plans
Helge Meinhard, CERN-IT Grid Deployment Board 04-Nov-2015
Report from WLCG Workshop 2017: WLCG Network Requirements GDB - CERN 12th of July 2017
Computing models, facilities, distributed computing
LHCOPN update Brookhaven, 4th of April 2017
Computing activities in Napoli
Business and Pricing Models
Methodology: Aspects: cost models, modelling of system, understanding of behaviour & performance, technology evolution, prototyping  Develop prototypes.
ATLAS Cloud Operations
HEPiX Spring 2014 Annecy-le Vieux May Martin Bly, STFC-RAL
Openlab Compute Provisioning Topics Tim Bell 1st March 2017
Cloud Providers and AARC
INFN Computing infrastructure - Workload management at the Tier-1
Andrea Chierici On behalf of INFN-T1 staff
Scientific Computing Strategy (for HEP)
CERN Data Centre ‘Building 513 on the Meyrin Site’
for the Offline and Computing groups
How to enable computing
WLCG: TDR for HL-LHC Ian Bird LHCC Referees’ meting CERN, 9th May 2017.
Dagmar Adamova, NPI AS CR Prague/Rez
HNSciCloud Technical Aspects
Olof Bärring LCG-LHCC Review, 22nd September 2008
UK Status and Plans Scientific Computing Forum 27th Oct 2017
ALICE Computing Model in Run3
WLCG Collaboration Workshop;
New strategies of the LHC experiments to meet
Input on Sustainability
EGI Webinar - Introduction -
Brian Matthews STFC EOSCpilot Brian Matthews STFC
14th International IEEE eScience Conference
WP6 – EOSC integration J-F. Perrin (ILL) 15th Jan 2019
Presentation transcript:

Evolution of CERN Facilities Ian Bird, Helge Meinhard – CERN CWP Workshop Sam Diego, 23/01/2017 23 Jan 2017 CWP - San Diego

CERN Facilities today 2017-18/19 Upgrade internal networking capacity Refresh tape infrastructure 2017: 225k cores  325k 150 PB raw  250 PB 2018: ??? 23 Jan 2017 CWP - San Diego

Under construction – 2nd network hub Provide redundancy to external connectivity (GEANT, LHCOPN, LHCONE, ESNet, etc.) Physically separate from main CC – in Prévessin Project started some time ago – originally conceived as part of Wigner project to ensure business continuity Delayed by lack of funding Hopefully to be delivered in 2017 23 Jan 2017 CWP - San Diego

CERN Prévessin Data Centre? Drivers: Optimise the overall cost of LHC computing for CERN over the long-term (~20 years) ”online” and “offline” Political sensitivity to many large (multi-MW) data centres in the Pays de Gex Replacement of Wigner DC needed circa 2019 (+ 1 or 2) Significantly increased demands on CERN computing for Run3, Run4 And others: neutrino platforms, HPC uses, design for FCC, etc., etc. Consideration that software trigger systems/HLT could be part of a shared facility to enable synergies, operations costs, resource re-use Commercial resources are not likely to be cost-effective/practical for many of these needs Guaranteed availability (e.g. for HLT, Tier 0, calibration, data quality) Bandwidth/latency for HLT Reasonably scaled local facility for the fixed part of the likely load will still be optimal cost Elastic/on-demand needs should be separate from this 23 Jan 2017 CWP - San Diego

New DC – study A study group, mandated by the CERN Directorate, was set up to investigate whether a new DC on the Prévessin site might be a better solution than multiple smaller DCs, and other alternatives for additional capacity to look at the feasibility: technical, cost and schedule as very high bandwidth would be required from pits to new DC Assumed starting point to be a copy/paste of the GreenITCube design of GSI/FAIR Reported end July Essentially a positive recommendation on feasibility, although significant costs for networking 23 Jan 2017 CWP - San Diego

What is the GreenITCube? A new DC recently built at GSI in preparation for the FAIR project needs 6 story steel frame building (27x30x20m) with space for 128 racks per floor Separate building for cooling infrastructure and cooling towers 2N power distribution up to 12 MW Primary and secondary cooling circuits with N+1 redundancy Passive water cooled racks for up to 35 kW/rack Design of cooling and power distribution for pairs of floors, i.e. for 4 MW Recovered heat used for heating of an adjacent building Construction started in December 2014 and opening ceremony in January 2016 with first servers running in March 2016 Is being reproduced in other locations 23 Jan 2017 CWP - San Diego

GSI Experience CERN Directorate: PUE is extremely good < 1.1 (1.05 achieved during commissioning) Cost Currently equipped for 4 MW at a cost of 12M€ and foreseen cost of 16M€ for full 12MW Timescale: the full construction took just over 1 year Potential capacity 12 MW is ~3 times what we have in the CERN DC Capacity increase available in steps of 4MW CERN Directorate: Decision made to provisionally go ahead with an open tender for a Turnkey DC, based on a high level Functional Specification, but with strict qualification criteria Final decision expected shortly 23 Jan 2017 CWP - San Diego

CERN cloud procurements Since ~2014, series of short CERN procurement projects of increasing scale and complexity 23 Jan 2017 CWP - San Diego

2nd 6th 1st 23rd 20th Mar. Nov. Aug. End: 18th of Dec. 2015 End: 31st of March 2015 ATLAS simulation jobs Single core VMs Up to 3k VMs for 45 days 1st Cloud Procurement End: 18th of Dec. 2015 Target all VOs, simulation jobs 4-core VMs, O(1000) instances 2nd Cloud Procurement End: 30th of Nov. 2016 Provided by OTC IaaS 4-core VMs, O(1000) instances 500TB of central storage (DPM) 1k public IPs through GÉANT 3rd Cloud Procurement 2015 2016 Agreement between IBM and CERN CERN PoC to evaluate: Resource provisioning Network configurations Compute performance Transparent extension of CERN’s T0 End: 13th of May 2016 Sponsored Account “evaluation of Azure as an IaaS” Any VO, any workload Targeting multiple DCs: Iowa, Dublin and Amsterdam End: 30th of Nov. 2015 23rd Mar. 20th Nov. 23 Jan 2017 CWP - San Diego

Recent activity: T-Systems Running Cores Batch resources fully loaded shared among VOs WAN largely used Sometimes even saturated Mixture of “CPU-intensive” and “network-intensive” tasks MC workloads easier to manage Workloads CPU intensive workloads successfully run With some supplier issues I/O intensive workloads now being tested Network remains a concern for bandwidth and control Some suppliers have no public IPv4 addresses!

HELIX NEBULA The Science Cloud Procurers: CERN, CNRS, DESY, EMBL-EBI, ESRF, IFAE, INFN, KIT, SURFSara, STFC Procurers have committed funds (>1.6M€), manpower, use-cases with applications & data, in-house IT resources Experts: Trust-IT & EGI.eu Objective: procure innovative IaaS level cloud services Fully and seamlessly integrating commercial cloud (Iaas) resources with in-house resources and European e-Infrastructures To form a hybrid cloud platform for science Services will be made available to end-users from many research communities: High-energy physics, astronomy, life sciences, neutron/photon sciences, long tail of science Co-funded via H2020 (Jan’16-Jun’18) as a Pre-Commercial Procurement (PCP) project: Grant Agreement 687614, total procurement volume: >5M€ R&D is a very important part – represents > 50% of contract volume 23 Jan 2017 CWP - San Diego

HNSciCloud – Challenges Compute Mostly HTC, integrating some HPC requirements Full support for containers at scale Storage Caching at provider’s site, if possible automatically (avoid managed storage) Network Connection via GÉANT Support of identity federation (eduGAIN) for IT managers Procurement Match of cloud providers’ business model with public procurement rules 23 Jan 2017 CWP - San Diego

HNSciCloud – Design Phase Contractors T-Systems, Huawei, Cyfronet, Divia IBM RHEA Group, T-Systems, exoscale, SixSq Indra, HPE, Advania, SixSq Other major players not interested or dropped out just before tender submission 23 Jan 2017 CWP - San Diego

Provisioning services Moving towards Elastic Hybrid IaaS model: In house resources at full occupation Elastic use of commercial & public clouds Assume “spot-market” style pricing OpenStack Resource Provisioning (>1 physical data centre) HTCondor Public Cloud VMs Containers Bare Metal and HPC (LSF) Volunteer Computing IT & Experiment Services End Users CI/CD APIs CLIs GUIs Experiment Pilot Factories 23 Jan 2017 CWP - San Diego

CERN Facilities – longer term CERN Meyrin + Prevessin data centres Aim to keep fully occupied Option: “Wigner-like” or hosting for business continuity Could be a cloud-like solution too 2nd Network hub Elasticity from cloud/commercial resources Use as required – within cost envelope Extended with other opportunistic resources HPC, <lhcexperiment>@home, etc. Relative weight will be driven (mostly) by cost 23 Jan 2017 CWP - San Diego