Niko Neufeld PH/LBC. Detector front-end electronics Eventbuilder network Eventbuilder PCs (software LLT) Eventfilter Farm up to 4000 servers Eventfilter.

Slides:



Advertisements
Similar presentations
Optimizing High Density Cooling Solutions for Data Centers
Advertisements

Using Copper Water Loop Heat Pipes to Efficiently Cool CPUs and GPUs Stephen Fried President Passive Thermal Technology, Inc.
Components of HVAC System
Terms 4 Definitions and Questions. Motherboard The main board of a computer, usually containing the circuitry for the central processing unit, keyboard,
Zycko’s data centre proposition The very best products and technologies Focus on what your customers need Turning products into solutions Wide range of.
Overclocking and Cooling
Cooling Product Positioning
Cloud Computing Data Centers Dr. Sanjay P. Ahuja, Ph.D FIS Distinguished Professor of Computer Science School of Computing, UNF.
IBM Energy & Environment © 2008 IBM Corporation Energy Efficiency in the Data Centre … and beyond Peter Richardson UK Green Marketing Leader.
Where Does the Power go in DCs & How to get it Back Foo Camp James Hamilton web: blog:
Copyright Green Revolution Cooling
Thermal Management Solutions from APW President Systems
Cooling Computer Systems By: Andrew Linthicum. Navigation Menu.
Refrigeration and Heat Pump Systems Refrigeration systems: To cool a refrigerated space or to maintain the temperature of a space below that of the surroundings.
Solar thermal energy Eng. Elamir Ahmed. Definition of solar thermal energy  Solar thermal energy is a renewable energy source.  Solar thermal uses technology.
CoolAir Temperature- and Variation-Aware Management for Free-Cooled Datacenters Íñigo Goiri, Thu D. Nguyen, and Ricardo Bianchini 1.
September 18, 2009 Critical Facilities Round Table 1 Introducing the Heat Wheel to the Data Center Robert (Dr. Bob) Sullivan, Ph.D. Data Center Infrastructure.
Niko Neufeld CERN PH/LBC. Detector front-end electronics Eventbuilder network Eventbuilder PCs (software LLT) Eventfilter Farm up to 4000 servers Eventfilter.
CNN Center John Hester Turner Properties, Inc.. CNN Center Built in ,583,000 square feet on 18 floors Five structures joined by a common atrium.
High Density Cooling Air conditioning for high density applications
Overview of Liquid Cooling Systems Peter Rumsey, Rumsey Engineers.
Air Conditioning and Computer Centre Power Efficiency The Reality Christophe Martel Tony Cass.
COMP 4923 A2 Data Center Cooling Danny Silver JSOCS, Acadia University.
A mother board is the main circuit board for the computer system (hence the name mother board ) The mother board holds all of the other components of a.
Farm Completion Beat Jost and Niko Neufeld LHCb Week St. Petersburg June 2010.
Definitions What is a network? A series of interconnected computers, linked together either via cabling or wirelessly. Often linked via a central server.
Smart Grid: Opportunities in the Data Center January 22, 2010 Charles O’Donnell, Vice President, Engineering Liebert AC Power Emerson Network Power.
1 5 December 2012 Laurent Roy Infrastructure / Electronics Upgrade -Cabling (long distance) -Rack and Crate (space, electrical distribution, cooling) -Detector.
Status and plans for online installation LHCb Installation Review April, 12 th 2005 Niko Neufeld for the LHCb Online team.
Physical Infrastructure Issues In A Large Centre July 8 th 2003 CERN.ch.
Management of the LHCb DAQ Network Guoming Liu * †, Niko Neufeld * * CERN, Switzerland † University of Ferrara, Italy.
Niko Neufeld PH/LBC. Detector front-end electronics Eventbuilder network Eventbuilder PCs (software LLT) Eventfilter Farm up to 4000 servers Eventfilter.
Schneider Electric: Total Cooling Solutions for Data Centers
Why it might be interesting to look at ARM Ben Couturier, Vijay Kartik Niko Neufeld, PH-LBC SFT Technical Group Meeting 08/10/2012.
Computing Facilities CERN IT Department CH-1211 Geneva 23 Switzerland t CF CERN Computer Centre Consolidation Project Vincent Doré IT Technical.
We can…. 2 GLOBAL REFERENCES Rev: 00 References :
COMP 4923 A2 Data Center Cooling Danny Silver JSOCS, Acadia University.
Niko Neufeld, CERN/PH. Online data filtering and processing (quasi-) realtime data reduction for high-rate detectors High bandwidth networking for data.
Management of the LHCb Online Network Based on SCADA System Guoming Liu * †, Niko Neufeld † * University of Ferrara, Italy † CERN, Geneva, Switzerland.
Energy Savings in CERN’s Main Data Centre
Common meeting of CERN DAQ teams CERN May 3 rd 2006 Niko Neufeld PH/LBC for the LHCb Online team.
Niko Neufeld HL-LHC Trigger, Online and Offline Computing Working Group Topical Workshop Sep 5 th 2014.
Niko Neufeld, CERN. Trigger-free read-out – every bunch-crossing! 40 MHz of events to be acquired, built and processed in software 40 Tbit/s aggregated.
Management of the LHCb DAQ Network Guoming Liu *†, Niko Neufeld * * CERN, Switzerland † University of Ferrara, Italy.
Niko Neufeld LHCC Detector Upgrade Review, June 3 rd 2014.
Future O 2 LHC P2 U. FUCHS, P. VANDE VYVRE.
ATCA COOLING PROJECT INTERSHIP FINAL PRESENTATION Piotr Koziol.
Introduction to DAQ Architecture Niko Neufeld CERN / IPHE Lausanne.
A two-stage system for the future cooling system.
Ventilation & Airflow NetShelter CX 18U,24U,38U Ambient office air is drawn in to the soundproofed air-intake chambers on either side of the NetShelter.
Cooling System Get the engine up to optimum operating Temperature as quickly as possible and maintains it at that temperature. Controls the heat produced.
Dell EMC Modular Data Centers
CANOVATE MOBILE (CONTAINER) DATA CENTER SOLUTIONS
Data Center Network Topologies
Unit 2: Chapter 2 Cooling.
Design of the thermosiphon Test Facilities 2nd Thermosiphon Workshop
The Data Center Challenge
ARAC/H/F Air-cooled water chillers, free-cooling chillers and heat pumps Range: kW.
Challenges in ALICE and LHCb in LHC Run3
Niko Neufeld LHCb Upgrade Online Computing Challenges CERN openlab Workshop on Data Center Technologies and Infrastructures, Mar 2017.
CERN Data Centre ‘Building 513 on the Meyrin Site’
COOLING & VENTILATION INFRASTRUCTURE
Cloud Computing Data Centers
Where Does the Power go in DCs & How to get it Back
Cloud Computing Data Centers
TEC assembly and commissioning at Lyon
Liebert DSE High efficiency thermal management
OCP Engineering Workshop Rack & Power, Advanced Cooling Solutions, Data Center Facility
Door Heat Exchanger: Specification
Presentation transcript:

Niko Neufeld PH/LBC

Detector front-end electronics Eventbuilder network Eventbuilder PCs (software LLT) Eventfilter Farm up to 4000 servers Eventfilter Farm up to 4000 servers UX85B Point 8 surface subfarm switch TFC x 100 Gbit/s subfarm switch Online storage Clock & fast commands 8800 Versatile Link 8800 Versatile Link throttle from PCIe40 Clock & fast commands 6 x 100 Gbit/s ECS Cooling for LHCb Online Upgrade - N. Neufeld 2

9000 links from detector Eventbuilder system of 500 – 600 servers and O(10) switches Eventfilter farm of up to 4000 servers, will start with ~ 1000 servers, O(100) switches Experiment Control System infrastructure, O(100) servers, storage O(10) Petabyte ItemPower Eventbuilder server500 W Eventbuilder switch5 kW Eventfilter serverup to 350 W Eventfilter switch300 W Controls server300 W Storage25 kW Total800 – 2000 kW Cooling for LHCb Online Upgrade - N. Neufeld 3

Online system cost is driven by number and type of interconnects shorter  cheaper faster  cheaper per unit of data transported power density higher  fewer racks, shorter distances, more compact systems spatial density higher  fewer containers Data-centre operation much easier on the surface in a non-controlled area Current LHCb data-centre is in UX85A Data-centre initial cost is definitely lowest for pre- fabricated (“container”) solution Data-centre operating cost depends on cooling solution (and to some extent on CPU technology ) 4 Cooling for LHCb Online Upgrade - N. Neufeld

Most compact system achieved by locating all Online components in a single location Power, space and cooling constraints allow such an arrangement only on the surface: containerized data- centre Cooling for LHCb Online Upgrade - N. Neufeld D2 & D1 (current ECS & farm New containerized data-centre 5

Baseline scenario Very little – some ECS infrastructure in D2 / D3 barracks Will continue with heat-exchanger doors using mixed water (max 200 kW) + some small amount of air- conditioning (mostly to keep the air reasonably dry ) Plan B in case of problem with long- distance versatile link No indication of any problem In that case need additional 400 kW (available today) Mixed-water air cooling (new rear- ddor heat-exchangers) Or direct liquid cooling in servers with central heat-exchangers to mixed water 6 Cooling for LHCb Online Upgrade - N. Neufeld

Existing infrastructure 70 kW on EOD (dual-feed, redundant) Used for critical services (ECS and storage) Will be kept for Run3 and beyond. A small increase (30 kW) in the same configuration is desirable 100 kW heat-load Cooling is done using mixed-water air-cooling in rear-door heat-exchangers System will be kept, probably with new (deeper/wider) racks, but same cooling principle These services are needed 24/7  redundancy for the mixed water facility is needed full redundancy with 2 nd group? or possibility to quickly connect mobile group? 7 Cooling for LHCb Online Upgrade - N. Neufeld

Main options identified until today: Direct Liquid Cooling (DLC) Natural free cooling (NFC) Water Cooled Heat Exchanger Doors (WCD) Racks (42U) will dissipate at least 20 kW Not considered: Classical data-centre airco using CRAC units, hot aisle, cold aisle  not very attractive for a completely new, non-legacy, containerized data-centre Solution must scale to 2 MW even though not everything needs to be deployed form day 1 Obviously want to go for a very high PUE (Power Usage Efficiency) == Power for IT / Power for data-centre Need cooling from 2017 as containers are being installed 8 Cooling for LHCb Online Upgrade - N. Neufeld

Easy to deploy Can be retro-fitted to exisiting servers Saves power by removing / shutting off server fans Potential for over-clocking the CPUs (more performance for same capex) ‘Excellent thermal stability (custom electronics boards, PCIe40) Can work with facility water from 15 to 50 C No constraints container layout Cost (manifold port + heat-sink + tubes) about 250 USD/server (not including heat exchanger with facility cooling) Part of the infrastructure (heat- sinks, tubes) need to be changed for every server generation Operational reliability needs to be tested Some airconditioning will be required anyhow: switches, memory modules, server power-supplies 9 Cooling for LHCb Online Upgrade - N. Neufeld

Relatively old solution (comes from the gaming market to the server) Prominent in proprietary solutions (e.g. HP Moonshot, IBM BlueGene) At least two vendor-agnostic providers: CoolIT, ASTec Principle: Liquid in closed loop goes through cooling plates on hot server elements (CPU, Memory, chipset, GPGPU, …) Connected via manifold to secondary loop 10 Cooling for LHCb Online Upgrade - N. Neufeld

Efficient connection to primary cooling at P8 which Tin can we get which Tout is acceptable flows Operational reliability tests will start soon Serviceability dense racks, repairs Detailed cost analysis Max in / max out temperature to evaporator towers 11 Cooling for LHCb Online Upgrade - N. Neufeld

Two-phase immersion cooling uses special low- boiling point fluids such as 3M Novec 12 Cooling for LHCb Online Upgrade - N. Neufeld

High density Non proprietary Novec and similar materials are non-sticky  equipment comes out dry Adapts to any-size (PCIe40!) Inherent flame-suppression Excellent thermal stability (overclocking possible) Density  only accessible from the top, incompatible with current rack-standards (OCP etc…) (this might change with wide-spread adoption) Cabling and other routine maintenance difficult – server must be removed, residual liquid Sealing of certain components vital (in particular hard-drives) Read also: om/it/whats-stopping-liquid- cooling/ om/it/whats-stopping-liquid- cooling/ 13 Cooling for LHCb Online Upgrade - N. Neufeld

AKA as Direct Air Cooling Air-side free cooling, where outside air is brought into the data center directly through filters or indirectly through heat exchangers. Adiabatic in which the air is brought to some sort of chamber and used along with water evaporation to cool the air Very “hip” now in data-centres, used by many sites achieving very low PUE (Facebook etc…) 14 Cooling for LHCb Online Upgrade - N. Neufeld

Very low operational cost Well suited for Geneva climate (Tmax < 35 C almost always) Can be bought ready integrated with many container solutions No modification of equipment High investment cost Will need some heating during some winter periods and certainly some drying – humidity control Requires careful control and monitoring (SCADA) – usually comes with the solution Puts constraints on rack and container layout Very inconvenient to work inside the racks 15 Cooling for LHCb Online Upgrade - N. Neufeld

LocationMax. Load [kW] UX85A D1 & D R-007 (SX85) Data-centre S82000 The upgraded LHCb online system has significantly larger power-needs than the current one, in particular in the event- building and event-filter. All new power-needs are in the to-be-built new data- centre  cooling solution the most important property of this project For existing, re-used location current power and battery backup are sufficient 16 Cooling for LHCb Online Upgrade - N. Neufeld