Download presentation
Presentation is loading. Please wait.
Published byShannon Dorsey Modified over 8 years ago
1
CERN - IT Department CH-1211 Genève 23 Switzerland www.cern.ch/i t Power and Cooling Challenges at CERN IHEPCCC Meeting April 24 th 2007 Tony Cass
2
CERN - IT Department CH-1211 Genève 23 Switzerland www.cern.ch/it Basic Issues Computing equipment is not becoming more energy efficient – Or, rather, not as rapidly as performance improves Rack power density is increasing – From 1.5kW to 8kW now with 15+kW foreseen Power demand will grow with increasing requirement for LHC computing – Conservative assumptions lead to ~20MW by 2020 – “Moore’s law” growth in capacity, as seen for LEP, leads to prediction of ~100MW by 2020. “Critical” IT loads are at planned capacity limit of 250kW now and demand is growing – “critical” => with infinite diesel backup in the event of severe power outage. – “physics” load loses power after <10 minutes if both French and Swiss power are unavailable. Housing Future Computing Equipment - 2
3
Evolution of power and performance INTEL and AMD processors installed at CERN INTEL and AMD processors installed at CERN 2 cores 1 GHz 2 cores 1 GHz 2 cores 2.4 GHz 2 cores 2.4 GHz 2 cores 2.8 GHz 2 cores 2.8 GHz 4 cores 2.2 GHz 4 cores 2.2 GHz 4 cores 3 GHz 4 cores 3 GHz 8 cores 2.33 GHz 8 cores 2.33 GHz
4
CERN - IT Department CH-1211 Genève 23 Switzerland www.cern.ch/it Basic Issues Computing equipment is not becoming more energy efficient – Or, rather, not as rapidly as performance improves Rack power density is increasing – From 1.5kW to 8kW now with 15+kW foreseen Power demand will grow with increasing requirement for LHC computing – Conservative assumptions lead to ~20MW by 2020 – “Moore’s law” growth in capacity, as seen for LEP, leads to prediction of ~100MW by 2020. “Critical” IT loads are at planned capacity limit of 250kW now and demand is growing – “critical” => with infinite diesel backup in the event of severe power outage. – “physics” load loses power after <10 minutes if both French and Swiss power are unavailable. Housing Future Computing Equipment - 4
5
CERN - IT Department CH-1211 Genève 23 Switzerland www.cern.ch/it Project Power Evolution Housing Future Computing Equipment - 5
6
CERN - IT Department CH-1211 Genève 23 Switzerland www.cern.ch/it Basic Issues Computing equipment is not becoming more energy efficient – Or, rather, not as rapidly as performance improves Rack power density is increasing – From 1.5kW to 8kW now with 15+kW foreseen Power demand will grow with increasing requirement for LHC computing – Conservative assumptions lead to ~20MW by 2020 – “Moore’s law” growth in capacity, as seen for LEP, leads to prediction of ~100MW by 2020. “Critical” IT loads are at planned capacity limit of 250kW now and demand is growing – “critical” => with infinite diesel backup in the event of severe power outage. – “physics” load loses power after <10 minutes if both French and Swiss power are unavailable. Housing Future Computing Equipment - 6
7
CERN - IT Department CH-1211 Genève 23 Switzerland www.cern.ch/it Follow on issues The Meyrin site (with the Computer Centre) is at ~maximum consumption – 66MVA; limited by autotransfer system (between French & Swiss supplies) and feed from Prévessin. Diesel capacity is limited to 350kW for CC – we will couple the two 300kVA UPS modules to gain headroom at the expense of redundancy; no clear growth path thereafter (redundant) local diesel capacity?? B513 is very poorly designed from a modern HVAC standpoint – cooling 2.5MW will be a struggle, although there a number of optimisations still to make. CFD simulations interesting, but hampered by lack of real data on server air flow rates. Housing Future Computing Equipment - 7
8
CERN - IT Department CH-1211 Genève 23 Switzerland www.cern.ch/it What are we doing? Convincing ourselves air cooling is OK – Mostly done; power density of up to ~20kW/rack looks achievable within optimally designed building (long and thin, not square; unobstructed airflows; rigorous hot/cold aisle separation) – As an alternative, prefer “open” racks with heat exchanger on the back to “closed” racks with internal air flow. better able to cope with failure or, and more likely, door openings. Studying future options – using the existing computer centre, for example by installing equipment at a higher power/m 2 density; – using the “barn” [adjacent to the current machine room]; – using alternative buildings on one of the CERN sites - e.g. the former water tank (B226), the B186 assembly hall and B927 on the Prévessin site; – renting or purchasing space in a computing centre in the Geneva area; – purchasing the full computing service from a service provider (e.g. Amazon’s Computing Cloud). – Shipping container options (stop gap for 2010/11 if we don’t have a definitive solution by then???) Housing Future Computing Equipment - 8
9
CERN - IT Department CH-1211 Genève 23 Switzerland www.cern.ch/it Summary Power demand will exceed capacity by 2010 at the latest. Considering options to deliver increased capacity – but we are behind schedule to meet the 2010 crossover so stopgap solutions may be necessary. Money will be needed – Intel consider construction cost of new centre is $6/W. 40-60M€ for 20MW facility? But a modular design would spread costs. – Operation is 350k€/compute-MW/year assuming an HVAC overhead of 30% Housing Future Computing Equipment - 9
Similar presentations
© 2024 SlidePlayer.com. Inc.
All rights reserved.