Computer Centre Upgrade Status & Plans Post-C5, June 27 th 2003 CERN.ch.

Slides:



Advertisements
Similar presentations
State Data Center Re-scoped Projects With emphasis on reducing load on cooling systems in OB2 April 4, 2012.
Advertisements

The CDCE BNL HEPIX – LBL October 28, 2009 Tony Chan - BNL.
GB Patent No Intnl Patent Pending - Peter deBoeck © Slick System of Plug-in Electrical Accessories The new concept in electrical wiring devices.
MICE Refurbishment of CERN RF equipment for MICE M. Vretenar, CERN AB/RF.
State Data Center Re-scoped Projects With emphasis on reducing load on cooling systems in OB2 April 4, 2012.
LHC UPS Systems and Configurations: Changes during the LS1 V. Chareyre / EN-EL LHC Beam Operation Committee 11 February 2014 EDMS No /02/2014.
Proposed Design and the Operation of a Disconnecting Device for Small EG Installations Prepared by Hendri Geldenhuys.
Managing a computerised PO Operating environment 1.
A New Building Data Center Upgrade capacity and technology step 2011 – May the 4 th– Hepix spring meeting Darmstadt (de) Pascal Trouvé (Facility Manager.
CERN IT Department CH-1211 Genève 23 Switzerland t Options for Expanding CERN’s Computing Capacity Without A New Building Medium term plans.
GE Critical Power © 2010 General Electric Company. All Rights Reserved. This material may not be copied or distributed in whole or in part, without prior.
For more information, please contact 19/11/2010 EN/CV/DC Chiller design and purchase by Elena Perez Rodriguez.
CSG Panel – May 10, 2006 Key Lessons for New Data Centers Project overview Requirements Lessons learned Going Forward.
The HiLumi LHC Design Study is included in the High Luminosity LHC project and is partly funded by the European Commission within the Framework Programme.
IT Department 29 October 2012 LHC Resources Review Board2 LHC Resources Review Boards Frédéric Hemmer IT Department Head.
Electrical Installation 2
Turkish Power System Overview March 2012
1 MICE-Video Conference Status of MICE Hall Contents: General status Electrical Mechanical Preparing for Steps II & III Risks and worries Andy Nichols,
CV activities on LHC complex during the long shutdown Serge Deleval Thanks to M. Nonis, Y. Body, G. Peon, S. Moccia, M. Obrecht Chamonix 2011.
 Interconnecting DER With Electric Power Systems POWER SYSTEM 2002 CONFERENCE -- IMPACT OF DISTRIBUTED GENERATION Image.
T H E U N I N T E R R U P T I B L E P O W E R P R O V I D E R SineWave - Installation rules.
Computer Operations Group Data Centre Facilities Hitendra Patel June 2015.
UPS network perturbations in SX2 Vincent Chareyre EN-EL-SN ALICE Technical Coordination Meeting 7 May 2010.
Transforming B513 into a Computer Centre for the LHC Era Tony Cass —
Lyndonville Electric Department Feasibility Analysis Review December 2,
112/09/2013. Summary  Events on the network  Incidents  Network reconfigurations  Network operation  Update on activities  Tests  Projects and.
Building A Computer Centre HEPiX Large Cluster SIG October 21 st 2002 CERN.ch.
Review of the Project for the Consolidation and Upgrade of the CERN Electrical Distribution System October 24-26, 2012 THE HIGH-VOLTAGE NETWORK OF CERN.
Physical Infrastructure Issues In A Large Centre July 8 th 2003 CERN.ch.
Safe-Room Status Intermediate ‘Conclusions’ M. Brugger (based on HSE discussion)
LS1 : EL Upgrades & Consolidation Chamonix– February 8 th 2012 F. Duval on behalf of EN/EL The goal of LS1 is to implement everything needed for a safe.
Computing Facilities CERN IT Department CH-1211 Geneva 23 Switzerland t CF CERN Computer Facilities Evolution Wayne Salter / Vincent Doré.
Infrastructure Improvements 2010 – November 4 th – Hepix – Ithaca (NY)
GIF in EHN1- layout, construction and schedule Adrian Fabich, EN-MEF-LE GIF users meeting, 12 th February 2014.
Consolidation and Upgrade of the SPS Electrical Networks Current Status JP. Sferruzza, on behalf of EN/EL.
1 MICE Hall and Infrastructure - Progress since CM18 Chris Nelson MICE CM19 Oct 2007.
2 IMPACT - THE FIRE PERMIT = Hot Work Permit 3 Welcome ! This course is linked to the use of IMPACT, so it is assumed that: You know how to use IMPACT.
Computing Facilities CERN IT Department CH-1211 Geneva 23 Switzerland t CF CERN Computer Centre Upgrade Project Wayne Salter HEPiX November.
Managing the CERN LHC Tier0/Tier1 centre Status and Plans March 27 th 2003 CERN.ch.
Phase II Purchasing LCG PEB January 6 th 2004 CERN.ch.
Utility Engineers, PC.  Generation  Transmission  Distribution.
Computing Facilities CERN IT Department CH-1211 Geneva 23 Switzerland t CF CERN Computer Centre Consolidation Project Vincent Doré IT Technical.
ILC Test Facility at New Muon Lab (NML) S. Nagaitsev Fermilab April 16, 2007.
J. Pedersen, ST/EL Consolidation of the CERN technical infrastructure and civil engineering. J. Pedersen ST 6 th ST workshop, Thoiry, 1 – 3 April 2003.
Overview of the main events related to TS equipment during 2007 Definition Number and category of the events Events and measures taken for each machine.
ABOC/ATC days Electrical Network: Selectivity and Protection Plan José Carlos GASCON TS–EL Group 22 January 2007.
Energy Savings in CERN’s Main Data Centre
Teknologi Pusat Data 12 Data Center Site Infrastructure Tier Standard: Topology Ida Nurhaida, ST., MT. FASILKOM Teknik Informatika.
RRB scrutiny Report: Muons. Internal scrutiny group for RRB report Internal scrutiny group composition – J. Shank, G. Mikenberg, F. Taylor, H. Kroha,
Computing Facilities CERN IT Department CH-1211 Geneva 23 Switzerland t CF CERN Computer Facilities Evolution Wayne Salter HEPiX May 2011.
Computer Centre Upgrade Status & Plans Post-C5, October 11 th 2002 CERN.ch.
MH...CH LECT-021 SYSTEMS CONCEPT Adopting a materials handling systems from overall optimization point of view. Adopting a materials handling systems.
ATC / ABOC 23 January 2008SESSION 6 / MTTR and Spare Parts AB / RF GROUP MTTR, SPARE PARTS AND STAND-BY POLICY FOR RF EQUIPMENTS C. Rossi on behalf of.
Vault Reconfiguration IT DMM January 23 rd 2002 Tony Cass —
CERN - IT Department CH-1211 Genève 23 Switzerland t Power and Cooling Challenges at CERN IHEPCCC Meeting April 24 th 2007 Tony Cass.
Uninterruptible Power Supply Improves Reliability at The Australian Synchrotron Sean Murphy – ARW 2013 Melbourne.
Design considerations for the FCC electrical network architecture FCC week 2016, Rome, April 2016 Davide Bozzini - CERN/EN/EL With the contribution.
Introduction to the Data Centre Track HEPiX St Louis 7 th November 2007 CERN.ch.
110 th November 2010EN/CV/DC Detector Cooling Project - advancement status of Work Package N3 Thermosiphon project Michele Battistin 18 May 2011.
West Cambridge Data Centre Ian Tasker Information Services.
Schedule and Milestones Roy Preece STFC RAL FAC 17 th April 2015.
S A Griffiths CM42 June 2015 Electrical & Control.
CERN Data Centre ‘Building 513 on the Meyrin Site’
Background B513 built in early 1970’s for mainframe era.
LV Safe Powering from UPS to Clients
the CERN Electrical network protection system
Electrical and electronic systems projects
Maintenance at ALBA facility: Strategy for the conventional services
Modular Edge-connected data centers
Modular Edge-connected data centers
Presentation transcript:

Computer Centre Upgrade Status & Plans Post-C5, June 27 th 2003 CERN.ch

CERN.ch 2 The Plan (June 2001)  Configure the vault as a second Machine Room  Install new computing equipment in the vault  Rework the 18kV supply to B513  Renew the UPS configuration  Renew Machine Room power distribution, air conditioning and fire protection This Talk Critical UPS with substation. Physics UPS in This Talk

CERN.ch 3

4

5 The Substation

CERN.ch 6 A New Substation for B513  Under the car park between B513 & Restaurant 2. –location selected 2001, agreement in principle at Site Committee in February  Initial design in October 2002, refined since. –Approved by the Site Committee on 7 th May –Tender sent out on 16 th May; Deadline for reply is today, envelope opening ~ July 7 th. »4 companies seem interested –Construction to start in August, finish by March 2004 »will lead to temporary loss of ~20 parking spaces so consider using Salėve side car park. »some car pool spaces will need relocating permanently. –Electrical equipment installation from April/May 2004 –Commissioning after accelerator shutdown in November 2004.

CERN.ch 7

8

9

10

CERN.ch 11 Supply Failure & Maintenance Issues  In addition to supporting a computing load of up to 2.5MW, the new substation design includes reliability and redundancy features to increase the likelihood that we can run uninterrupted for long periods. –This extends to aspects of the design in the machine room—e.g. dual power inputs for the PDUs.  The next few slides present the impact of failure at the various different levels –18kV supply –18kV switchboards –Transformers –400V switchboards –UPS modules.

CERN.ch 12 18kV Supply  Improved redundancy of 18kV supply as B513 will be in the 18kV loop, not on a spur.  Up to 250kW of equipment maintained indefinitely on Diesel power. –The Diesel Generators are now tested annually. IT must insist this continues.  No power for physics systems if both EDF and EoS supplies fail (i.e. the autotransfer mechanism fails)  Temporary supply is available from the West Area to cover tests. The power available is limited to 6.5MVA, but is adequate to cover needs in winter.  Frequency: –Failure: ?? –Maintenance: Annual

CERN.ch 13 18kV Switchboards  Two Cases: –Switchboard for safety systems and critical IT services »Failover system guarantees continued power. Although this is limited to a maximum of 2MVA, the safety systems and critical service load is anyway limited to less than 2MVA by the power available from the Diesel Generators. –Switchboard for general and physics services »Maximum of 1.2MVA available from the safety system switchboard. This is enough to support a reduced level of air conditioning in the vault and the machine room, adequate to support the 250kW critical load. »No power available to support the physics load.  Frequency: –Failure: extremely rare –Maintenance: Every 10 years

CERN.ch 14 Transformers  Two cases –Transformer for Physics services »Failover system ensures continuity of power, but systems behind UPS system connected to failing transformer are at risk from any power fluctuations. Full UPS protection can be assured (even with one transformer down), but this requires a manual intervention. –Transformer for general services »Failure leads to loss of power for part of the overall equipment. This will include part of the air conditioning equipment and the reduced cooling capacity will not be sufficient to maintain acceptable conditions should the failure occur in summer. »Manual intervention allows load of up to 2MVA which is adequate to support the power needs of air conditioning in winter.  Frequency –Failure: extremely rare –Maintenance: annual, but cut required only every 3 years

CERN.ch V Switchboards  Same two cases –Switchboard for Physics services »Failover system ensures continuity of power, but systems behind UPS system connected to failing switchboard are at risk from any power fluctuations. Full UPS protection can be assured (even with one switchboard down), but this requires a manual intervention. –Switchboard for general services »Failure leads to loss of power for part of the overall equipment. This will include part of the air conditioning equipment and the reduced cooling capacity will not be sufficient to maintain acceptable conditions should the failure occur in summer. »Manual intervention allows load of up to 2MVA which is adequate to support the power needs of air conditioning in winter.  Frequency –Failure: extremely rare –Maintenance: every 6 years

CERN.ch 16 UPS Modules  Each UPS system has N+1 redundancy, i.e. failure of a single module is transparent for the load. Two modules must fail for there to be any impact.  UPS module failure normally switches systems to the standard mains supply. If two UPS modules fail: –No loss of service, but physics systems would run unprotected. –UPS coverage maintained for sytems with redundant power supplies connected to critical and physics UPS.  If failure of two UPS modules leads to loss of power –1/3 rd of physics services lost –systems with redundant power supplies connected to critical and physics UPS continue running.  Frequency –Failure: ~once per 10 years catastrophically –Maintenance: Annual, but exposure to mains failure only during bypass module maintenance (~2 hours).

CERN.ch 17 Reliability Conclusions  Overall, there is an improved level of reliability and redundancy as compared to the present situation. –c.f. changes to earthing arrangements following incident in April.  In itself, the electrical supply system should allow continuous operation of the computer centre over a long period. –Unfortunately, maintenance on the 18kV switchboards requires a shutdown every 10 years.  BUT: both a smooth power supply and adequate air conditioning are required. Improved redundancy of supply to the air conditioning system is desirable. This is being studied now.

CERN.ch 18 The Machine Room

CERN.ch 19 Machine Room Upgrade: Why?  More Power –We want to be able to use up to 2MW »c.f. 600kW available today »requires normabarre density for ~3MW for flexibility  Better layout –Have only related services connected to a given normabarre »today’s layout causes confusion during power failures u (although these won’t happen in future…)  And, while we’re at it, clean up rubbish and obsolete cables from the false floor void. –Likely to be air flow under the false floor

Current Machine Room Layout Problem: Normabarres run one way, services run the other…. Services

Proposed Machine Room Layout 18m double rows of racks 12 Mario racks or 36 19” racks 528 box PCs 105kW U PCs 288kW 324 disk servers 120kW(?) 9m double rows of racks for critical servers Aligned normabarres

CERN.ch 22 Power Distribution  One PDU per 18m double row –400A/phase, supports load up to 260kW –4 9m normabarres in 2 pairs »27 connections »50A/phase for systems without power factor correction »100A/phase for power factor corrected systems –1 18m normabarre –1 outlet unused  12 PDUs for Physics –3.1MW total »we will not be (greatly) limited by the available power when locating racks and services u BUT: problems with, e.g. disk servers which need 3 16A sockets each. We need to reduce the impact redundant power supplies have on the use of normabarre connections.  Other PDUs as necessary for critical equipment –including power to networking areas in Salėve side corridor

CERN.ch 23 Power Factor Correction  Power supplies on early PCs had a power factor of only 0.7. –This gives high currents in the neutral conductor and the electrical distribution must be designed carefully. –Filters installed at the PDU level give overall power factor correction so the UPS sees a well behaved load. »Without filters, bigger UPS and upstream PDUs are required  EU directive requires PC systems delivered after 1 st January 2002 to have power factor correction. –Latest batch of PCs have power factor better than  We propose to assume power factor of 0.95 for future equipment. –Avoids purchase and installation of filters (save ~500KCHF) –PDUs can support full 400A/phase, (without correction, current in neutral conductor reaches 400A at about 250A/phase). »Need ~14 PDUs, not ~21  further financial saving.

CERN.ch 24 Machine Room Air Conditioning  Upgrade needed…  B513 corridors will not be able to take cold air from the machine room.

CERN.ch 25 Air Conditioning Upgrade Plans  Detailed air conditioning upgrade plan still in preparation. Key points as they are understood today are –additional chiller to be installed in 2005 (once the new substation is commissioned) –Machine room cooling probably requires at least some cold air to be routed under the false floor »some ducting exists already on Salėve wall. More may be needed. –Need to improve redundancy »or, at least, understand the impact and frequency of the different possible failures of critical elements (including the difference between failure in winter and in summer).  Major upgrade foreseen for Winter of 2004/5.

CERN.ch 26 Some Small Issues

CERN.ch 27 Two concerns  Will power demand exceed capacity of current UPS before 2006? –If so, use the 250kW UPS to provide additional capacity (850kW total). Supply redundancy delayed until  61cm 2 floor tiles are no longer available. –520 tiles bought in 2001—all of the remaining stock from Steel Systems. »They said we would be able to buy more later, although at a significantly higher price, but now say they can no longer do this. –Stock expected to be enough for 10 years, but we should go carefully. –Long term, consider replacing the false floor in offices, corridors and the barn to liberate tiles for the machine room.

CERN.ch 28 Summary

CERN.ch 29 Summary  Substation construction to start soon –Apologies for the disruption… –Commissioning at end of 2004  Future Machine Room Layout firming up –Comments/objections soon! –Electrical supply upgrade to start from September.  Power factor of at least 0.95 assumed for future equipment.  Air conditioning upgrade being planned now –Would like increased redundancy –Almost certain to be air flow under the false floor –B513 corridors will no longer be able to extract cold air from the machine room.