Remote Hosting Project

Slides:



Advertisements
Similar presentations
Express5800/ft series servers Product Information Fault-Tolerant General Purpose Servers.
Advertisements

NASK RESEARCH AND ACADEMIC COMPUTER NETWORK Collocation at the NASK Collocation and Hosting Centre.
CERN IT Department CH-1211 Genève 23 Switzerland t The Wigner Data Centre An Extension to the CERN Data Centre.
© Crown copyright Met Office SRNWP Interoperability Workshop, ECMWF January 2008 SRNWP Interoperability Alan Dickinson Met Office.
André Augustinus 16 June 2003 DCS Workshop Safety.
ROAD ACCIDENT FUND VENDOR BRIEFING SESSION RAF/ 2012 /00032 Date: 15 January 2013 Time: 11:00.
Network Design and Implementation
DoD Information Technology Security Certification and Accreditation Process (DITSCAP) Phase III – Validation Thomas Howard Chris Pierce.
Documenting the Existing Network - Starting Points IACT 418 IACT 918 Corporate Network Planning.
Lesson 11-Virtual Private Networks. Overview Define Virtual Private Networks (VPNs). Deploy User VPNs. Deploy Site VPNs. Understand standard VPN techniques.
Zamtel Call Centre Tender PreBid Workshop 10 th March 2014.
CERN IT Department CH-1211 Genève 23 Switzerland t Options for Expanding CERN’s Computing Capacity Without A New Building Medium term plans.
Washington Metropolitan Area Transit Authority Pre-Proposal Conference Sourcing and Contracts Management System (CMS) Solution Request for Proposal FQ
Computing Facilities CERN IT Department CH-1211 Geneva 23 Switzerland t CF CERN Business Continuity Overview Wayne Salter HEPiX April 2012.
For more information, please contact 19/11/2010 EN/CV/DC Chiller design and purchase by Elena Perez Rodriguez.
SITA Presentation to Portfolio Committee on Police on SITA Services Provided to SAPS - FY2010/ October 2011 By: Acting Chairperson of Board Ms Fébé.
IT Department 29 October 2012 LHC Resources Review Board2 LHC Resources Review Boards Frédéric Hemmer IT Department Head.
Status Report on Tier-1 in Korea Gungwon Kang, Sang-Un Ahn and Hangjin Jang (KISTI GSDC) April 28, 2014 at 15th CERN-Korea Committee, Geneva Korea Institute.
Computing Facilities CERN IT Department CH-1211 Geneva 23 Switzerland t CF CERN Remote Hosting First Experiences Wayne Salter (with input.
CERN IT Department CH-1211 Genève 23 Switzerland t The new (remote) Tier 0 What is it, and how will it be used? The new (remote) Tier 0 What.
Resources and Financial Plan Sue Foffano WLCG Resource Manager C-RRB Meeting, 12 th October 2010.
IT Infrastructure Chap 1: Definition
Electronic Systems Support LHC Crates 2005 Paul Harwood PH Electronic Systems Support Group (PH-ESS) Meeting Sept 1, 2004.
Definitions What is a network? A series of interconnected computers, linked together either via cabling or wirelessly. Often linked via a central server.
TELE202 Lecture 5 Packet switching in WAN 1 Lecturer Dr Z. Huang Overview ¥Last Lectures »C programming »Source: ¥This Lecture »Packet switching in Wide.
E.Soundararajan R.Baskaran & M.Sai Baba Indira Gandhi Centre for Atomic Research, Kalpakkam.
Computing Facilities CERN IT Department CH-1211 Geneva 23 Switzerland t CF CERN Computer Facilities Evolution Wayne Salter / Vincent Doré.
ALMA Archive Operations Impact on the ARC Facilities.
CERN European Organization for Nuclear Research M. Nonis – ACCU Meeting – 3/12/2008 HOSTEL BOOKING SOFTWARE.
Computing Facilities CERN IT Department CH-1211 Geneva 23 Switzerland t CF CERN Computer Centre Upgrade Project Wayne Salter HEPiX November.
Phase II Purchasing LCG PEB January 6 th 2004 CERN.ch.
Computing Facilities CERN IT Department CH-1211 Geneva 23 Switzerland t CF Automatic server registration and burn-in framework HEPIX’13 28.
Network design Topic 6 Testing and documentation.
 The End to the Means › (According to IBM ) › 03.ibm.com/innovation/us/thesmartercity/in dex_flash.html?cmp=blank&cm=v&csr=chap ter_edu&cr=youtube&ct=usbrv111&cn=agus.
Computing Facilities CERN IT Department CH-1211 Geneva 23 Switzerland t CF CERN Computer Centre Consolidation Project Vincent Doré IT Technical.
Chapter 12 The Network Development Life Cycle
CERN - IT Department CH-1211 Genève 23 Switzerland t IT Dept Presentation [September 2009] - 1 User Support - Future Changes in Policy and.
Network management Network management refers to the activities, methods, procedures, and tools that pertain to the operation, administration, maintenance,
Teknologi Pusat Data 12 Data Center Site Infrastructure Tier Standard: Topology Ida Nurhaida, ST., MT. FASILKOM Teknik Informatika.
Computing Facilities CERN IT Department CH-1211 Geneva 23 Switzerland t CF CERN Computer Facilities Evolution Wayne Salter HEPiX May 2011.
Purchasing supplies at CERN Procurement Rules Someone FP-PI.
CERN - IT Department CH-1211 Genève 23 Switzerland t Power and Cooling Challenges at CERN IHEPCCC Meeting April 24 th 2007 Tony Cass.
© 2007 Cisco Systems, Inc. All rights reserved.Cisco Public ITE PC v4.0 Chapter 1 1 Planning a Network Upgrade Working at a Small-to-Medium Business or.
Contract: EIE/07/069/SI Duration: October 2007 – March 2010Version: July 7, 2009 Ventilation for buildings Energy performance of buildings Guidelines.
Dell EMC Modular Data Centers
Oracle & HPE 3PAR.
ROAD ACCIDENT FUND VENDOR BRIEFING SESSION RAF/ 2012 /00032
NATIONAL MEDIA VISITS Visits for print, radio, agency media during LS1 April April 2014 Hello, In order to reply to the huge amount of visit requests.
Remote Hosting Project
NERC Published Lessons Learned Summary
Tender Evaluation and Award Process
Purchasing supplies at CERN
Supported by: PCR Status report AESAG, September
Purchasing supplies at CERN
CERN Data Centre ‘Building 513 on the Meyrin Site’
Purchasing supplies at CERN
- Period before Tenders Submission –
Luca dell’Agnello INFN-CNAF
Olof Bärring LCG-LHCC Review, 22nd September 2008
Luca dell’Agnello Daniele Cesini GDB - 13/12/2017
Low temperature chiller Market Survey
Supported by: PCR Status report AESAG, September
Managing Clouds with VMM
Information and Site Visit for tender: “supply and installation of furniture and related services” EIOPA/OP/50/2014 Frankfurt, 12 May 2014.
THE EMPLOYER Right, of access to the site.
Information and Site Visit for tender: “supply and installation of furniture and related services” EIOPA/OP/50/2014 Frankfurt, 12 May 2014.
Operationalizing Export Certification and Regionalization Programmes
IT-4464/EN Bidders conference
Status of discussion after 7th meeting
Presentation transcript:

Remote Hosting Project LHCOPN Meeting May 2012 David Foster Material from Wayne Salter

Some History - I How to provide resources once the CERN CC is “full”? (Lack of power) Studies for a new CC on Prévessin site Four conceptual designs (2008/2009) Lack of on site experience Expensive! (Major capex investment) Interest from Norway to provide a remote hosting facility Some interesting ideas. Slow moving politics. Call for tender from all member states

Use of Remote Hosting Logical extension of physics data processing Batch and disk storage split across the two sites Business continuity Benefit from the remote hosting site to implement a more complete business continuity strategy for IT services

Some History - II Small scale local hosting in Geneva to gain experience Lack of critical power / gain experience Price enquiry for local hosting capacity 17 racks and up to 100kW Running successfully since summer 2010 Call for interest at FC June 2010 How much facility for 4MCHF/year? Is such an approach technically feasible? Is such an approach financially interesting? Deadline end of November 2010 Response Surprising level of interest – 23+ proposals Wide variation of solutions and capacity offered Many offering > 2MW (one even > 5MW) Assumptions and offers not always clearly understood Wide variation in electricity tariffs (factor of 8!)

Some History - III Many visits and discussions in 2010/2011 Official decision to go ahead taken in spring 2011 and all potential bidders informed Several new consortia expressed interest Call for tender Sent out on 12th Sept Specification with as few constraints as possible Draft SLA included A number of questions for clarification were received and answered (did people actually read the documents?) Replies were due by 7th Nov

Tender Specification - I Contract length 3+1+1+1+1 Reliable hosting of CERN equipment in a separated area with controlled access Including all infrastructure support and maintenance Provision of full configured racks including intelligent PDUs Services which cannot be done remotely Reception, unpacking and physical installation of servers All network cabling according to CERN specification Smart ‘hands and eyes’ Repair operations and stock management Retirement operations

Tender Specification - II Real time monitoring info to be provided into CERN monitoring system Installation of remote controlled cameras on request Equipment for adjudication: 2U CPU servers ~1kW and redundant PSUs 4U disk servers or SAS JBOD with 36 disks of 2TB ~ 450W and redundant PSUs Brocade routers up to 33U with 8 PSUs ~7kW 2U HP switches with redundant PSUs ~400W 1U KVM switches with single PSU 4 disk servers for every 3 CPU servers

Tender Specification - III Central star point: 4 racks for fibre patching 4 racks for UTP patching 2 racks for central switches All necessary racks for routers (see slides on SLA) External WAN connectivity: 2x100Gpbs capacity (separate paths) to the most convenient PoP in the GEANT network 100Gbps capacity as 1x100Gbps, or 3x40Gbps or 10x10Gbps

Parameters - I Provided for adjudication but will be reworked with contractor.   2013 2014 2015 2016 2017 2018 2019 Q1 Q3 Q4 Delivery Non-critical 50 350 250 650 500 Critical 150 100 Retirement 400 200 Total Installed 900 1150 1800 1400 1900 1650 300 600 450 Total 1200 1500 2400 2100 Numbers are KW The equipment shall be installed and operational within the racks within two working weeks of each delivery.

Parameters - II January 2013 10 racks for LAN routers and 2 racks for WAN routers  October 2016 2 additional racks for LAN routers  October 2018 4 additional racks for LAN routers Each router shall be installed and operational within its rack within one working week following delivery, including all necessary cabling. In the case of the delivery of multiple routers at one time, it shall be one elapsed working week for each delivered router.   2013 2014 2015 2016 2017 2018 2019 Q1 Q3 Q4 Delivery Non-critical 28 14 Critical 56 Retirement Total Installed 42 70 Total 84 98 126

Parameters - III CERN equipment shall run at all times Micro cuts Maintenance With the exception of downstream of in-room switchboards for non-critical equipment Equipment interruption once every two years < 4 hours For non-critical equipment this can be abrupt For critical equipment a 10 minute buffer must be provided to allow equipment to be switched off in a controlled manner Switch off to be triggered automatically

Parameters - IV Operating inlet temperature range 14-27°C with limited excursions up to a maximum of 32° C (based on the ASHRAE recommendations) Real time monitoring parameters: Current power usage of CERN equipment – 10 minutes Current capacity of UPS systems (if used) – 1 minute Infrastructure alarms indicating faults potentially affecting CERN equipment, including any loss of redundancy – 5 seconds Relevant temperature and humidity readings for the cooling infrastructure, e.g. inlet and return temperature of cooling air (and/or water), humidity of inlet air – 1 minute For retirements all equipment shall be removed and prepared for disposal or shipment within 4 weeks

Parameters - V The following summarises the expected failure rates of IT server components: Disks – 2% annual failure rate, i.e. 2 failures per year for every 100 disks Other standard components: Cooling fans, CPU, Disk, Memory module, PSU and RAID controller – 5 interventions per 100 servers per year More complex repairs such as replacing the mainboard, backplane or investigations of unknown failures – 3 interventions per 100 servers per year  The following repair times shall be respected: Disks – 8 working hours. Other standard components: Cooling fans, CPU, Disk, Memory module, PSU and RAID controller – 8 working hours. More complex repairs such as replacing the mainboard, backplane or investigations of unknown failures – to be shipped to the appropriate vendor within 10 working days. The following summarises the expected failure rates of networking equipment: Switches – 5 repair operations for every 100 switches per year. Routers – 5 repair operations per router per year. 4 working hours.

Parameters - VI Smart ‘hands and eyes’ supported to be provided during working hours (8 hours between 07:00 and 19:00 CET) 5 per hundred IT systems and 10 per network equipment per year Intervention time: Non-critical equipment – 1 working day Critical equipment, including networking – 4 working hours At the end of the contract all equipment shall be removed from racks and prepared for disposal or shipment within 8 weeks. Penalties defined for non respect of the SLA

Summary of Tender Replies Country Bid Declined Belgium 1   Finland France Germany Hungary 2 Norway 5 4 Poland Portugal Spain Sweden 3 Switzerland UK 16 14

Evaluation Process The financial offers were reviewed and in some cases corrected The technical compliance of a number of offers were reviewed (those which were within a similar price range) Meetings were held with 5 consortia to ensure that we understood correctly what was being offered they had correctly understood what we were asking for errors were discovered in their understanding Site selected and approved at FC 14th March

Wigner Data Centre in Budapest And the winner is…. Wigner Data Centre in Budapest (formerly KFKI RMKI)

Data Centre Location Easy reach to the airport and city centre Huge area within the fence Highly secure area Area only for research institutions Optimal environment

Status and Future Plans Contract should be signed on Friday (this week!) During 2012 Tender for network connectivity Small test installation Define and agree working procedures and reporting Define and agree SLA Integrate with CERN monitoring/ticketing system Define what equipment we wish to install and how it should be operated 2013 1Q 2013: install initial capacity (100kW plus networking) and beginning larger scale testing 4Q 2013: install further 500kW Goal for 1Q 2014 to be in production as IaaS with first Business Continuity services