Commodity Data Center Design

Slides:



Advertisements
Similar presentations
SAN DIEGO SUPERCOMPUTER CENTER at the UNIVERSITY OF CALIFORNIA, SAN DIEGO Green Datacenter Initiatives at SDSC Matt Campbell SDSC Data Center Services.
Advertisements

Computer Room Requirements for High Density Rack Mounted Servers Rhys Newman Oxford University.
Data Center Design Issues Bill Tschudi, LBNL
Cooling Strategies for Small IT Rooms Presented by Jim Magallanes TechniCool Innovations Inc Upham Street Suite B-1, Broomfield.
Protective Packaging and Materials Handling
1 * Other names and brands may be claimed as the property of others. Copyright © 2010, Intel Corporation. Data Center Efficiency with Optimized Cooling.
Commodity Data Center Design James Hamilton
Columbia University’s Advanced Concepts Data Center Pilot June 17, 2011.
MidAmerican Energy Holdings Company Telecom Power Infrastructure Analysis Premium Power for Colocation Telecom Power Infrastructure Analysis February 27,
Cooling Product Positioning
Cloud Computing Data Centers Dr. Sanjay P. Ahuja, Ph.D FIS Distinguished Professor of Computer Science School of Computing, UNF.
PG&E and Altera Data Center Energy Efficiency Project.
Where Does the Power go in DCs & How to get it Back Foo Camp James Hamilton web: blog:
©2008 Pearson Education, Inc., Upper Saddle River, NJ. All rights reserved. This material is protected under all copyright laws as they currently exist.
Return of the Large Data Center. Computing Trends Computing power is now cheap, power hungry, and hot. Supercomputers are within reach of all R1 universities.
Utility-Function-Driven Energy- Efficient Cooling in Data Centers Authors: Rajarshi Das, Jeffrey Kephart, Jonathan Lenchner, Hendrik Hamamn IBM Thomas.
02/24/09 Green Data Center project Alan Crosswell.
Cloud Computing Economics Ville Volanen
CIT 470: Advanced Network and System AdministrationSlide #1 CIT 470: Advanced Network and System Administration Data Centers.
Architecture for Modular Data Centers James Hamilton 2007/01/17
Computer Room Experiences A medium sized tier-2 site view Pete Gronbech GridPP Project Manager HEPIX April 2012.
Learning Objectives Determine the need for warehouses and inventory management Identify the various types of warehouses Identify and describe the different.
24 x 7 Energy Efficiency February, 2007 William Tschudi
September 18, 2009 Critical Facilities Round Table 1 Introducing the Heat Wheel to the Data Center Robert (Dr. Bob) Sullivan, Ph.D. Data Center Infrastructure.
 Install new air cooled high efficiency screw chiller (variable speed)  Install new fan coils with ECM motors and low temperature heating coils and proper.
Architecture for Modular Data Centers James Hamilton 2007/01/08
GE Critical Power © 2010 General Electric Company. All Rights Reserved. This material may not be copied or distributed in whole or in part, without prior.
© 2010 Colt Telecom Group Limited. All rights reserved. Next Generation Data Centre Design Akber Jaffer 2.
Best Practices in HVAC Design/Retrofit
Data Centre Design. PROJECT BRIEF Develop a low energy data centre for a major computing company. Building is to be a prime example of sustainable, low.
1 Copyright © 2011, Elsevier Inc. All rights Reserved. Chapter 6 Authors: John Hennessy & David Patterson.
Mission Energy: Energy Efficient Cooling How Data Centre energy consumption and operating costs can be significantly reduced through readily available.
Chapter 3 Storage Systems
Strata IT Training Chapter 16 Green Networking. Green Servers Power Management –Power off monitor settings –Automatic server shutdown policy Rack mounted.
Lots of hype, little science But – lots of this hype is real, and there are many challenging engineering problems Initially, we focus on data centers:
Overview of Data Center Energy Use Bill Tschudi, LBNL
Thermal Design Project Final Report John Wallerich Principal Engineer Wallerich Group, LLC.
Thermal-aware Issues in Computers IMPACT Lab. Part A Overview of Thermal-related Technologies.
Dealing with Hotspots in Datacenters Caused by High-Density Computing Peter Hannaford Director of Business Development EMEA.
Proposed Cowlitz PUD Airflow Management Pilot Presented at the RTF IT Subcommittee March 5, 2013 Kerstin Rock, Director Michelle Lichtenfels, Senior Program.
MH...CH LECT011 What is Material Handling? Materials handling is the science and art both involving the moving, packing and storing of substance.
2 Automation Power Datacenter Facilities Automation - The Company Belgian Engineering Company founded in 1961 by private investors Market leader in conversion,
The Data Center Challenge
Energy Efficient Data Centre Design
Computing Facilities CERN IT Department CH-1211 Geneva 23 Switzerland t CF CERN Computer Centre Consolidation Project Vincent Doré IT Technical.
Authors: William Tschudi, Lawrence Berkeley National Lab Stephen Fok, Pacific Gas and Electric Company Stephen Fok, Pacific Gas and Electric Company Presented.
Increasing DC Efficiency by 4x Berkeley RAD Lab
Key Customer ChallengesCustomer Pain Points How You can Help the CustomerProductsSolutionsServices Increasing Density Difficult to maintain 300 cfm per.
3.3 Radiation In cold countries, solar hot water systems are often used to heat a house. In a typical solar hot water system, water from a storage tank.
CERN - IT Department CH-1211 Genève 23 Switzerland t Power and Cooling Challenges at CERN IHEPCCC Meeting April 24 th 2007 Tony Cass.
All content in this presentation is protected – © 2008 American Power Conversion Corporation Row Cooling.
1 PCE 2.1: The Co-Relationship of Containment and CFDs Gordon Johnson Senior CFD Manager at Subzero Engineering CDCDP (Certified Data Center Design Professional)
Data Center Energy Efficiency SC07 Birds of a Feather November, 2007 William Tschudi
Dell EMC Modular Data Centers
CANOVATE MOBILE (CONTAINER) DATA CENTER SOLUTIONS
Overview: Cloud Datacenters II
OCP for Telco The challenge and the solution
The Data Center Challenge
CERN Data Centre ‘Building 513 on the Meyrin Site’
Data Center Research Roadmap
IT Equipment Efficiency
Cloud Computing Data Centers
IT Equipment Efficiency
Where Does the Power go in DCs & How to get it Back
OCP for Telco The challenge and the solution
Architecture for Modular Data Centers
Commodity Data Center Design
Direct Current (DC) Data Center
Cloud Computing Data Centers
Energy Efficiency for the Machine (the Gigafactory) that Makes the Machine March 9, 2018.
Presentation transcript:

Commodity Data Center Design James Hamilton 2007-10-08 JamesRH@microsoft.com http://research.microsoft.com/~jamesrh

Containerized Products Nortel Steel Enclosure Containerized telecom equipment Caterpillar Portable Power Rackable Systems Concentro 1,152 Systems in 40’ (9,600 cores/3.5 PB) Datatainer ZoneBox Rackable Systems Container Cooling Model Google WillPower Will Whitted Petabox Internet Archive Brewster Kahle Sun Project Black Box 242 systems in 20’ 1/21/2007

Cooling, Feedback, & Air Handling Gains Verari Intel Tighter control of air-flow increased delta-T and overall system efficiency Expect increased use of special enclosures, variable speed fans, and warm machine rooms CRACs closer to servers for tighter temp control feedback loop Container takes one step further with very little air in motion, variable speed fans, & tight feedback between CRAC and load Intel 1/21/2007

Shipping Container as Data Center Module Contains network gear, compute, storage, & cooling Just plug in power, network, & chilled water Increased cooling efficiency Variable water & air flow Better air flow management (higher delta-T) 80% air handling power reductions (Rackable Systems) Bring your own data center shell Just central networking, power, cooling, security & admin center Can be stacked 3 to 5 high Less regulatory issues (e.g. no building permit) Avoids (for now) building floor space taxes Political/Social issues USA PATRIOT act concerns & regional restrictions Move resources closer to customer (CDN mini-centers) Single customs clearance on import Single FCC compliance certification Distributed, incremental fast built mini-centers 1/21/2007

Manufacturing & H/W Admin. Savings Factory racking, stacking & packing much more efficient Robotics and/or inexpensive labor Avoid layers of packaging Systems->packing box->pallet->container Materials cost and wastage and labor at customer site Data Center power & cooling expensive consulting contracts Data centers are still custom crafted rather than prefab units Move skill set to module manufacturer who designs power & cooling once Installation design to meet module power, network, & cooling specs More space efficient Power densities in excess of 1250 W/sq ft Rooftop or parking lot installation acceptable (with security) Stack 3 to 5 high Service-Free H/W admin contracts can exceed 25% of systems cost Sufficient redundancy that it just degrades over time At end of service, return for remanufacture & recycling 20% to 50% of systems outages caused by Admin error (A. Brown & D. Patterson) 1/21/2007

Systems & Power Density Estimating datacenter power density difficult (15+ year horizon) Power is 40% of DC costs Power + Mechanical: 55% of cost Shell is roughly 15% of DC cost Cheaper to waste floor than power Typically 100 to 200 W/sq ft Rarely as high as 350 to 600 W/sq ft Modular DC eliminates impossible shell to power trade-off Add modules until power is absorbed 480VAC to container High efficiency DC distribution within High voltage to rack can save >5% over 208VAC Over 20% of entire DC costs is in power redundancy Batteries able to supply up to 12 min at some facilities N+2 generation at over $2M each Instead, use more smaller, cheaper data centers Eliminate redundant power & bulk of shell costs Resource equalization 1/21/2007

Where do you Want to Compute Today? Slides posted soon to: http://research.microsoft.com/~JamesRH 10/08/2007