Thermal Management of Heterogeneous Data Centers

Slides:



Advertisements
Similar presentations
Data Center Design Issues Bill Tschudi, LBNL
Advertisements

Team 1: Aaron, Austin, Dan, Don, Glenn, Mike, Patrick.
Alastair Dewhurst, Dimitrios Zilaskos RAL Tier1 Acknowledgements: RAL Tier1 team, especially John Kelly and James Adams Maximising job throughput using.
1 * Other names and brands may be claimed as the property of others. Copyright © 2010, Intel Corporation. Data Center Efficiency with Optimized Cooling.
Model predictive control for energy efficient cooling and dehumidification Tea Zakula Leslie Norford Peter Armstrong.
SQL Server 2000 Clustering Dave Fackler. Agenda Windows 2000 Clustering SQL Server 2000 Clustering Implementation Tips.
Dinker Batra CLUSTERING Categories of Clusters. Dinker Batra Introduction A computer cluster is a group of linked computers, working together closely.
Green Cloud Computing Hadi Salimi Distributed Systems Lab, School of Computer Engineering, Iran University of Science and Technology,
Chapter 15 Application of Computer Simulation and Modeling.
Network Redesign and Palette 2.0. The Mission of GCIS* Provide all of our users optimal access to GCC’s technology resources. *(GCC Information Services:
VIRTUALISATION OF HADOOP CLUSTERS Dr G Sudha Sadasivam Assistant Professor Department of CSE PSGCT.
 Install new air cooled high efficiency screw chiller (variable speed)  Install new fan coils with ECM motors and low temperature heating coils and proper.
Team Members Lora zalmover Roni Brodsky Academic Advisor Professional Advisors Dr. Natalya Vanetik Prof. Shlomi Dolev Dr. Guy Tel-Zur.
Maintaining a Microsoft SQL Server 2008 Database SQLServer-Training.com.
Sumit Kumar Archana Kumar Group # 4 CSE 591 : Virtualization and Cloud Computing.
APC InfraStruxure TM Central Smart Plug-In for HP Operations Manager Manage Power, Cooling, Security, Environment, Rack Access and Physical Layer Infrastructure.
Computer Networks Unit 1 – BTA 3O Ms. Chytra. Introduction to Networks Most people working in an office with more than a few computers will be using some.
David Bendit System Administrator Mars Space Flight Facility Arizona State University.
Multimedia & Networking Lab
Case Study 2 – TeraBit Inc. DMico Johnson Hans Schmidt.
Routing Protocol Evaluation David Holmer
Copyright 2010 – Johnson Controls, Inc. 1 A Day in the Life of a Smart Campus Clay Nesler VP, Global Energy & Sustainability Johnson Controls
NETWORK FILE ACCESS SECURITY Daniel Mattingly EKU, Dept. of Technology, CEN/CET.
Strata IT Training Chapter 16 Green Networking. Green Servers Power Management –Power off monitor settings –Automatic server shutdown policy Rack mounted.
1 Finding Constant From Change: Revisiting Network Performance Aware Optimizations on IaaS Clouds Yifan Gong, Bingsheng He, Dan Li Nanyang Technological.
Grid Computing at The Hartford Condor Week 2008 Robert Nordlund
Thermal-aware Issues in Computers IMPACT Lab. Part A Overview of Thermal-related Technologies.
BNL Tier 1 Service Planning & Monitoring Bruce G. Gibbard GDB 5-6 August 2006.
Software Architecture for Dynamic Thermal Management in Datacenters Tridib Mukherjee Graduate Research Assistant IMPACT Lab ( Department.
Thermal Aware Data Management in Cloud based Data Centers Ling Liu College of Computing Georgia Institute of Technology NSF SEEDM workshop, May 2-3, 2011.
Green Computing Metrics: Power, Temperature, CO2, … Computing system: Many-cores, Clusters, Grids and Clouds Algorithm and model: task scheduling, CFD.
Xi He Golisano College of Computing and Information Sciences Rochester Institute of Technology Rochester, NY THERMAL-AWARE RESOURCE.
Thermal Management in Datacenters Ayan Banerjee. Thermal Management using task placement Tasks: Requires a certain number of servers (cores) for a specified.
Adaptable Approach to Estimating Thermal Effects in a Data Center Environment Corby Ziesman IMPACT Lab Arizona State University.
Background Data Centre (DC) energy consumption doubled between 2000 and 2005 and grew by 50% from 2005 to 2010, consuming 1.5% of global primary energy.
LQCD Computing Project Overview
Monitoring Windows Server 2012
DISTRIBUTED SYSTEMS Principles and Paradigms Second Edition ANDREW S
Title of the Poster Supervised By: Prof.*********
SMART VACUUM CLEANER PRESENTED BY PROJECT PRESENTATION CSE 591
Thermal-aware Task Placement in Data Centers (part 4)
Forecasting with Cyber-physical Interactions in Data Centers (part 3)
Process Layout Chapter #6.
Management of Virtual Machines in Grids Infrastructures
Clinton A Jones Eastern Kentucky University Department of Technology
Chapter 1: Introduction
Creating an Oracle Database
Dresdner Bank & Wyndham International: The Business Case for Server Consolidation Chapter 3 Case 2.
Distributed Multimedia Systems
Computing Resource Allocation and Scheduling in A Data Center
THE STEPS TO MANAGE THE GRID
Grid Computing Colton Lewis.
LQCD Computing Operations
TYPES OF SERVER. TYPES OF SERVER What is a server.
3.2 Virtualisation.
Management of Virtual Machines in Grids Infrastructures
3.2 Virtualisation.
Campus Locator – Definition Phase (May04-04)
Seema Bandyopadhyay and Edward J. Coyle
Variables Print Quality Factors that influence Print Quality
HC Hyper-V Module GUI Portal VPS Templates Web Console
Analytics for the IoT: A Deep Dive into Algorithms
Distributed Databases
Zero Machine Failure Chen Linchevski, CEO, Co-founder.
Language Processors Application Domain – ideas concerning the behavior of a software. Execution Domain – Ideas implemented in Computer System. Semantic.
Subject Name: Operating System Concepts Subject Number:
COMP755 Advanced Operating Systems
Brown County General Hospital Virtualization Project
Overview of CSE and UW Computing Facilities
Yining ZHAO Computer Network Information Center,
Presentation transcript:

Thermal Management of Heterogeneous Data Centers David Bendit System Administrator Mars Space Flight Facility Arizona State University

Introduction Mars Space Flight Facility Located in the Moeur Building (West of the MU) Medium-sized server room Roughly 12’x25’ Over 100 physical servers Small cluster 26 physical machines, each running 4 Xen instances Total of 104 nodes optimally

Motivation Chiller had a bad habit of dying over weekends Come in Monday morning to a 90°F+ server room IMPACT research has studied the HPCI Cluster HCPI is more homogenous, dedicated to cluster processing Proper cooling setup More standard layout MSFF server room is sub-optimal setup, which doesn’t have as much research

Problem Statement What qualities are important in selecting a thermal-aware job scheduling algorithm for heterogeneous data centers Based on those qualities, which algorithms fare the best How does the heterogeneity of the data center affect the performance of the algorithms chosen?

Approach Physical Virtual Temperature sensors throughout the server room Wireless, single-hop sensor network Logging of cluster jobs and as many physical server temps as feasible Virtual Simulation of various scheduling strategies using FloVENT Will take into account server room layout and air flow

Difficulties Physical Originally supposed to work with another team to create temperature sensor network My contact in that team ended up dropping the class Logging cluster job traces Limited experience with new scheduling software at MSFF Logging individual server temperatures Heterogeneity of hardware means that not all supports reporting temperatures over SNMP or other means

Difficulties, cont. Virtual Server room floor plans We’ve placed things around where they fit, without more pre-planning than general ideas Needed to create this, and left out important details (namely, air vents and the like) Simulation software License is expired and needs to be renewed before any simulations can be run

Difficulties, cont. General MSFF is NASA funded As a US Government facility, we’re not able to let foreign nationals into the server room Frequent changes to the server room itself As data storage requirements, etc. change, we change equipment, so things can change week-to-week Results may not be implementable Because of very specific user requirements, useful changes may not be implemented anyway

Moving Forward Project will continue past the length of this course We will correlate cluster activity with machine and ambient temperatures End goal Reduce overall strain on the chiller My (rather pessimistic) prediction Because of large number of other computing resources in the server room, and always-on nature of our setup, cluster has negligible impact on overall temperature Need more wide-ranging changes for real temperature reduction

Questions?