Early Experiences with Energy-Aware (EAS) Scheduling

Slides:



Advertisements
Similar presentations
Server Technology Confidential Page 1, June 22,2009 Server Technology Global Services.
Advertisements

Cloud Computing: Theirs, Mine and Ours Belinda G. Watkins, VP EIS - Network Computing FedEx Services March 11, 2011.
DoD HPC Modernization Program & Move Toward Emerging Architectures
Walter Binder University of Lugano, Switzerland Niranjan Suri IHMC, Florida, USA Green Computing: Energy Consumption Optimized Service Hosting.
Deploying GMP Applications Scott Fry, Director of Professional Services.
Network Design and Implementation
All content in this presentation is protected – © 2008 American Power Conversion Corporation Rael Haiboullin System Engineer Change Manager.
Green Cloud Computing Hadi Salimi Distributed Systems Lab, School of Computer Engineering, Iran University of Science and Technology,
Information Technology Center Introduction to High Performance Computing at KFUPM.
Windows Server System TM Overview IT Expectations: Do More with Less.
Presented by Scalable Systems Software Project Al Geist Computer Science Research Group Computer Science and Mathematics Division Research supported by.
IT PLANNING Enterprise Architecture (EA) & Updates to the Plan.
NOT FOR PUBLIC DISTRIBUTION State of Minnesota Technology Summary February 24, 2011.
1 PSC update July 3, 2008 Ralph Roskies, Scientific Director Pittsburgh Supercomputing Center
1 Maui High Performance Computing Center Open System Support An AFRL, MHPCC and UH Collaboration December 18, 2007 Mike McCraney MHPCC Operations Director.
11 SERVER CLUSTERING Chapter 6. Chapter 6: SERVER CLUSTERING2 OVERVIEW  List the types of server clusters.  Determine which type of cluster to use for.
Adaptive Server Farms for the Data Center Contact: Ron Sheen Fujitsu Siemens Computers, Inc Sever Blade Summit, Getting the.
Aim High…Fly, Fight, Win NWP Transition from AIX to Linux Lessons Learned Dan Sedlacek AFWA Chief Engineer AFWA A5/8 14 MAR 2011.
Next Generation of Apache Hadoop MapReduce Arun C. Murthy - Hortonworks Founder and Architect Formerly Architect, MapReduce.
Chapter 2 Computer Clusters Lecture 2.1 Overview.
1 May 12, 2010 Federal Data Center Consolidation Initiative.
Copyright © 2008 Altair Engineering, Inc. All rights reserved. PBS GridWorks - Efficient Application Scheduling in Distributed Environments Dr. Jochen.
Server Virtualization: Navy Network Operations Centers
Low-Power Wireless Sensor Networks
Integrating Fine-Grained Application Adaptation with Global Adaptation for Saving Energy Vibhore Vardhan, Daniel G. Sachs, Wanghong Yuan, Albert F. Harris,
What is Ubiquity™? ubiquity (noun): the state of being, or seeming to be, everywhere at once Ubiquity™ is a powerful building management system that.
November , 2009SERVICE COMPUTATION 2009 Analysis of Energy Efficiency in Clouds H. AbdelSalamK. Maly R. MukkamalaM. Zubair Department.
1 Overview 1.Motivation (Kevin) 1.5 hrs 2.Thermal issues (Kevin) 3.Power modeling (David) Thermal management (David) hrs 5.Optimal DTM (Lev).5 hrs.
Cloud Computing in NASA Missions Dan Whorton CTO, Stinger Ghaffarian Technologies June 25, 2010 All material in RED will be updated.
TASK PACKAGING Module 1 UNIT IV ADDITIONAL TOPICS " Copyright 2002, Information Spectrum, Inc. All Rights Reserved."
PET Summer Institute Kim Kido | Univ. Hawaii Manoa.
Ohio Supercomputer Center Cluster Computing Overview Summer Institute for Advanced Computing August 22, 2000 Doug Johnson, OSC.
The Red Storm High Performance Computer March 19, 2008 Sue Kelly Sandia National Laboratories Abstract: Sandia National.
Cray XT3 Experience so far Horizon Grows Bigger Richard Alexander 24 January 2006
Headline in Arial Bold 30pt HPC User Forum, April 2008 John Hesterberg HPC OS Directions and Requirements.
Issues Autonomic operation (fault tolerance) Minimize interference to applications Hardware support for new operating systems Resource management (global.
History of our Virtualization Project Discussion Points 1.How we started 2.Foundations of a Virtualization Project 3.Physical to Virtual 4.ELA 5.Lessons.
HPCMP Benchmarking Update Cray Henry April 2008 Department of Defense High Performance Computing Modernization Program.
VGreen: A System for Energy Efficient Manager in Virtualized Environments G. Dhiman, G Marchetti, T Rosing ISLPED 2009.
Scalable Systems Software for Terascale Computer Centers Coordinator: Al Geist Participating Organizations ORNL ANL LBNL.
1 Cray Inc. 11/28/2015 Cray Inc Slide 2 Cray Cray Adaptive Supercomputing Vision Cray moves to Linux-base OS Cray Introduces CX1 Cray moves.
MROrder: Flexible Job Ordering Optimization for Online MapReduce Workloads School of Computer Engineering Nanyang Technological University 30 th Aug 2013.
June 30 - July 2, 2009AIMS 2009 Towards Energy Efficient Change Management in A Cloud Computing Environment: A Pro-Active Approach H. AbdelSalamK. Maly.
Program Update Presented by Larry Davis, Deputy Director September 2009 Department of Defense High Performance Computing Modernization Program.
Rob Davidson, Partner Technology Specialist Microsoft Management Servers: Using management to stay secure.
Scalable and Coordinated Scheduling for Cloud-Scale computing
ISI Research Opportunities at Maui High Performance Computing Center ISI Research Opportunities at Maui High Performance Computing Center Developed for.
VISION & CHANDLER MIGRATION 10/03/2013. Executive Summary Last year we requested several critical Vision enhancements immediately post go-live.
Xi He Golisano College of Computing and Information Sciences Rochester Institute of Technology Rochester, NY THERMAL-AWARE RESOURCE.
Final Implementation of a High Performance Computing Cluster at Florida Tech P. FORD, X. FAVE, K. GNANVO, R. HOCH, M. HOHLMANN, D. MITRA Physics and Space.
Process Manager Specification Rusty Lusk 1/15/04.
OSG Area Coordinator’s Report: Workload Management Maxim Potekhin BNL May 8 th, 2008.
Ensieea Rizwani An energy-efficient management mechanism for large-scale server clusters By: Zhenghua Xue, Dong, Ma, Fan, Mei 1.
Tackling I/O Issues 1 David Race 16 March 2010.
Scheduling a 100,000 Core Supercomputer for Maximum Utilization and Capability September 2010 Phil Andrews Patricia Kovatch Victor Hazlewood Troy Baer.
Mobile Analyzer A Distributed Computing Platform Juho Karppinen Helsinki Institute of Physics Technology Program May 23th, 2002 Mobile.
Software Architecture of Sensors. Hardware - Sensor Nodes Sensing: sensor --a transducer that converts a physical, chemical, or biological parameter into.
Hosted Services Led by Jason Gross, Terrice McClain, & Jen Paulin
Our Experience with Desktop Virtualization
Server Virtualization IT Steering Committee, March 11, 2009
SOLAR PROGRAMS Presented By Benjamin H. Jordan, P.E. 4/2016.
Green IT Focus: Server Virtualization
OpenPBS – Distributed Workload Management System
Database Services at CERN Status Update
Computing Resource Allocation and Scheduling in A Data Center
Cray Announces Cray Inc.
Towards Green Aware Computing at Indiana University
Optena: Enterprise Condor
Presentation transcript:

Early Experiences with Energy-Aware (EAS) Scheduling Presentation Date

Project Motivation Overview Efforts To date Results Ongoing Efforts Agenda Project Motivation Overview Efforts To date Results Ongoing Efforts Summary

Idle nodes consume approximately 50% of the power of an active node. Project Motivation With the average HPCMP HPC System Utilization in the 80% range, opportunities exist to power off idle compute nodes, reducing energy costs for the idle nodes and associated cooling. Scheduling of reservations, system maintenance events, and scheduling of large node count jobs all create opportunities to power off idle nodes. Idle nodes consume approximately 50% of the power of an active node.

Project initiated via HPCMP Energy Efficiency Initiatives 8/11. Overview Project initiated via HPCMP Energy Efficiency Initiatives 8/11. Motivation: Explore the feasibility of powering off resources that are idle to save energy. There are commercial implementations on cyclic workloads that have proven successful, but is it feasible for large non-cyclic HPC workloads? Theory: Typically HPCMP systems run ~ 80% utilization. If we can power off 10% of system’s nodes there would be significant energy and monetary savings to the program. Annual HPCMP kWh saving by x% of systems for y hours/day. Architecture 10% 12h/D 15% 12h/D 10% 24h/D 15% 24h/D XE6 224012 336018 448024 672036 Altix 263725 395588 527450 791175 Subtotal 487737 731606 975474 1463211 Site (1.8 PUE) 390190 585284 780379 1170569 Total 877927 1316890 1755853 2633780

Overview (continued) Issues considered No reduction in service delivery or increased expansion factors Software/system viability: Architecture considerations: e.g. CRAY XT5 architecture will not recover from off node No significant performance degradation for jobs Idle nodes use 50%-60% of active nodes System reliability affects Cooling requirement reductions

Standardized system interfaces created for system control. Efforts To date Altair base code for commercial application significantly adapted to the scale of HPC systems/job mix, and non-cyclic job model. Standardized system interfaces created for system control. Code updated to be backfilled aware. Significant development/testing on ARL Test & Development Systems (TDS) passive mode live mode Network Affects test on Tana (CRAY XE6, 256 cores)

Efforts To date (continued) Production on Army Research Laboratory DoD Supercomputing Resource Center (ARL DSRC) systems Harold (SGI Altix ICE Quad Core Xeon Cluster 10,752 cores) TOW (SGI Altix ICE Quad Core Xeon Cluster 6,656 cores) ICECUBE (SGI Altix ICE Quad Core Xeon Cluster TDS, 96 cores) MJM (Linux Networx Evolocity II Xeon Cluster, 3.248 cores) Development applicable to Mana: US and other IB clusters Abutil (Appro Xtreme-X Series Cluster 1,824 cores) Cabutil (Appro Xtreme-X Series Cluster 1,008 cores) Production on Maui High Performance Computing Center (MHPCC) system Mana (Dell Quad Core Xeon Cluster (9,216 cores) Transitioned slowly to “live mode” starting 25 Apr 12 Muutil (Appro Quad Core Xeon Cluster EAS/PBSPro 11.2 installed. Ran in passive mode for 2 months. Waiting for approval.

Harold (SGI Altix ICE 10,752 cores) Results

Harold Results (continued) Moved reservation nodes from racks 17-21 to racks 1-5. Improved backfill jobs and freed up nodes for longer running jobs Scheduled PM on 5/27 shows nodes powered down as no pending jobs could complete prior to PM reservation. Emergency PM on 6/12 shows nodes powered down due to system inactivity.

Harold Results (continued) Harold EAS Savings Harold Utilization Apr 11,498 hrs ( 1,690 kWh) - 95.5% for Apr (7,396.061 hrs) May 31,604 hrs ( 4,645 kWh) - 89.9% for May (7,194,403 hrs) Jun 30,102 hrs ( 4,424 kWh) Jul 1,273 hrs ( 187 kWh) Aug 18,385 hrs ( 3,280 kWh) Limited Opportunity to save energy on Harold at its current level of utilization. Estimated Yearly Savings is $6,000

Harold Lifecycle Utilization OS Upgrade Harold took 7 months from start of production to reach 80% utilization. $200K of power cost expended over life cycle for idle nodes. Vendor ELT Start of Production (6 months to exceed 80%) CAP

TOW (SGI Altix ICE 6,656 cores) Results

TOW Results (continued) Lustre issues with too many nodes powering-on EAS disabled while node startup issues addressed. Lustre upgrade and tuning/EAS resumed on a limited set of nodes. Low EAS usage since 5/11 with increased workload/130 node dedicated project TOW EAS savings TOW Utilization Apr 15,555 hrs ( 2,286 kWh ) - 59.9% for Apr (2,872,756) May 1,333 hrs ( 195 kWh ) - 73.5% for May (3,617,621) Jun 6,335 hrs ( 931 kWh ) Jul 4,023 hrs ( 591 kWh) Aug 7,337 hrs (1,078 kWh)

Abutil (Appro Xtreme Cluster 1,408 cores) Results

Abutil Results (continued) EAS Configured on Abutil in February EAS was disabled for short period for user code testing Abutil EAS Savings Abutil Utilization Apr 29,267 hrs ( 11,024 kWh ) - 3.9% for Apr ( 49,924 hrs ) May 38,044 hrs ( 13,934 kWh ) - 3.6% for May ( 47,803 hrs ) Jun 37,007 hrs ( 13,544 kWh ) Jul 21,602 hrs ( 7,819 kWh ) Aug 21,195 hrs ( 7.826 kWh ) July timeframe implemented keep-alive feature and increased the powered-on nodes. Daily Average of 53 of 88 Nodes Powered-off Estimated Yearly Savings is approx $22k

ARL Cabutil Results

ARL Cabutil (continued) EAS Configured on Cabutil in March EAS Disabled for a Short Period to Test User Code Cabutil EAS Savings Cabutil Utilization Apr 18,900 hrs ( 6,128 kWh ) - 0.8% for Apr ( 5,541 hrs) May 18,701 hrs ( 6,063 kWh ) - 7.9% for May ( 58,163 hrs) Jun 20,840 hrs ( 7,383 kWh ) Jul 19,530 hrs ( 6,798 kWh ) Aug 10,306 hrs ( 3,390 kWh) Estimated Yearly Savings is approx $11k Daily Average of 37 of 44 nodes powered-off

MHPCC Mana Results

MHPCC Mana Results

MHPCC Mana Results (continued)

MHPCC Mana Results (continued) EAS Configured on Mana in March Mana Monthly Savings: Apr: 495,022 hrs ( 74,253 kWh ) - virtual savings May 247,144 hrs ( 37,071 kWh ) - about 50% of the system was under EAS control Jun 203,243 hrs ( 30,486 kWh ) - less than 50% was under EAS control Jul 261,925 hrs ( 39,273 kWh ) Aug 212,803 hrs ( 31,920 kWh ) Mana Estimated Yearly Savings approx $160k

Implement EAS on TI-12 systems across the HPCMP Ongoing Efforts Upgrade to PBSPro 11.3 Currently testing at ARL Utility Server Implement EAS on AFRL, ERDC and NAVY DSRC Utility Servers Implement EAS on TI-12 systems across the HPCMP

EAS found to be feasible on many architectures. Summary EAS found to be feasible on many architectures. EAS needs to be more tightly integrated with scheduler. EAS appears to be compatible with the IBM iDataPlex TI-12 systems. HPC Vendor resistance has been encountered but EAS node shutdown/startup provides an opportunity to perform proactive hardware checks, improving long term reliability.

Measured power savings are encouraging. Thanks to all! Summary (continued) Measured power savings are encouraging. Thanks to all! ARL (Project Sponsorship & Executive Agent). Lockheed Martin (LM) (Architectural Implementation, Algorithm Development, Performance/Feasibility Evaluation). Altair (Algorithm Development, Scheduler Interface) Instrumental (Architectural Implementation, HPCMP Coordination) Air Force Research Laboratory Maui High Performance Computing Center (MHPCC) ORS: XE6 evaluation. SGI (System monitoring, Architectural Implementation). APPRO: Support for EAS on HPCMP Utility Servers

Questions Questions?