Toward Runtime Power Management of Exascale Networks by On/Off Control of Links Ehsan Totoni University of Illinois-Urbana Champaign, PPL Charm++ Workshop,

Slides:



Advertisements
Similar presentations
Sabyasachi Ghosh Mark Redekopp Murali Annavaram Ming-Hsieh Department of EE USC KnightShift: Enhancing Energy Efficiency by.
Advertisements

Green Network Project Contract
Misbah Mubarak, Christopher D. Carothers
Hadi Goudarzi and Massoud Pedram
ElasticTree: Saving Energy in Data Center Networks Brandon Heller, Srini Seetharaman, Priya Mahadevan, Yiannis Yiakoumis, Puneed Sharma, Sujata Banerjee,
The Who, What, Why and How of High Performance Computing Applications in the Cloud Soheila Abrishami 1.
Green Cloud Computing Hadi Salimi Distributed Systems Lab, School of Computer Engineering, Iran University of Science and Technology,
Institute of Networking and Multimedia, National Taiwan University, Jun-14, 2014.
Power Cost Reduction in Distributed Data Centers Yuan Yao University of Southern California 1 Joint work: Longbo Huang, Abhishek Sharma, LeanaGolubchik.
Small-World Graphs for High Performance Networking Reem Alshahrani Kent State University.
Dr. Gengbin Zheng and Ehsan Totoni Parallel Programming Laboratory University of Illinois at Urbana-Champaign April 18, 2011.
“ElasticTree: Saving energy in data center networks“ by Brandon Heller, Seetharaman, Mahadevan, Yiakoumis, Sharma, Banerjee, McKeown presented by Nicoara.
Application Models for utility computing Ulrich (Uli) Homann Chief Architect Microsoft Enterprise Services.
ECE669 L12: Interconnection Network Performance March 9, 2004 ECE 669 Parallel Computer Architecture Lecture 12 Interconnection Network Performance.
NCKU CSIE CIAL1 Principles and Protocols for Power Control in Wireless Ad Hoc Networks Authors: Vikas Kawadia and P. R. Kumar Publisher: IEEE JOURNAL ON.
Datacenter Power State-of-the-Art Randy H. Katz University of California, Berkeley LoCal 0 th Retreat “Energy permits things to exist; information, to.
3. Interconnection Networks. Historical Perspective Early machines were: Collection of microprocessors. Communication was performed using bi-directional.
Adaptive Web Caching Lixia Zhang, Sally Floyd, and Van Jacob-son. In the 2nd Web Caching Workshop, Boulder, Colorado, April 25, System Laboratory,
1 A Novel Mechanism for Flooding Based Route Discovery in Ad hoc Networks Jian Li and Prasant Mohapatra Networks Lab, UC Davis.
Strategies for Implementing Dynamic Load Sharing.
OS Fall ’ 02 Performance Evaluation Operating Systems Fall 2002.
Performance Engineering and Debugging HPC Applications David Skinner
Ehsan Totoni Josep Torrellas Laxmikant V. Kale Charm Workshop April 29th, 2014.
ElasticTree: Saving Energy in Data Center Networks 許倫愷 2013/5/28.
Ekrem Kocaguneli 11/29/2010. Introduction CLISSPE and its background Application to be Modeled Steps of the Model Assessment of Performance Interpretation.
Folklore Confirmed: Compiling for Speed = Compiling for Energy Tomofumi Yuki INRIA, Rennes Sanjay Rajopadhye Colorado State University 1.
Application-specific Topology-aware Mapping for Three Dimensional Topologies Abhinav Bhatelé Laxmikant V. Kalé.
1J. Kim Web Science & Technology Forum Enabling Hardware Technology for Web Science John Kim Department of Computer Science KAIST.
Interoperabilidade com Windows Azure Computação em nuvem no Windows Azure com suporte a PHP, Java, Ruby e Python Rogerio Cordeiro Technical Evangelist.
Towards Sustainable Portable Computing through Cloud Computing and Cognitive Radios Vinod Namboodiri Wichita State University.
Principles of Scalable HPC System Design March 6, 2012 Sue Kelly Sandia National Laboratories Abstract: Sandia National.
جلسه دهم شبکه های کامپیوتری به نــــــــــــام خدا.
Cloud Computing Energy efficient cloud computing Keke Chen.
Copyright © 2011, Programming Your Network at Run-time for Big Data Applications 張晏誌 指導老師:王國禎 教授.
Introduction to Cloud Computing
Planning and Designing Server Virtualisation.
Department of Computer Science at Florida State LFTI: A Performance Metric for Assessing Interconnect topology and routing design Background ‒ Innovations.
Energy Usage in Cloud Part2 Salih Safa BACANLI. Cooling Virtualization Energy Proportional System Conclusion.
Extreme scale parallel and distributed systems – High performance computing systems Current No. 1 supercomputer Tianhe-2 at petaflops Pushing toward.
Extreme-scale computing systems – High performance computing systems Current No. 1 supercomputer Tianhe-2 at petaflops Pushing toward exa-scale computing.
Scalable Reconfigurable Interconnects Ali Pinar Lawrence Berkeley National Laboratory joint work with Shoaib Kamil, Lenny Oliker, and John Shalf CSCAPES.
1 Distributed Energy-Efficient Scheduling for Data-Intensive Applications with Deadline Constraints on Data Grids Cong Liu and Xiao Qin Auburn University.
Scalable Multi-Class Traffic Management in Data Center Backbone Networks Amitabha Ghosh (UtopiaCompression) Sangtae Ha (Princeton) Edward Crabbe (Google)
Parallelization of Classification Algorithms For Medical Imaging on a Cluster Computing System 指導教授 : 梁廷宇 老師 系所 : 碩光通一甲 姓名 : 吳秉謙 學號 :
Thermal-aware Issues in Computers IMPACT Lab. Part A Overview of Thermal-related Technologies.
Michihiro Koibuchi(NII, Japan ) Tomohiro Otsuka(Keio U, Japan ) Hiroki Matsutani ( U of Tokyo, Japan ) Hideharu Amano ( Keio U/ NII, Japan ) An On/Off.
A lower bound to energy consumption of an exascale computer Luděk Kučera Charles University Prague, Czech Republic.
Non-Data-Communication Overheads in MPI: Analysis on Blue Gene/P P. Balaji, A. Chan, W. Gropp, R. Thakur, E. Lusk Argonne National Laboratory University.
Network-Aware Scheduling for Data-Parallel Jobs: Plan When You Can
Modeling Billion-Node Torus Networks Using Massively Parallel Discrete-Event Simulation Ning Liu, Christopher Carothers 1.
Tufts Wireless Laboratory School Of Engineering Tufts University Paper Review “An Energy Efficient Multipath Routing Protocol for Wireless Sensor Networks”,
1 Optimal operation of energy storage in buildings: The use of hot water system Emma Johansson Supervisors: Sigurd Skogestad and Vinicius de Oliveira.
HPC HPC-5 Systems Integration High Performance Computing 1 Application Resilience: Making Progress in Spite of Failure Nathan A. DeBardeleben and John.
Ensieea Rizwani An energy-efficient management mechanism for large-scale server clusters By: Zhenghua Xue, Dong, Ma, Fan, Mei 1.
Storage Systems CSE 598d, Spring 2007 OS Support for DB Management DB File System April 3, 2007 Mark Johnson.
CprE 458/558: Real-Time Systems (G. Manimaran)1 CprE 458/558: Real-Time Systems Energy-aware QoS packet scheduling.
Topology How the components are connected. Properties Diameter Nodal degree Bisection bandwidth A good topology: small diameter, small nodal degree, large.
Jennifer Rexford Fall 2010 (TTh 1:30-2:50 in COS 302) COS 561: Advanced Computer Networks Energy.
Data Centers and Cloud Computing 1. 2 Data Centers 3.
A CASE STUDY IN USING MASSIVELY PARALLEL SIMULATION FOR EXTREME- SCALE TORUS NETWORK CO-DESIGN Misbah Mubarak, Rensselaer Polytechnic Institute Christopher.
Fermi National Accelerator Laboratory & Thomas Jefferson National Accelerator Facility SciDAC LQCD Software The Department of Energy (DOE) Office of Science.
Input and Output Optimization in Linux for Appropriate Resource Allocation and Management James Avery King.
Energy Aware Network Operations
Green cloud computing 2 Cs 595 Lecture 15.
Performance Evaluation of Adaptive MPI
Cloud computing Anton Boyko .NET Developer.
ElasticTree: Saving Energy in Data Center Networks
Virtualization.
In-network computation
Towards Predictable Datacenter Networks
Presentation transcript:

Toward Runtime Power Management of Exascale Networks by On/Off Control of Links Ehsan Totoni University of Illinois-Urbana Champaign, PPL Charm++ Workshop, April

Power challenge  Power is a major challenge  Blue Waters consuming up to 13 MW  Enough to electrify a small town  Power and cooling infrastructure  Up to 30% of power in network  Projected for future by Peter Kogge  Saving 25% power in current Cray XT system by turning down network  Work from Sandia 2Ehsan Totoni

Network link power  Network is not “energy proportional”  Consumption is not related to utilization  Near peak most of the time  Unlike processor  Recent study:  Work from Google in ISCA’10  50% of power in network of non-HPC data center  When CPU’s underutilized  Up to 65% of network’s power is in links 3Ehsan Totoni

Exascale networks  Dragonfly  IBM PERCS in Power 775 machines  Cray Aries network in XC30 “Cascade”  DOE Exascale Report  High dimensional Tori  5D Torus in IBM Blue Gen/Q  6D Torus in K Computer  Higher radix -> a lot of links! 4Ehsan Totoni

Communication patterns  Applications’ communication patterns are different  Network topology designed for a wide range of applications MILC NPB CG 5Ehsan Totoni

Fraction of links ever used 6Ehsan Totoni

Nearest neighbor usage 7Ehsan Totoni

More expensive links 8Ehsan Totoni

Nearest neighbor 9Ehsan Totoni

Solution to power waste  Many of the links are never used  For common applications  Are networks over-built? Maybe  FFTs are crucial  But processors are also overbuilt  Let’s make them “energy proportional”  Consume according to workload  Just like processors  Turn off unused links  Commercial network exists (Motorola) 10Ehsan Totoni

Runtime system solution  Hardware can cause delays  According to related work  Not enough application knowledge  Small window size  Compiler does not have enough info  Input dependent program flow  Application does not know hardware  Significant programming burden to expose  Runtime system is the best  mediates all communication  knows the application  knows the hardware 11Ehsan Totoni

Feasibility  Not probably available for your cluster downstairs  Need to convince hardware vendors  Runtime hints to hardware, small delay penalty if wrong  Multiple jobs: interference  Isolated allocations are becoming common  Blue Genes allocate cubes already  Capability machines are for big jobs 12Ehsan Totoni

Software design choices  Random mapping and indirect routing have similar performance but different link usages 13Ehsan Totoni

Power model  We saw many links that are never used  Used links are not used all the time  For only a fraction of iteration time  Compute-communicate paradigm  A power model for “network capacity utilization”  “Average” utilization of all the links  Assume that links are turned magically on and off  At the exact right time  No switching overhead  Example: network used one tenth of iteration time 14Ehsan Totoni

Model results 15Ehsan Totoni

Scheduling on/offs  Runtime roughly knows when a message will arrive  For common iterative HPC applications  Low noise systems (e.g. IBM Blue Genes)  There is a delay for switching the link  10 μ s for current implementation  Much smaller than iteration time  Runtime can be conservative  Schedule “on”s earlier  Similar to having more switching delay 16Ehsan Totoni

Delay overhead 17Ehsan Totoni

Results summary 18Ehsan Totoni

Questions? Are you convinced? 19Ehsan Totoni