2nd Workshop on Energy for Sustainable Science at Research Infrastructures Report on parallel session A3 Wayne Salter on behalf of Dr. Mike Ashworth (STFC)

Slides:



Advertisements
Similar presentations
Computer Room Requirements for High Density Rack Mounted Servers Rhys Newman Oxford University.
Advertisements

Founded in 2010: UCL, Southampton, Oxford and Bristol Key Objectives of the Consortium: Prove the concept of shared, regional e-infrastructure services.
Accelerators for HPC: Programming Models Accelerators for HPC: StreamIt on GPU High Performance Applications on Heterogeneous Windows Clusters
LEIT (ICT7 + ICT8): Cloud strategy - Cloud R&I: Heterogeneous cloud infrastructures, federated cloud networking; cloud innovation platforms; - PCP for.
Multi-core and tera- scale computing A short overview of benefits and challenges CSC 2007 Andrzej Nowak, CERN
1 Computational models of the physical world Cortical bone Trabecular bone.
©2009 HP Confidential template rev Ed Turkel Manager, WorldWide HPC Marketing 4/7/2011 BUILDING THE GREENEST PRODUCTION SUPERCOMPUTER IN THE.
CHEP 2012 Computing in High Energy and Nuclear Physics Forrest Norrod Vice President and General Manager, Servers.
Device Tradeoffs Greg Stitt ECE Department University of Florida.
Appro Xtreme-X Supercomputers A P P R O I N T E R N A T I O N A L I N C.
2. Computer Clusters for Scalable Parallel Computing
CURRENT AND FUTURE HPC SOLUTIONS. T-PLATFORMS  Russia’s leading developer of turn-key solutions for supercomputing  Privately owned  140+ employees.
Zhao Lixing.  A supercomputer is a computer that is at the frontline of current processing capacity, particularly speed of calculation.  Supercomputers.
This project and the research leading to these results has received funding from the European Community's Seventh Framework Programme.
GPU System Architecture Alan Gray EPCC The University of Edinburgh.
Prof. Srinidhi Varadarajan Director Center for High-End Computing Systems.
Cloud Computing Data Centers Dr. Sanjay P. Ahuja, Ph.D FIS Distinguished Professor of Computer Science School of Computing, UNF.
Some Thoughts on Technology and Strategies for Petaflops.
CalStan 3/2011 VIRAM-1 Floorplan – Tapeout June 01 Microprocessor –256-bit media processor –12-14 MBytes DRAM – Gops –2W at MHz –Industrial.
Lecture 1: Introduction to High Performance Computing.
Heterogeneous Computing Dr. Jason D. Bakos. Heterogeneous Computing 2 “Traditional” Parallel/Multi-Processing Large-scale parallel platforms: –Individual.
Introduction to Reconfigurable Computing Greg Stitt ECE Department University of Florida.
1 Copyright © 2012, Elsevier Inc. All rights reserved. Chapter 1 Fundamentals of Quantitative Design and Analysis Computer Architecture A Quantitative.
GPU Programming with CUDA – Accelerated Architectures Mike Griffiths
Chapter 2 Computer Clusters Lecture 2.3 GPU Clusters for Massive Paralelism.
1J. Kim Web Science & Technology Forum Enabling Hardware Technology for Web Science John Kim Department of Computer Science KAIST.
Copyright 2009 Fujitsu America, Inc. 0 Fujitsu PRIMERGY Servers “Next Generation HPC and Cloud Architecture” PRIMERGY CX1000 Tom Donnelly April
Sensor-Based Fast Thermal Evaluation Model For Energy Efficient High-Performance Datacenters Q. Tang, T. Mukherjee, Sandeep K. S. Gupta Department of Computer.
N. GSU Slide 1 Chapter 02 Cloud Computing Systems N. Xiong Georgia State University.
Last Time Performance Analysis It’s all relative
Sogang University Advanced Computing System Chap 1. Computer Architecture Hyuk-Jun Lee, PhD Dept. of Computer Science and Engineering Sogang University.
Publication: Ra Inta, David J. Bowman, and Susan M. Scott. Int. J. Reconfig. Comput. 2012, Article 2 (January 2012), 1 pages. DOI= /2012/ Naveen.
Are Supercomputers returning to Investment Banking?
March 9, 2015 San Jose Compute Engineering Workshop.
Cray Innovation Barry Bolding, Ph.D. Director of Product Marketing, Cray September 2008.
October 12, 2004Thomas Sterling - Caltech & JPL 1 Roadmap and Change How Much and How Fast Thomas Sterling California Institute of Technology and NASA.
A lower bound to energy consumption of an exascale computer Luděk Kučera Charles University Prague, Czech Republic.
Next Generation Operating Systems Zeljko Susnjar, Cisco CTG June 2015.
Introduction to Research 2011 Introduction to Research 2011 Ashok Srinivasan Florida State University Images from ORNL, IBM, NVIDIA.
© 2009 IBM Corporation Motivation for HPC Innovation in the Coming Decade Dave Turek VP Deep Computing, IBM.
© Copyright 2012 Hewlett-Packard Development Company, L.P. The information contained herein is subject to change without notice. HP Update IDC HPC Forum.
Patryk Lasoń, Marek Magryś
High Performance Computing
Power and Cooling at Texas Advanced Computing Center Tommy Minyard, Ph.D. Director of Advanced Computing Systems 42 nd HPC User Forum September 8, 2011.
Lenovo - Eficiencia Energética en Sistemas de Supercomputación Miguel Terol Palencia Arquitecto HPC LENOVO.
Energy Savings in CERN’s Main Data Centre
Tackling I/O Issues 1 David Race 16 March 2010.
CERN VISIONS LEP  web LHC  grid-cloud HL-LHC/FCC  ?? Proposal: von-Neumann  NON-Neumann Table 1: Nick Tredennick’s Paradigm Classification Scheme Early.
CERN - IT Department CH-1211 Genève 23 Switzerland t Power and Cooling Challenges at CERN IHEPCCC Meeting April 24 th 2007 Tony Cass.
Parallel Computers Today Oak Ridge / Cray Jaguar > 1.75 PFLOPS Two Nvidia 8800 GPUs > 1 TFLOPS Intel 80- core chip > 1 TFLOPS  TFLOPS = floating.
All content in this presentation is protected – © 2008 American Power Conversion Corporation Row Cooling.
BLUE GENE Sunitha M. Jenarius. What is Blue Gene A massively parallel supercomputer using tens of thousands of embedded PowerPC processors supporting.
Societal applications of large scalable parallel computing systems ARTEMIS & ITEA Co-summit, Madrid, October 30th 2009.
Profiling and Characterising Software Application Energy Consumption Ian Osborne ICT KTN and EEC SIG.
Earth System Modelling: an HPC perspective Mike Ashworth & Rupert Ford Scientific Computing Department and STFC Hartree Centre STFC Daresbury Laboratory.
Why Parallel/Distributed Computing Sushil K. Prasad
Extreme Scale Infrastructure
INFN Computing Outlook The Bologna Initiative
Appro Xtreme-X Supercomputers
OCP: High Performance Computing Project
Clustered Systems Introduction
Low Latency Analytics HPC Clusters
What is Parallel and Distributed computing?
Barcelona Supercomputing Center
Course Description: Parallel Computer Architecture
Chapter 1 Introduction.
Exascale Programming Models in an Era of Big Computation and Big Data
The University of Adelaide, School of Computer Science
Utsunomiya University
Presentation transcript:

2nd Workshop on Energy for Sustainable Science at Research Infrastructures Report on parallel session A3 Wayne Salter on behalf of Dr. Mike Ashworth (STFC)

Talks A Comprehensive Approach to Energy Efficiency in Data Centers for High-Performance Computing by Prof. Thomas C. Schulthess (CSCS) Exploiting mobile phone technology to build energy efficient supercomputers: the Mont Blanc project by Dr. Simon McIntosh-Smith (University of Bristol) Roadmap towards Ultimately-Efficient Datacenters by Dr. Bruno Michel (IBM) Energy Savings in CERN’s main Data Centre by Wayne Salter (CERN)

Summary - I First three talks followed a common theme – Computing needs growing rapidly – As a result the power needed for computing growing very quickly too – Not sustainable – Need major change in technology and/or use Fourth talk – Discussed some concrete measures taken to improve the energy efficiency of an old existing DC

A Comprehensive Approach to Energy Efficiency in Data Centers for High-Performance Computing - I Discussed investments in Switzerland for HPC One of the goals of HP2C program was to push innovation in algorithm and application software design to take advantage better of capabilities of modern HPC H/W – Massive concurrency (multithreading and high node count) – Hybrid nodes (CPU+GPU) Three pronged approach: – Efficient DC (new DC built using cooling from lake already has PUE of 1.2 and this is likely to improve with more loading) – Efficient computers (designed specifically for the applications) – Efficient applications Most energy in computers today is moving data and not compute New system (Cray XC30 Piz Daint) – Next generation network interconnect developed through DARPA HPCS program – Hybrid nodes with Intel CPU and NVIDIA GPU Discussed improvements based on COSMO weather predication application

A Comprehensive Approach to Energy Efficiency in Data Centers for High-Performance Computing - II Main points: – Use of free cooling using lake water has resulted in an efficient data centre at CSCS with a PUE of ~ 1.2 for 2-3 MW – The dynamical core of the COSMO weather forecast code used by Meteo Swiss has been adapted to exploit GPU hardware – An improvement in efficiency of 10x has been achieved for the COSMO code combining 1.5x from the building, 1.75x from the new system, 1.49x from the new code and 2.64x from use of hybrid nodes – Future power efficiency improvements are more likely to come from applications development than from hardware

Exploiting mobile phone technology to build energy efficient supercomputers: the Mont Blanc - I HPC systems growing in performance but also power – Average power consumption of top 10 in 2008 was 1.5MW and now 6MW (5x increase in 5 years) – Future limiting factor not necessary the delivery of power but rather its cost – Europe is major HPC player but has no technology of its own – However, is strong in embedded computing How to build an affordable Exoflop machine? – Need revolutionary rather than evolutionary approach Mont Blanc project – European project led by Barcelona Supercomputing centre – Leverage commodity and embedded power-efficient technology from mobile market – Proof of concept has been made with a compute card containing Samsung CPU, GPU, DRAM and NAND memory, and NIC. – 15 cards/blade, 9 blades per chassis and 4 blade chassis in a rack – Delivers 17.2 TFLOPS for 8.2 kW – Ambitious development roadmap

Exploiting mobile phone technology to build energy efficient supercomputers: the Mont Blanc - II Main points: – Following the historical replacement of vector processors by commodity microprocessors – “killer micros”, there may be a similar coup by the “killer mobiles” – Europe is well placed with strengths in embedded computing for mobile devices – The Mont Blanc project aims to produce an Exascale system based on ARM microprocessors with a 200 Pflop/s system consuming 10 MW in 2017 and an Exaflop system at 20 MW around 2020

Ro admap towards Ultimately-Efficient Datacenters - I We need three paradigm changes: – Moving from cold air cooling to hot water energy re-use – Analyse the system in term of efficiency not performance – From areal device size scaling to volumetric density scaling – build in 3D not in 2D - vertical integration Hot water cooling with waste heat re-use – SuperMUC I prototype at ETH Zurich iDataPlex cluster with 3.2PFLOPS (20k CPUs/160k cores) 4MW, PUE 1.15, 90% heat for re-use =>40% less energy consumption Analysing systems in terms of compute efficiency and density – Shows that we are still 4 orders of magnitude worse than a human brain => use as an example – Transistors occupy only 1ppm of system volume – Majority of energy used for communication (dependent on wire length and scales quadratically) => need to look at volumetric scaling – Some ideas presented on how to move from 2D to high density 3D chip design with interlayer cooling and electrochemical chip powering – Aim to develop 1 PFLOPS machine in 10 litres Can also learn from allometric scaling in biology

Roadmap towards Ultimately-Efficient Datacenters - II Main points: – We should move from cold air cooling to hot water energy re-use – We must analyse systems in term of efficiency not performance – Vertical integration will enable dense architectures which improve efficiency through chip stacking and interlayer cooling – Moore’s Law goes 3D – Such an ultra-dense 3D system will achieve 1 Pflop/s in 10 litres

Energy Savings in CERN’s main Data Centre Move from low power density mainframes to rack mounted servers led to cooling issues – Solved by cold/hot aisle separation in 2008 Further improvements made to improve efficiency – Modification of air handling to increase substantially the use of free cooling => very low requirement for chillers – More aggressive temperature environment – Savings of > 6.2 GWh for a 2.6 MW DC Achieved with relative simple cost-effective measures Further measures foreseen