Computer and Computational Sciences Division Los Alamos National Laboratory Ideas that change the world Achieving Usability and Efficiency in Large-Scale.

Slides:



Advertisements
Similar presentations
Hadi Salimi Distributed Systems Labaratory, School of Computer Engineering, Iran University of Science and Technology, Fall
Advertisements

Priority Research Direction (I/O Models, Abstractions and Software) Key challenges What will you do to address the challenges? – Develop newer I/O models.
High Performance Computing Course Notes Grid Computing.
Distributed Systems 1 Topics  What is a Distributed System?  Why Distributed Systems?  Examples of Distributed Systems  Distributed System Requirements.
July Terry Jones, Integrated Computing & Communications Dept Fast-OS.
1 BGL Photo (system) BlueGene/L IBM Journal of Research and Development, Vol. 49, No. 2-3.
Kei Davis and Fabrizio Petrini Europar 2004, Pisa Italy 1 CCS-3 P AL A CASE STUDY.
New Challenges in Cloud Datacenter Monitoring and Management
Computer System Architectures Computer System Software
Tanenbaum & Van Steen, Distributed Systems: Principles and Paradigms, 2e, (c) 2007 Prentice-Hall, Inc. All rights reserved DISTRIBUTED SYSTEMS.
Lecture 3 – Parallel Performance Theory - 1 Parallel Performance Theory - 1 Parallel Computing CIS 410/510 Department of Computer and Information Science.
Chapter 3: Operating-System Structures System Components Operating System Services System Calls System Programs System Structure Virtual Machines System.
Cloud Computing 1. Outline  Introduction  Evolution  Cloud architecture  Map reduce operation  Platform 2.
Cluster Reliability Project ISIS Vanderbilt University.
Extreme scale parallel and distributed systems – High performance computing systems Current No. 1 supercomputer Tianhe-2 at petaflops Pushing toward.
Extreme-scale computing systems – High performance computing systems Current No. 1 supercomputer Tianhe-2 at petaflops Pushing toward exa-scale computing.
Presented by Reliability, Availability, and Serviceability (RAS) for High-Performance Computing Stephen L. Scott and Christian Engelmann Computer Science.
Loosely Coupled Parallelism: Clusters. Context We have studied older archictures for loosely coupled parallelism, such as mesh’s, hypercubes etc, which.
4.2.1 Programming Models Technology drivers – Node count, scale of parallelism within the node – Heterogeneity – Complex memory hierarchies – Failure rates.
Directed Reading 2 Key issues for the future of Software and Hardware for large scale Parallel Computing and the approaches to address these. Submitted.
An Approach To Automate a Process of Detecting Unauthorised Accesses M. Chmielewski, A. Gowdiak, N. Meyer, T. Ostwald, M. Stroiński
Workshop on the Future of Scientific Workflows Break Out #2: Workflow System Design Moderators Chris Carothers (RPI), Doug Thain (ND)
ESIP Federation Air Quality Cluster Partner Agencies.
Advanced Computer Networks Topic 2: Characterization of Distributed Systems.
Cray Innovation Barry Bolding, Ph.D. Director of Product Marketing, Cray September 2008.
Advanced Principles of Operating Systems (CE-403).
Issues Autonomic operation (fault tolerance) Minimize interference to applications Hardware support for new operating systems Resource management (global.
NIH Resource for Biomolecular Modeling and Bioinformatics Beckman Institute, UIUC NAMD Development Goals L.V. (Sanjay) Kale Professor.
PARALLEL COMPUTING overview What is Parallel Computing? Traditionally, software has been written for serial computation: To be run on a single computer.
Authors: Ronnie Julio Cole David
Computer and Computational Sciences Division Los Alamos National Laboratory On the Feasibility of Incremental Checkpointing for Scientific Computing Jose.
GRID Overview Internet2 Member Meeting Spring 2003 Sandra Redman Information Technology and Systems Center and Information Technology Research Center National.
Diskless Checkpointing on Super-scale Architectures Applied to the Fast Fourier Transform Christian Engelmann, Al Geist Oak Ridge National Laboratory Februrary,
Ruth Pordes November 2004TeraGrid GIG Site Review1 TeraGrid and Open Science Grid Ruth Pordes, Fermilab representing the Open Science.
U N I V E R S I T Y O F S O U T H F L O R I D A Hadoop Alternative The Hadoop Alternative Larry Moore 1, Zach Fadika 2, Dr. Madhusudhan Govindaraju 2 1.
1 THE EARTH SIMULATOR SYSTEM By: Shinichi HABATA, Mitsuo YOKOKAWA, Shigemune KITAWAKI Presented by: Anisha Thonour.
Workshop on Parallelization of Coupled-Cluster Methods Panel 1: Parallel efficiency An incomplete list of thoughts Bert de Jong High Performance Software.
Barriers to Industry HPC Use or “Blue Collar” HPC as a Solution Presented by Stan Ahalt OSC Executive Director Presented to HPC Users Conference July 13,
7. Grid Computing Systems and Resource Management
+ Clusters Alternative to SMP as an approach to providing high performance and high availability Particularly attractive for server applications Defined.
COMP381 by M. Hamdi 1 Clusters: Networks of WS/PC.
Hadoop/MapReduce Computing Paradigm 1 CS525: Special Topics in DBs Large-Scale Data Management Presented By Kelly Technologies
DR. SIMING LIU SPRING 2016 COMPUTER SCIENCE AND ENGINEERING UNIVERSITY OF NEVADA, RENO Session 2 Computer Organization.
3/12/2013Computer Engg, IIT(BHU)1 PARALLEL COMPUTERS- 1.
Lawrence Livermore National Laboratory 1 Science & Technology Principal Directorate - Computation Directorate Scalable Fault Tolerance for Petascale Systems.
3/12/2013Computer Engg, IIT(BHU)1 PARALLEL COMPUTERS- 3.
CS4315A. Berrached:CMS:UHD1 Introduction to Operating Systems Chapter 1.
Tackling I/O Issues 1 David Race 16 March 2010.
CDA-5155 Computer Architecture Principles Fall 2000 Multiprocessor Architectures.
Background Computer System Architectures Computer System Software.
Presented by Fault Tolerance Challenges and Solutions Al Geist Network and Cluster Computing Computational Sciences and Mathematics Division Research supported.
PARALLEL AND DISTRIBUTED PROGRAMMING MODELS U. Jhashuva 1 Asst. Prof Dept. of CSE om.
BLUE GENE Sunitha M. Jenarius. What is Blue Gene A massively parallel supercomputer using tens of thousands of embedded PowerPC processors supporting.
Presented by Robust Storage Management On Desktop, in Machine Room, and Beyond Xiaosong Ma Computer Science and Mathematics Oak Ridge National Laboratory.
Lecture 13 Parallel Processing. 2 What is Parallel Computing? Traditionally software has been written for serial computation. Parallel computing is the.
COMP7330/7336 Advanced Parallel and Distributed Computing MapReduce - Introduction Dr. Xiao Qin Auburn University
INTRODUCTION TO HIGH PERFORMANCE COMPUTING AND TERMINOLOGY.
Computer Organization and Architecture Lecture 1 : Introduction
CompSci 280 S Introduction to Software Development
Hadoop Aakash Kag What Why How 1.
For Massively Parallel Computation The Chaotic State of the Art
Fault-Tolerant NoC-based Manycore system: Reconfiguration & Scheduling
Grid Computing.
CSC 480 Software Engineering
Architecture & Organization 1
Architecture & Organization 1
with Computational Scientists
Subject Name: Operating System Concepts Subject Number:
LO2 – Understand Computer Software
Presentation transcript:

Computer and Computational Sciences Division Los Alamos National Laboratory Ideas that change the world Achieving Usability and Efficiency in Large-Scale Parallel Computing Systems Kei Davis and Fabrizio Petrini Performance and Architectures Lab (PAL), CCS-3

Kei Davis and Fabrizio Petrini Europar 2004, Pisa Italy 2 CCS-3 P AL Schedule n Introduction n Break n Existing Systems n Break n Case Study n Break n A New Approach

Kei Davis and Fabrizio Petrini Europar 2004, Pisa Italy 3 CCS-3 P AL Part 1: Introduction 1. The need for more capability 2. The big issues 3. A taxonomy of systems in three dimensions

Kei Davis and Fabrizio Petrini Europar 2004, Pisa Italy 4 CCS-3 P AL The Need for More Capability The most constant difficulty in contriving the engine has arisen from the desire to reduce the time in which the calculations were executed to the shortest which is possible. Charles Babbage, Our interest is in scientific computing—large-scale, numerical, parallel applications run on large-scale parallel machines.

Kei Davis and Fabrizio Petrini Europar 2004, Pisa Italy 5 CCS-3 P AL Definitions n Computing capacity: total deliverable computing power from a system or set of systems. (Power— rate of delivery) n Computing capability: computing power available to a single application. Highest-end computing is primarily concerned with capability—why else build such machines?

Kei Davis and Fabrizio Petrini Europar 2004, Pisa Italy 6 CCS-3 P AL The Need for Large-Scale Parallel Machines n It is the insatiable demand for ever more computational capability that has driven the creation of many Tflop-scale parallel machines (Earth Simulator, LANL’s ASCI Q, LLNL’s Thunder and BlueGene/L, etc.) n Petaflop machines are on the horizon, for example DARPA HPCS program (High Productivity Computing Systems)

Kei Davis and Fabrizio Petrini Europar 2004, Pisa Italy 7 CCS-3 P AL One-upmanship? Is this merely one-upmanship with the Japanese? From The Roadmap for the Revitalization of High-End Computing, Computing Research Association: […] there is a growing recognition that a new set of scientific and engineering discoveries could be catalyzed by access to very-large-scale computer systems—those in the 100 teraflop to petaflop range.

Kei Davis and Fabrizio Petrini Europar 2004, Pisa Italy 8 CCS-3 P AL Requirements for ASCI In our own arena, Advanced Simulation and Computing (ASC) for stockpile stewardship; climate, ocean, and urban infrastructure modeling, etc., Within 10 years, estimates of the demand for Capability and general physics arguments indicate a machine of 1000TF=1 PetaFlop (PF) will be needed to execute the most demanding jobs. Such demand is inevitable; it should not be viewed, however, as some plateau in required Capability: there are sound technical reasons to expect even greater Capability demand in the future.

Kei Davis and Fabrizio Petrini Europar 2004, Pisa Italy 9 CCS-3 P AL Large Component Count n Increases in performance will be achieved through single processor improvements and increases in component count n For example, BlueGene/L will have 133,120 processors and 608,256 memory modules n The large component count will make any assumption of complete reliability unrealistic 133,120 processors 608,256 DRAM

Kei Davis and Fabrizio Petrini Europar 2004, Pisa Italy 10 CCS-3 P AL Sensitivity to Failures n In a large-scale machine a failure of a single component usually causes a significant fraction of the system to fail because 1. Components are strongly coupled (e.g., a failure of a fan will lead to other failures due to overheating) 2. The state of the application is not stored redundantly, and loss of any state is catastrophic 3. In capability mode, many processing nodes are running the same application, and are tightly coupled together

Kei Davis and Fabrizio Petrini Europar 2004, Pisa Italy 11 CCS-3 P AL The Need for Transparent Fault-Tolerance n System software must be resilient to failures, to allow continuing execution of in the presence of failures n Most of the investment is in the application software (250M$/year for MPI software in the ASCI TriLabs) n Economical constraints impose a limited level of redundancy n Other considerations include cost of development, scalability and efficiency

Kei Davis and Fabrizio Petrini Europar 2004, Pisa Italy 12 CCS-3 P AL The JASON’s Report n A recent report from the JASON’s, a committee of distinguished scientists chartered by the US government, raised the sensitive question of whether ASCI machines can be used as capability engines n For that to be possible, major advances in fault- tolerance are needed n The recommendation of the report is to skip one generation of supercomputers, due to the lack of good technical/scientific solutions

Kei Davis and Fabrizio Petrini Europar 2004, Pisa Italy 13 CCS-3 P AL MTBF as a Function of System Size

Kei Davis and Fabrizio Petrini Europar 2004, Pisa Italy 14 CCS-3 P AL Failure Distribution (ASCI Blue Mountain)

Kei Davis and Fabrizio Petrini Europar 2004, Pisa Italy 15 CCS-3 P AL State of the Art in Large- Scale Supercomputers n We can assemble large-scale systems by wiring together hardware and “bolting together” software components n But we have almost no control on the machine: not only faults but also performance anomalies

Kei Davis and Fabrizio Petrini Europar 2004, Pisa Italy 16 CCS-3 P AL 1.2 The Big Issues From DoE Office of Science By the end of this decade petascale computers with thousands of times more computational power than any in current use will be vital tools for expanding the frontiers of science and for addressing vital national priorities. These systems will have tens to hundreds of thousands of processors, an unprecedented level of complexity, and will require significant new levels of scalability and fault management. [Emphasis added]

Kei Davis and Fabrizio Petrini Europar 2004, Pisa Italy 17 CCS-3 P AL Office of Science cont’d Current and future large-scale parallel systems require that such services be implemented in a fast and scalable manner so that the OS/R does not become a performance bottleneck. Without reliable, robust operating systems and runtime environments the computational science research community will be unable to easily and completely employ future generations of extreme-scale systems for scientific discovery.

Kei Davis and Fabrizio Petrini Europar 2004, Pisa Italy 18 CCS-3 P AL DARPA Defence Advanced Research Projects Administration (DARPA) High Productivity Computing Systems (HPCS) mission: Provide economically viable high productivity systems for the national security and industrial user communities with the following design attributes in the latter part of this decade: Performance Programmability Portability Robustness

Kei Davis and Fabrizio Petrini Europar 2004, Pisa Italy 19 CCS-3 P AL Our Translation n Performance—achieving achievable performance (not, e.g., some percentage of theoretical peak) n Programmability/portability—standard interfaces, transparency of mechanisms for fault tolerance n Robustness—graceful failover

Kei Davis and Fabrizio Petrini Europar 2004, Pisa Italy 20 CCS-3 P AL 1.3 A Taxonomy of Systems Q: Is it a supercomputer or just a cluster? A: It is a continuum along multiple dimensions. A taxonomy of systems of three dimensions: n Degree of integration of compute node; n Collective primitives provided by the network interface, programmability, global address space; n Degree of integration of system software.

Kei Davis and Fabrizio Petrini Europar 2004, Pisa Italy 21 CCS-3 P AL Note This taxonomy is useful for our explication, but we make no claims that it n is canonical, n that it captures highly specialized architectures (for example custom-designed special-purpose digital processors, vector processors, floating-point processors). We are concerned with the big `general purpose’ parallel machines.

Kei Davis and Fabrizio Petrini Europar 2004, Pisa Italy 22 CCS-3 P AL Compute Node Degree of integration of compute node between processors, memory, and network interface n Single processor—SMP—multiple CPU cores per chip n Number of levels of cache, proximity of caches to CPU core n Proximity of network interface to CPU core: on- chip—off-chip direct connection—separated by I/O interface

Kei Davis and Fabrizio Petrini Europar 2004, Pisa Italy 23 CCS-3 P AL Network Interface n Collective primitives provided by network interface: none—functionally rich; n Programmability of network interface: none— general purpose n Provision of virtual global address space

Kei Davis and Fabrizio Petrini Europar 2004, Pisa Italy 24 CCS-3 P AL System software n Degree of integration of system software much more about this later…