Lawrence Livermore National Laboratory BRdeS-1 Science & Technology Principal Directorate - Computation Directorate How to Stop Worrying and Learn to Love.

Slides:



Advertisements
Similar presentations
Lawrence Livermore National Laboratory ROSE Compiler Project Computational Exascale Workshop December 2010 Dan Quinlan Chunhua Liao, Justin Too, Robb Matzke,
Advertisements

Program Analysis and Tuning The German High Performance Computing Centre for Climate and Earth System Research Panagiotis Adamidis.
Parallel computer architecture classification
This work was performed under the auspices of the U.S. Department of Energy by Lawrence Livermore National Laboratory under Contract DE-AC52-07NA27344.
Priority Research Direction Key challenges General Evaluation of current algorithms Evaluation of use of algorithms in Applications Application of “standard”
Performance Metrics Inspired by P. Kent at BES Workshop Run larger: Length, Spatial extent, #Atoms, Weak scaling Run longer: Time steps, Optimizations,
Parallel Research at Illinois Parallel Everywhere
Prof. Srinidhi Varadarajan Director Center for High-End Computing Systems.
Introductory Courses in High Performance Computing at Illinois David Padua.
Click to add text Introduction to the new mainframe: Large-Scale Commercial Computing © Copyright IBM Corp., All rights reserved. Chapter 3: Scalability.
March 18, 2008SSE Meeting 1 Mary Hall Dept. of Computer Science and Information Sciences Institute Multicore Chips and Parallel Programming.
Claude TADONKI Mines ParisTech – LAL / CNRS / INP 2 P 3 University of Oujda (Morocco) – October 7, 2011 High Performance Computing Challenges and Trends.
July Terry Jones, Integrated Computing & Communications Dept Fast-OS.
Bronis R. de Supinski Center for Applied Scientific Computing Lawrence Livermore National Laboratory June 2, 2005 The Most Needed Feature(s) for OpenMP.
SC’07 7 th International APART Workshop Panel 11 November 2007 David Koester, Ph.D.
What Great Research ?s Can RAMP Help Answer? What Are RAMP’s Grand Challenges ?
A Parallel Structured Ecological Model for High End Shared Memory Computers Dali Wang Department of Computer Science, University of Tennessee, Knoxville.
1 New Architectures Need New Languages A triumph of optimism over experience! Ian Watson 3 rd July 2009.
Lecture 1: Introduction to High Performance Computing.
Panel Summary Andrew Hanushevsky Stanford Linear Accelerator Center Stanford University XLDB 23-October-07.
The hybird approach to programming clusters of multi-core architetures.
Leveling the Field for Multicore Open Systems Architectures Markus Levy President, EEMBC President, Multicore Association.
Programming for High Performance Computers John M. Levesque Director Cray’s Supercomputing Center Of Excellence.
Are their more appropriate domain-specific performance metrics for science and engineering HPC applications available then the canonical “percent of peak”
CIS4930/CDA5125 Parallel and Distributed Systems Florida State University CIS4930/CDA5125: Parallel and Distributed Systems Instructor: Xin Yuan, 168 Love,
The WRF Model The Weather Research and Forecasting (WRF) Model is a mesoscale numerical weather prediction system designed for both atmospheric research.
AN EXTENDED OPENMP TARGETING ON THE HYBRID ARCHITECTURE OF SMP-CLUSTER Author : Y. Zhao 、 C. Hu 、 S. Wang 、 S. Zhang Source : Proceedings of the 2nd IASTED.
[Tim Shattuck, 2006][1] Performance / Watt: The New Server Focus Improving Performance / Watt For Modern Processors Tim Shattuck April 19, 2006 From the.
4.2.1 Programming Models Technology drivers – Node count, scale of parallelism within the node – Heterogeneity – Complex memory hierarchies – Failure rates.
Directed Reading 2 Key issues for the future of Software and Hardware for large scale Parallel Computing and the approaches to address these. Submitted.
Alternative ProcessorsHPC User Forum Panel1 HPC User Forum Alternative Processor Panel Results 2008.
LLNL-PRES Lawrence Livermore National Laboratory, P. O. Box 808, Livermore, CA This work performed under the auspices of the U.S. Department.
Application Scaling. Doug Kothe, ORNL Paul Muzio, CUNY Jonathan Carter, LLBL Bronis de Supinski, LLNL Mike Heroux, Sandia Phil Jones, LANL Brent Leback,
A New Parallel Debugger for Franklin: DDT Katie Antypas User Services Group NERSC User Group Meeting September 17, 2007.
Introduction to Research 2011 Introduction to Research 2011 Ashok Srinivasan Florida State University Images from ORNL, IBM, NVIDIA.
MESQUITE: Mesh Optimization Toolkit Brian Miller, LLNL
Motivation: Sorting is among the fundamental problems of computer science. Sorting of different datasets is present in most applications, ranging from.
Lawrence Livermore National Laboratory S&T Principal Directorate - Computation Directorate Tools and Scalable Application Preparation Project Computation.
Workshop on Parallelization of Coupled-Cluster Methods Panel 1: Parallel efficiency An incomplete list of thoughts Bert de Jong High Performance Software.
Experiences with Achieving Portability across Heterogeneous Architectures Lukasz G. Szafaryn +, Todd Gamblin ++, Bronis R. de Supinski ++ and Kevin Skadron.
Bronis R. de Supinski and Jeffrey S. Vetter Center for Applied Scientific Computing August 15, 2000 Umpire: Making MPI Programs Safe.
Interconnection network network interface and a case study.
Exscale – when will it happen? William Kramer National Center for Supercomputing Applications.
Lenovo - Eficiencia Energética en Sistemas de Supercomputación Miguel Terol Palencia Arquitecto HPC LENOVO.
Challenges in the Next Generation Internet Xin Yuan Department of Computer Science Florida State University
Lawrence Livermore National Laboratory 1 Science & Technology Principal Directorate - Computation Directorate Scalable Fault Tolerance for Petascale Systems.
Gengbin Zheng Xiang Ni Laxmikant V. Kale Parallel Programming Lab University of Illinois at Urbana-Champaign.
Other Tools HPC Code Development Tools July 29, 2010 Sue Kelly Sandia is a multiprogram laboratory operated by Sandia Corporation, a.
Tackling I/O Issues 1 David Race 16 March 2010.
Applications Scaling Panel Jonathan Carter (LBNL) Mike Heroux (SNL) Phil Jones (LANL) Kalyan Kumaran (ANL) Piyush Mehrotra (NASA Ames) John Michalakes.
Is MPI still part of the solution ? George Bosilca Innovative Computing Laboratory Electrical Engineering and Computer Science Department University of.
HPC University Requirements Analysis Team Training Analysis Summary Meeting at PSC September Mary Ann Leung, Ph.D.
Hybrid Parallel Implementation of The DG Method Advanced Computing Department/ CAAM 03/03/2016 N. Chaabane, B. Riviere, H. Calandra, M. Sekachev, S. Hamlaoui.
PERFORMANCE OF THE OPENMP AND MPI IMPLEMENTATIONS ON ULTRASPARC SYSTEM Abstract Programmers and developers interested in utilizing parallel programming.
Defining the Competencies for Leadership- Class Computing Education and Training Steven I. Gordon and Judith D. Gardiner August 3, 2010.
Conclusions on CS3014 David Gregg Department of Computer Science
Performance Technology for Scalable Parallel Systems
For Massively Parallel Computation The Chaotic State of the Art
John Levesque Director Cray Supercomputing Center of Excellence
for the Offline and Computing groups
Gengbin Zheng Xiang Ni Laxmikant V. Kale Parallel Programming Lab
Massive Parallelization of SAT Solvers
Alternative Processor Panel Results 2008
Hybrid Programming with OpenMP and MPI
Coordinator: DKRZ Weather Climate HPC
Major Topics in Operating Systems
EE 4xx: Computer Architecture and Performance Programming
3 Questions for Cluster and Grid Use
Department of Computer Science, University of Tennessee, Knoxville
Presentation transcript:

Lawrence Livermore National Laboratory BRdeS-1 Science & Technology Principal Directorate - Computation Directorate How to Stop Worrying and Learn to Love the Systems We Have Today September 10, 2009 Bronis R. de Supinski Center for Applied Scientific Computing Performance Measures x.x, x.x, and x.x

BRdeS-2 Science & Technology Principal Directorate - Computation Directorate Questions 1 and 5: Current metrics are inadequate; Universal metrics are unlikely; Will anyone pay?  Optimization on emerging systems requires new foci and metrics Extremely large core counts Reduced memory and memory bandwidth  Old issues remain but new ones are added Mean time to interrupt growing in importance Memory footprints must decrease Power efficiency as well as FLOP rate  Raw FLOP rates do not characterize application limits  Percent of peak merely translates FLOP rates into system specific number  Parallel efficiency and scalability are helpful but still miss application limits  Time-to-solution; Application work/unit time do not translate across applications Number of Processors (Problem Size) Time to Solution Diag-CG unscalable Multigrid-CG scalable

BRdeS-3 Science & Technology Principal Directorate - Computation Directorate Question 2: We require major software advances to use future systems effectively Debugging 10 3 Cores Load Balance 10 4 Cores Fault Tolerance 10 5 Cores Multicore 10 6 Cores Vector FP Units/ Accelerators? 10 7 Cores Power? 10 8 Cores Purple BG/L Petascale Exascale LLNL is creating essential capabilities for these core counts

BRdeS-4 Science & Technology Principal Directorate - Computation Directorate Questions 3 and 4: A hybrid of current programming models is the only sensible solution  We must make better use of vector floating point units and we can avoid more general accelerators only so long  New hardware can simplify parallelism and expose more  Better support for switching thread contexts  Transactional memory  Thread level speculation Evolution of existing programming models will allow ubiquitous support of these features and more  MPI and OpenMP satiisfy critical programming model requirements  Industry acceptance and participation  More than research funding vehicles  Highly usable as evidenced by user acceptance  I don’t know what the future parallel programming model will be but it will be called MPI