HPC Benchmarking, Rudi Eigenmann, 2006 SPEC Benchmark Workshop this is important for machine procurements and for understanding where HPC technology is.

Slides:



Advertisements
Similar presentations
SPEC High Performance Group (An Introduction)
Advertisements

SPEC ENV2002: Environmental Simulation Benchmarks Wesley Jones
SPECs High-Performance Benchmark Suite, SPEC HPC Rudi Eigenmann Purdue University Indiana, USA.
SPEC OMP Benchmark Suite H. Saito, G. Gaertner, W. Jones, R. Eigenmann, H. Iwashita, R. Lieberman, M. van Waveren, and B. Whitney SPEC High-Performance.
Höchstleistungsrechenzentrum Stuttgart Matthias M üller SPEC Benchmarks for Large Systems Matthias Mueller, Kumaran Kalyanasundaram, G. Gaertner, W. Jones,
SPEC HPG Benchmarks for HPC Systems Kumaran Kalyanasundaram for SPEC High-Performance Group Kumaran Kalyanasundaram, PhD Chair, SPEC HPG Manager, SGI Performace.
Höchstleistungsrechenzentrum Stuttgart Matthias M üller SPEC Benchmarks at HLRS Matthias Mueller High Performance Computing Center Stuttgart
Refining High Performance FORTRAN Code from Programming Model Dependencies Ferosh Jacob University of Alabama Department of Computer Science
Parallel Processing with OpenMP
On the Interaction of Tiling and Automatic Parallelization Zhelong Pan, Brian Armstrong, Hansang Bae Rudolf Eigenmann Purdue University, ECE
Overview Motivations Basic static and dynamic optimization methods ADAPT Dynamo.
The Who, What, Why and How of High Performance Computing Applications in the Cloud Soheila Abrishami 1.
Chapter 4 M. Keshtgary Spring 91 Type of Workloads.
Senior Design Project: Parallel Task Scheduling in Heterogeneous Computing Environments Senior Design Students: Christopher Blandin and Dylan Machovec.
Scheduling Considerations for building Dynamic Verification Tools for MPI Sarvani Vakkalanka, Michael DeLisi Ganesh Gopalakrishnan, Robert M. Kirby School.
Introduction CS 524 – High-Performance Computing.
1 Cetus – An Extensible Compiler Infrastructure Sang Ik Lee Troy Johnson Rudolf Eigenmann ECE, Purdue University.
NPACI: National Partnership for Advanced Computational Infrastructure Supercomputing ‘98 Mannheim CRAY T90 vs. Tera MTA: The Old Champ Faces a New Challenger.
On the Integration and Use of OpenMP Performance Tools in the SPEC OMP2001 Benchmarks Bernd Mohr 1, Allen D. Malony 2, Rudi Eigenmann 3 1 Forschungszentrum.
1/28/2004CSCI 315 Operating Systems Design1 Operating System Structures & Processes Notice: The slides for this lecture have been largely based on those.
Parallelizing Compilers Presented by Yiwei Zhang.
Tile Reduction: the first step towards tile aware parallelization in OpenMP Ge Gan Department of Electrical and Computer Engineering Univ. of Delaware.
Why Threads Are A Bad Idea (for most purposes) John Ousterhout Sun Microsystems Laboratories
Genetic Algorithms: A Tutorial
1 Reasons for parallelization Can we make GA faster? One of the most promising choices is to use parallel implementations. The reasons for parallelization.
Virtualization Concept. Virtualization  Real: it exists, you can see it.  Transparent: it exists, you cannot see it  Virtual: it does not exist, you.
STRATEGIES INVOLVED IN REMOTE COMPUTATION
Budapest, November st ALADIN maintenance and phasing workshop Short introduction to OpenMP Jure Jerman, Environmental Agency of Slovenia.
© Fujitsu Laboratories of Europe 2009 HPC and Chaste: Towards Real-Time Simulation 24 March
Statistical Performance Analysis for Scientific Applications Presentation at the XSEDE14 Conference Atlanta, GA Fei Xing Haihang You Charng-Da Lu July.
Predicting performance of applications and infrastructures Tania Lorido 27th May 2011.
ICOM 5995: Performance Instrumentation and Visualization for High Performance Computer Systems Lecture 7 October 16, 2002 Nayda G. Santiago.
National Center for Supercomputing Applications University of Illinois at Urbana-Champaign Cell processor implementation of a MILC lattice QCD application.
Twelve Ways to Fool the Masses When Giving Performance Results on Parallel Supercomputers – David Bailey (1991) Eileen Kraemer August 25, 2002.
A Framework for Elastic Execution of Existing MPI Programs Aarthi Raveendran Tekin Bicer Gagan Agrawal 1.
S AN D IEGO S UPERCOMPUTER C ENTER N ATIONAL P ARTNERSHIP FOR A DVANCED C OMPUTATIONAL I NFRASTRUCTURE On pearls and perils of hybrid OpenMP/MPI programming.
A Framework for Elastic Execution of Existing MPI Programs Aarthi Raveendran Graduate Student Department Of CSE 1.
System Software for Parallel Computing. Two System Software Components Hard to do the innovation Replacement for Tradition Optimizing Compilers Replacement.
A Model for Computational Science Investigations Supercomputing Challenge
From lecture slides for Computer Organization and Architecture: Designing for Performance, Eighth Edition, Prentice Hall, 2010 CS 211: Computer Architecture.
HPCMP Benchmarking Update Cray Henry April 2008 Department of Defense High Performance Computing Modernization Program.
Computer and Computational Sciences Division Los Alamos National Laboratory On the Feasibility of Incremental Checkpointing for Scientific Computing Jose.
Nanco: a large HPC cluster for RBNI (Russell Berrie Nanotechnology Institute) Anne Weill – Zrahia Technion,Computer Center October 2008.
August 2001 Parallelizing ROMS for Distributed Memory Machines using the Scalable Modeling System (SMS) Dan Schaffer NOAA Forecast Systems Laboratory (FSL)
Weather Research & Forecasting Model Xabriel J Collazo-Mojica Alex Orta Michael McFail Javier Figueroa.
Motivation: Sorting is among the fundamental problems of computer science. Sorting of different datasets is present in most applications, ranging from.
Theme 3: Science of Science Policy to Address National Needs Summary December 4, 2008.
Data Structures and Algorithms in Parallel Computing Lecture 7.
6.1 Advanced Operating Systems Lies, Damn Lies and Benchmarks Are your benchmark tests reliable?
A Simulation Framework to Automatically Analyze the Communication-Computation Overlap in Scientific Applications Vladimir Subotic, Jose Carlos Sancho,
High-level Interfaces for Scalable Data Mining Ruoming Jin Gagan Agrawal Department of Computer and Information Sciences Ohio State University.
edit type on title master Fortran ISV Release I to L LINPACK TOP500 Technical Systems Division * Scalable Computing Lab 2 Hsin-Ying Lin
An Evaluation of Data-Parallel Compiler Support for Line-Sweep Applications Daniel Chavarría-Miranda John Mellor-Crummey Dept. of Computer Science Rice.
Slide 1 NEMOVAR-LEFE Workshop 22/ Slide 1 Current status of NEMOVAR Kristian Mogensen.
1 HPJAVA I.K.UJJWAL 07M11A1217 Dept. of Information Technology B.S.I.T.
Whole Test Suite Generation. Abstract Not all bugs lead to program crashes, and not always is there a formal specification to check the correctness of.
ECE 259 / CPS 221 Advanced Computer Architecture II (Parallel Computer Architecture) Evaluation – Metrics, Simulation, and Workloads Copyright 2004 Daniel.
Michael J. Voss and Rudolf Eigenmann PPoPP, ‘01 (Presented by Kanad Sinha)
Societal applications of large scalable parallel computing systems ARTEMIS & ITEA Co-summit, Madrid, October 30th 2009.
Pitfalls: Time Dependent Behaviors CS433 Spring 2001 Laxmikant Kale.
Parallel Programming Models
4- Performance Analysis of Parallel Programs
CSE 311 Foundations of Computing I
CSCI1600: Embedded and Real Time Software
Software Acceleration in Hybrid Systems Xiaoqiao (XQ) Meng IBM T. J
Why Threads Are A Bad Idea (for most purposes)
Why Threads Are A Bad Idea (for most purposes)
Why Threads Are A Bad Idea (for most purposes)
CSCI1600: Embedded and Real Time Software
Presentation transcript:

HPC Benchmarking, Rudi Eigenmann, 2006 SPEC Benchmark Workshop this is important for machine procurements and for understanding where HPC technology is heading HPC Benchmarking and Performance Evaluation With Realistic Applications Brian Armstrong, Hansang Bae, Rudolf Eigenmann, Faisal Saied, Mohamed, Sayeed, Yili Zheng Purdue University Benchmarking has two important goals 1. Assess the performance of High-Performance Computer Platforms 2. Measure and show opportunities for progress in HPC this is important to quantify and compare scientific research contributions and to set new directions for research

HPC Benchmarking, Rudi Eigenmann, 2006 SPEC Benchmark Workshop Why Talk About Benchmarking? There is no progress if you cant measure it use benchmark applications unknown to others; give no reference use applications that have the same name as known benchmarks, but that show better performance of your innovation use only those benchmarks out of a suite that show good performance on your novel technique use only the benchmarks out of the suite that dont break your technique modify the benchmark source code change data set parameters use the debug data set use a few loops out of the full programs only measure loop performance but label it as the full application dont mention in your paper why you have chosen the benchmarks in this way and what changes you have made time the interesting part of the program only; exclude overheads measure interesting overheads only, exclude large unwanted items 12 ways to fool the scientist (with computer performance evaluation)

HPC Benchmarking, Rudi Eigenmann, 2006 SPEC Benchmark Workshop Benchmarks Need to be Representative and Open Representative Benchmarks: –Represent real problems Open Benchmarks: –No proprietary strings attached –Source code and performance data can be freely distributed With these goals in mind, SPECs High-performance group was formed in 1994

HPC Benchmarking, Rudi Eigenmann, 2006 SPEC Benchmark Workshop Why is Benchmarking with Real Application Hard? Simple Benchmarks are Overly Easy to Run Realistic Benchmarks Cannot be Abstracted from Real Applications Today's Realistic Applications May Not be Tomorrows Applications Benchmarking is not Eligible for Research Funding Maintaining Benchmarking Efforts is Costly Proprietary Full-Application Benchmarks Cannot Serve as Yardsticks

HPC Benchmarking, Rudi Eigenmann, 2006 SPEC Benchmark Workshop SPEC HPC2002 Includes three (suites of) codes –SPECchem, used in chemical and pharmaceutical industries (gamess) 110,000 lines of Fortran and C –SPECenv, weather forecast application (WRF) 160,000 lines of Fortran and C –SPECseis, used in the search for oil and gas 20,000 lines of Fortran and C All codes include several data sets and are available in a serial and a parallel variant ( MPI, OpenMP, hybrid execution is possible ). SPEC HPC used in TAP list - Top Application Performers –the rank list of HPC systems based on realistic applications – Emphasis on most realistic applications, no programming model favored

HPC Benchmarking, Rudi Eigenmann, 2006 SPEC Benchmark Workshop Ranklist of Supercomputers based on Realistic Applications (SPEC HPC, medium data set)

HPC Benchmarking, Rudi Eigenmann, 2006 SPEC Benchmark Workshop Can We Learn the Same from Kerrnel Benchmarks?

HPC Benchmarking, Rudi Eigenmann, 2006 SPEC Benchmark Workshop MPI Communication (percent of overall runtime)

HPC Benchmarking, Rudi Eigenmann, 2006 SPEC Benchmark Workshop I/O Behavior Disk Read/Write Times and Volumes Only SPECseis has parallel I/O. SPECenv and SPECchem perform I/O on a single processor. HPL has no I/O

HPC Benchmarking, Rudi Eigenmann, 2006 SPEC Benchmark Workshop I/O Volume, Time, and Effective Bandwidth

HPC Benchmarking, Rudi Eigenmann, 2006 SPEC Benchmark Workshop Memory Footprints

HPC Benchmarking, Rudi Eigenmann, 2006 SPEC Benchmark Workshop Conclusions There is a dire need for basing performance evaluation and benchmarking results on realistic applications. The SPEC HPC the main criteria for real-application benchmarking: relevance and openness. Kernel benchmarks are the best choices for measuring individual system components. However, there is a large range of questions that can only be answered satisfactorily using real-application benchmarks. Benchmarking with real applications is hard and there are many challenges, but there is no replacement.