SPEC OMP Benchmark Suite H. Saito, G. Gaertner, W. Jones, R. Eigenmann, H. Iwashita, R. Lieberman, M. van Waveren, and B. Whitney SPEC High-Performance.

Slides:



Advertisements
Similar presentations
SPEC High Performance Group (An Introduction)
Advertisements

SPEC ENV2002: Environmental Simulation Benchmarks Wesley Jones
SPEC MPI2007 Benchmarks for HPC Systems Ron Lieberman Chair, SPEC HPG HP-MPI Performance Hewlett-Packard Company Dr. Tom Elken Manager, Performance Engineering.
SPECs High-Performance Benchmark Suite, SPEC HPC Rudi Eigenmann Purdue University Indiana, USA.
HPC Benchmarking, Rudi Eigenmann, 2006 SPEC Benchmark Workshop this is important for machine procurements and for understanding where HPC technology is.
Höchstleistungsrechenzentrum Stuttgart Matthias M üller SPEC Benchmarks for Large Systems Matthias Mueller, Kumaran Kalyanasundaram, G. Gaertner, W. Jones,
SPEC HPG Benchmarks for HPC Systems Kumaran Kalyanasundaram for SPEC High-Performance Group Kumaran Kalyanasundaram, PhD Chair, SPEC HPG Manager, SGI Performace.
Höchstleistungsrechenzentrum Stuttgart Matthias M üller Current Efforts of SPEC HPG Application Benchmarks for High Performance Computing IPSJ SIGMPS 2003.
Höchstleistungsrechenzentrum Stuttgart Matthias M üller Overview of SPEC HPG Benchmarks SPEC BOF SC2003 Matthias Mueller High Performance Computing Center.
OpenMP.
Nios Multi Processor Ethernet Embedded Platform Final Presentation
C O M P U T A T I O N A L R E S E A R C H D I V I S I O N Scaling of the Community Atmospheric Model to ultrahigh resolution Michael F. Wehner Lawrence.
NPACI Parallel Computing Institute August 19-23, 2002 San Diego Supercomputing Center S an D IEGO S UPERCOMPUTER C ENTER N ATIONAL P ARTNERSHIP FOR A DVANCED.
Parallel Processing with OpenMP
Introduction to Openmp & openACC
The OpenUH Compiler: A Community Resource Barbara Chapman University of Houston March, 2007 High Performance Computing and Tools Group
PERFORMANCE ANALYSIS OF MULTIPLE THREADS/CORES USING THE ULTRASPARC T1 (NIAGARA) Unique Chips and Systems (UCAS-4) Dimitris Kaseridis & Lizy K. John The.
Autonomic Systems Justin Moles, Winter 2006 Enabling autonomic behavior in systems software with hot swapping Paper by: J. Appavoo, et al. Presentation.
NewsFlash!! Earth Simulator no longer #1. In slightly less earthshaking news… Homework #1 due date postponed to 10/11.
Optimized Parallel Approach for 3D Modelling of Forest Fire Behaviour G. Accary, O. Bessonov, D. Fougère, S. Meradji, D. Morvan Institute for Problems.
Lecture 7: 9/17/2002CS170 Fall CS170 Computer Organization and Architecture I Ayman Abdel-Hamid Department of Computer Science Old Dominion University.
Chapter 4 M. Keshtgary Spring 91 Type of Workloads.
Trace-Based Automatic Parallelization in the Jikes RVM Borys Bradel University of Toronto.
CSCE 212 Chapter 4: Assessing and Understanding Performance Instructor: Jason D. Bakos.
CS2214 Recitation Presented By Veejay Sani. Benchmarking SPEC CPU2000 Integer Benchmark Floating Point Benchmark We will only deal with Integer Benchmark.
Parallellization of a semantic discovery tool for a search engine Kalle Happonen Wray Buntine.
Copyright © The McGraw-Hill Companies, Inc. Permission required for reproduction or display. Parallel Programming with MPI and OpenMP Michael J. Quinn.
On the Integration and Use of OpenMP Performance Tools in the SPEC OMP2001 Benchmarks Bernd Mohr 1, Allen D. Malony 2, Rudi Eigenmann 3 1 Forschungszentrum.
Lecture 5 Today’s Topics and Learning Objectives Quinn Chapter 7 Predict performance of parallel programs Understand barriers to higher performance.
Unified Parallel C at LBNL/UCB FT Benchmark in UPC Christian Bell and Rajesh Nishtala.
SPEC 2006 CSE 820. Michigan State University Computer Science and Engineering Q1. What is SPEC? SPEC is the Standard Performance Evaluation Corporation.
CPU Performance Assessment As-Bahiya Abu-Samra *Moore’s Law *Clock Speed *Instruction Execution Rate - MIPS - MFLOPS *SPEC Speed Metric *Amdahl’s.
1 Titanium Review: Ti Parallel Benchmarks Kaushik Datta Titanium NAS Parallel Benchmarks Kathy Yelick U.C. Berkeley September.
Budapest, November st ALADIN maintenance and phasing workshop Short introduction to OpenMP Jure Jerman, Environmental Agency of Slovenia.
Executing OpenMP Programs Mitesh Meswani. Presentation Outline Introduction to OpenMP Machine Architectures Shared Memory (SMP) Distributed Memory MPI.
A COMPARISON MPI vs POSIX Threads. Overview MPI allows you to run multiple processes on 1 host  How would running MPI on 1 host compare with POSIX thread.
University of Maryland Compiler-Assisted Binary Parsing Tugrul Ince PD Week – 27 March 2012.
Computer Performance Computer Engineering Department.
TECH 6 VLIW Architectures {Very Long Instruction Word}
Uncovering the Multicore Processor Bottlenecks Server Design Summit Shay Gal-On Director of Technology, EEMBC.
Recap Technology trends Cost/performance Measuring and Reporting Performance What does it mean to say “computer X is faster than computer Y”? E.g. Machine.
S AN D IEGO S UPERCOMPUTER C ENTER N ATIONAL P ARTNERSHIP FOR A DVANCED C OMPUTATIONAL I NFRASTRUCTURE On pearls and perils of hybrid OpenMP/MPI programming.
C OMPUTER O RGANIZATION AND D ESIGN The Hardware/Software Interface 5 th Edition Chapter 1 Computer Abstractions and Technology Sections 1.5 – 1.11.
Simulating a $2M Commercial Server on a $2K PC Alaa R. Alameldeen, Milo M.K. Martin, Carl J. Mauer, Kevin E. Moore, Min Xu, Daniel J. Sorin, Mark D. Hill.
 Copyright, HiPERiSM Consulting, LLC, George Delic, Ph.D. HiPERiSM Consulting, LLC (919) P.O. Box 569, Chapel Hill, NC.
Hybrid MPI and OpenMP Parallel Programming
Lecture 9 TTH 03:30AM-04:45PM Dr. Jianjun Hu CSCE569 Parallel Computing University of South Carolina Department of.
A User-Lever Concurrency Manager Hongsheng Lu & Kai Xiao.
From lecture slides for Computer Organization and Architecture: Designing for Performance, Eighth Edition, Prentice Hall, 2010 CS 211: Computer Architecture.
Performance Lecture notes from MKP, H. H. Lee and S. Yalamanchili.
Parallel Programming with MPI and OpenMP
TM Parallel Concepts An introduction. TM The Goal of Parallelization Reduction of elapsed time of a program Reduction in turnaround time of jobs Overhead:
Threaded Programming Lecture 2: Introduction to OpenMP.
EGEE-III INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks John Gordon SA1 Face to Face CERN, June.
Jim Conboy / DPGTF-H 20-May Toric at JET – Status & Tools Work in Progress – See for latest.
Single Node Optimization Computational Astrophysics.
Benchmarking and Applications. Purpose of Our Benchmarking Effort Reveal compiler (and run-time systems) weak points and lack of adequate automatic optimizations.
Lab Activities 1, 2. Some of the Lab Server Specifications CPU: 2 Quad(4) Core Intel Xeon 5400 processors CPU Speed: 2.5 GHz Cache : Each 2 cores share.
Introduction to HPC Debugging with Allinea DDT Nick Forrington
First INFN International School on Architectures, tools and methodologies for developing efficient large scale scientific computing applications Ce.U.B.
Introduction to Performance Tuning Chia-heng Tu PAS Lab Summer Workshop 2009 June 30,
OpenMP Lab Antonio Gómez-Iglesias Texas Advanced Computing Center.
Introduction to Parallel Computing: MPI, OpenMP and Hybrid Programming
Chen Jin (HT016952H) Zhao Xu Ying (HT016907B)
Performance Lecture notes from MKP, H. H. Lee and S. Yalamanchili.
Morgan Kaufmann Publishers
CERN Benchmarking Cluster
An Empirical Analysis of Java Performance Quality
CSCE 212 Chapter 4: Assessing and Understanding Performance
COMS 361 Computer Organization
Presentation transcript:

SPEC OMP Benchmark Suite H. Saito, G. Gaertner, W. Jones, R. Eigenmann, H. Iwashita, R. Lieberman, M. van Waveren, and B. Whitney SPEC High-Performance Group

Structure of talk Description of SPEC OMP Benchmarks Scalability of SPEC OMP Benchmarks Conclusion

SPEC OMP Benchmark suite developed by SPEC HPG (High Performance Group) Benchmark suite for performance testing of shared memory processor systems Uses OpenMP versions of SPEC CPU2000 benchmarks and candidates

Why Did SPEC Choose OpenMP? Benchmark suite is focused on SMP systems OpenMP is a standard, and is applicable to Fortran, C, and C++. Directive based OpenMP allows serial version to remain largely intact. Quickest path to parallel code conversion.

OMP/CPU2000 Differences Larger working set sizes, 1.6GB for OMPM2001, 6.5 GB for OMPL2001; it is 200MB for CPU2000 Longer run times (>1000 s/cpu for CPU2000 vs >10,000 s/cpu for OMP2000 medium) Focus on SMP systems, and issued by HPG SPEC OMP based on work for CPU2000, SPEC OMP mixes integer and FP in one suite

OMP/CPU2000 Similarities Same tools used to run the benchmarks Similar run and reporting rules Uses geometric mean to calculate overall performance relative to a baseline system Similar output format

OMP vs CPU2000

SPEC OMP Benchmark Principles Source code based Limited code and directive modifications Focused on SMP performance Requires a base run with no source modifications single set of compiler flags for all benchmarks SPEC supplied tools required to run benchmark

OMPM2001 Benchmarks

SPEC OMP Benchmark Reference Runtimes

Program Memory Footprints

Benchmarks with good scaling

Benchmarks with good scaling up to 64 CPUs

Benchmarks with superlinear scaling

Benchmarks with poor scaling

Conclusion SPEC OMP is a realistic set of benchmarks for SMP systems. Up to date results are available on the SPEC web site at Parallelization using OpenMP can be done relatively quickly, and on complex sections. Scalability up to 128 CPUs is possible using OpenMP.

SPEC OMP is Here Purchase through SPEC Academic discount available Membership encouraged