CDA 3101 Fall 2013 Introduction to Computer Organization Benchmarks 30 August 2013.

Slides:



Advertisements
Similar presentations
CS2100 Computer Organisation Performance (AY2014/2015) Semester 2.
Advertisements

810:142 Lecture 2: Performance Fall 2006 Chapter 4: Performance Adapted from Mary Jane Irwin at Penn State University for Computer Organization and Design,
TU/e Processor Design 5Z032 1 Processor Design 5Z032 The role of Performance Henk Corporaal Eindhoven University of Technology 2009.
Performance COE 308 Computer Architecture Prof. Muhamed Mudawar Computer Engineering Department King Fahd University of Petroleum and Minerals.
Computer Performance CS350 Term Project-Spring 2001 Elizabeth Cramer Bryan Driskell Yassaman Shayesteh.
Lecture 7: 9/17/2002CS170 Fall CS170 Computer Organization and Architecture I Ayman Abdel-Hamid Department of Computer Science Old Dominion University.
Computer Organization and Architecture (AT70.01) Comp. Sc. and Inf. Mgmt. Asian Institute of Technology Instructor: Dr. Sumanta Guha Slide Sources: Patterson.
CSCE 212 Chapter 4: Assessing and Understanding Performance Instructor: Jason D. Bakos.
EECE476: Computer Architecture Lecture 12: Evaluating Performance …and the Importance of Benchmarks! Chapter 4.3, 4.4, 4.5 There is more material in this.
CPU Performance Evaluation: Cycles Per Instruction (CPI)
Performance ICS 233 Computer Architecture and Assembly Language Dr. Aiman El-Maleh College of Computer Sciences and Engineering King Fahd University of.
EECC550 - Shaaban #1 Lec # 3 Spring Computer Performance Evaluation: Cycles Per Instruction (CPI) Most computers run synchronously utilizing.
CIS629 Fall Lecture Performance Overview Execution time is the best measure of performance: simple, intuitive, straightforward. Two important.
1 CSE SUNY New Paltz Chapter 2 Performance and Its Measurement.
1 Recap. 2 Measuring Performance  A computer user: response time (execution time).  A computer center manager - throughput - the total amount of work.
Copyright © 1998 Wanda Kunkle Computer Organization 1 Chapter 2.5 Comparing and Summarizing Performance.
Computer Performance Evaluation: Cycles Per Instruction (CPI)
CS/ECE 3330 Computer Architecture Chapter 1 Performance / Power.
Chapter 4 Assessing and Understanding Performance
CIS429/529 Winter 07 - Performance - 1 Performance Overview Execution time is the best measure of performance: simple, intuitive, straightforward. Two.
1 Chapter 4. 2 Measure, Report, and Summarize Make intelligent choices See through the marketing hype Key to understanding underlying organizational motivation.
Benchmarks Programs specifically chosen to measure performance Must reflect typical workload of the user Benchmark types Real applications Small benchmarks.
SPEC 2006 CSE 820. Michigan State University Computer Science and Engineering Q1. What is SPEC? SPEC is the Standard Performance Evaluation Corporation.
CMSC 611: Advanced Computer Architecture Benchmarking Some material adapted from Mohamed Younis, UMBC CMSC 611 Spr 2003 course slides Some material adapted.
1 Computer Performance: Metrics, Measurement, & Evaluation.
Where Has This Performance Improvement Come From? Technology –More transistors per chip –Faster logic Machine Organization/Implementation –Deeper pipelines.
Benchmarks Prepared By : Arafat El-madhoun Supervised By:eng. Mohammad temraz.
CSE 340 Computer Architecture Summer 2014 Understanding Performance
Computer Performance Computer Engineering Department.
BİL 221 Bilgisayar Yapısı Lab. – 1: Benchmarking.
CDA 3101 Fall 2013 Introduction to Computer Organization Computer Performance 28 August 2013.
10/19/2015Erkay Savas1 Performance Computer Architecture – CS401 Erkay Savas Sabanci University.
1 CS/EE 362 Hardware Fundamentals Lecture 9 (Chapter 2: Hennessy and Patterson) Winter Quarter 1998 Chris Myers.
ACMSE’04, ALDepartment of Electrical and Computer Engineering - UAH Execution Characteristics of SPEC CPU2000 Benchmarks: Intel C++ vs. Microsoft VC++
Computer Architecture
CEN 316 Computer Organization and Design Assessing and Understanding Performance Mansour AL Zuair.
CMSC 611: Advanced Computer Architecture Benchmarking Some material adapted from Mohamed Younis, UMBC CMSC 611 Spr 2003 course slides Some material adapted.
Lecture2: Performance Metrics Computer Architecture By Dr.Hadi Hassan 1/3/2016Dr. Hadi Hassan Computer Architecture 1.
1  1998 Morgan Kaufmann Publishers How to measure, report, and summarize performance (suorituskyky, tehokkuus)? What factors determine the performance.
Performance Performance
TEST 1 – Tuesday March 3 Lectures 1 - 8, Ch 1,2 HW Due Feb 24 –1.4.1 p.60 –1.4.4 p.60 –1.4.6 p.60 –1.5.2 p –1.5.4 p.61 –1.5.5 p.61.
1 Lecture 2: Performance, MIPS ISA Today’s topics:  Performance equations  MIPS instructions Reminder: canvas and class webpage:
September 10 Performance Read 3.1 through 3.4 for Wednesday Only 3 classes before 1 st Exam!
Performance – Last Lecture Bottom line performance measure is time Performance A = 1/Execution Time A Comparing Performance N = Performance A / Performance.
Lec2.1 Computer Architecture Chapter 2 The Role of Performance.
Performance COE 301 Computer Organization Dr. Muhamed Mudawar College of Computer Sciences and Engineering King Fahd University of Petroleum and Minerals.
ECE/CS 552: Benchmarks, Means and Amdahl’s Law © Prof. Mikko Lipasti Lecture notes based in part on slides created by Mark Hill, David Wood, Guri Sohi,
CMSC 611: Advanced Computer Architecture Performance & Benchmarks Some material adapted from Mohamed Younis, UMBC CMSC 611 Spr 2003 course slides Some.
Jan. 5, 2000Systems Architecture II1 Machine Organization (CS 570) Lecture 2: Performance Evaluation and Benchmarking * Jeremy R. Johnson Wed. Oct. 4,
Computer Architecture CSE 3322 Web Site crystal.uta.edu/~jpatters/cse3322 Send to Pramod Kumar, with the names and s.
Performance COE 301 / ICS 233 Computer Organization Prof. Muhamed Mudawar College of Computer Sciences and Engineering King Fahd University of Petroleum.
BITS Pilani, Pilani Campus Today’s Agenda Role of Performance.
CSE 340 Computer Architecture Summer 2016 Understanding Performance.
June 20, 2001Systems Architecture II1 Systems Architecture II (CS ) Lecture 1: Performance Evaluation and Benchmarking * Jeremy R. Johnson Wed.
Measuring Performance and Benchmarks Instructor: Dr. Mike Turi Department of Computer Science and Computer Engineering Pacific Lutheran University Lecture.
Measuring Performance II and Logic Design
Lecture 2: Performance Evaluation
CS161 – Design and Architecture of Computer Systems
Assessing and Understanding Performance
Defining Performance Which airplane has the best performance?
CSCE 212 Chapter 4: Assessing and Understanding Performance
CMSC 611: Advanced Computer Architecture
Performance of computer systems
Performance ICS 233 Computer Architecture and Assembly Language
CMSC 611: Advanced Computer Architecture
Performance of computer systems
COMS 361 Computer Organization
Parameters that affect it How to improve it and by how much
Benchmarks Programs specifically chosen to measure performance
CS161 – Design and Architecture of Computer Systems
Presentation transcript:

CDA 3101 Fall 2013 Introduction to Computer Organization Benchmarks 30 August 2013

Overview Benchmarks Popular benchmarks –Linpack –Intel’s iCOMP SPEC Benchmarks MIPS Benchmark Fallacies and Pitfalls

Benchmarks Benchmarks measure different aspects of component and system performance Ideal situation: use real workload Types of Benchmarks Risk: adjust design to benchmark requirements –(partial) solution: use real programs and update constantly Engineering or scientific applications Software development tools Transaction processing Office applications Real programs Kernels Toy benchmarks Synthetic benchmarks

A /  Benchmark Story 1.You create a benchmark called the vmark 2.Run it on lots of different computers 3.Publish the vmarks in 4.vmark and become popular –Users start buying their PCs based on vmark –Vendors would be banging on your door 5.Vendors examine the vmark code and fix up their compilers and/or microarchitecture to run vmark 6.Your vmark benchmark has been broken  7.Create vmark 2.0

Performance Reports Reproducibility –Include hardware / software configuration (SPEC) –Evaluation process conditions Summarizing performance –Total time: –Arithmetic mean: AM = 1/n * Σ exec time i –Harmonic mean: HM = n / Σ (1/rate i ) –Weighted mean: WM = Σ w i * exec time i –Geometric mean: GM = (Π exec time ratio i ) 1/n GM (X i ) X i GM (Y i ) Y i = GM

Ex.1: Linpack Benchmark “Mother of all benchmarks” Time to solve a dense systems of linear equations DO I = 1, N DY(I) = DY(I) + DA * DX(I) END DO Metrics –R peak : system peak Gflops –N max : matrix size that gives the highest Gflops –N 1/2 : matrix size that achieves half the rated R max Gflops –R max : the Gflops achieved for the N max size matrix Used in

Ex.2: Intel’s iCOMP Index 3.0 New version (3.0) reflects: Mix of instructions for existing and emerging software. Increasing use of 3D, multimedia, and Internet software. Benchmarks 2 integer productivity applications (20% each) 3D geometry and lighting calculations (20%) FP engineering and finance programs and games (5%) Multimedia and Internet application (25%. ) Java application (10%) Weighted GM of relative performance –Baseline processor: Pentium II processor at 350MHz

Ex.3: SPEC CPU Benchmarks System Performance Evaluation Corporation Need to update/upgrade benchmarks –Longer run time –Larger problems –Application diversity Rules to run and report –Baseline and optimized –Geometric mean of normalized execution times –Reference machine: Sun Ultra5_10 ( 300-MHz SPARC, 256MB ) CPU2006: latest SPEC CPU benchmark (4 th version) –12 integer and 17 floating point programs Metrics: response time and throughput

Ex.3: SPEC CPU Benchmarks Previous Benchmarks, now retired

Ex.3: SPEC CPU Benchmarks Observe: We will use SPEC 2000 & 2006 CPU benchmarks in this set of notes. Task: However, you are asked to read about SPEC 2006 CPU benchmark suite, described at Result: Compare SPEC 2006 with SPEC 2000 data to answer the extra-credit questions in Homework #2.

SPEC CINT2000 Benchmarks gzip CCompression vpr CFPGA Circuit Placement and Routing gcc CC Programming Language Compiler mcf CCombinatorial Optimization crafty CGame Playing: Chess parser CWord Processing eon C++Computer Visualization perlbmk CPERL Programming Language gap CGroup Theory, Interpreter vortex CObject-oriented Database bzip2 CCompression twolf CPlace and Route Simulator

SPEC CFP2000 Benchmarks wupwise F77Physics / Quantum Chromodynamics swim F77Shallow Water Modeling mgrid F77Multi-grid Solver: 3D Potential Field applu F77Parabolic / Elliptic Partial Differential Equations mesa C3-D Graphics Library galgel F90Computational Fluid Dynamics art CImage Recognition / Neural Networks equake CSeismic Wave Propagation Simulation facerec F90Image Processing: Face Recognition ammp CComputational Chemistry lucas F90Number Theory / Primality Testing fma3d F90Finite-element Crash Simulation sixtrack F77High Energy Nuclear Physics Accelerator Design apsi F77Meteorology: Pollutant Distribution

SPECINT2000 Metrics SPECint2000: The geometric mean of 12 normalized ratios (one for each integer benchmark) when each benchmark is compiled with "aggressive" optimization SPECint_base2000: The geometric mean of 12 normalized ratios when compiled with "conservative" optimization SPECint_rate2000: The geometric mean of 12 normalized throughput ratios when compiled with "aggressive" optimization SPECint_rate_base2000: The geometric mean of 12 normalized throughput ratios when compiled with "conservative" optimization

SPECint_base2000 Results Alpha/Tru MHz Mips/IRIX 400MHz Intel/NT MHz

SPECfp_base2000 Results Alpha/Tru MHz Mips/IRIX 400MHz Intel/NT MHz

Effect of CPI: SPECint95 Ratings Microarchitecture improvements CPU time = IC * CPI * clock cycle

Effect of CPI: SPECfp95 Ratings Microarchitecture improvements

SPEC Recommended Readings  SPEC 2006 – Survey of Benchmark Programs  SPEC 2006 Benchmarks - Journal Articles on Implementation Techniques and Problems  SPEC 2006 Installation, Build, and Runtime Issues

Another Benchmark: MIPS  Millions of Instructions Per Second MIPS = IC / (CPUtime * 10 6 ) Comparing apples to oranges Flaw: 1 MIPS on one processor does not accomplish the same work as 1 MIPS on another –It is like determining the winner of a foot race by counting who used fewer steps –Some processors do FP in software (e.g. 1FP = 100 INT) –Different instructions take different amounts of time Useful for comparisons between 2 processors from the same vendor that support the same ISA with the same compiler (e.g. Intel’s iCOMP benchmark)

Fallacies and Pitfalls Ignoring Amdahl’s law Using clock rate or MIPS as a performance metric Using the Arithmetic Mean of normalized CPU times (ratios) instead of the Geometric Mean Using hardware-independent metrics –Using code size as a measure of speed Synthetic benchmarks predict performance –They do not reflect the behavior of real programs The geometric mean of CPU times ratios is proportional to the total execution time [NOT!!]

Conclusions Performance is specific to a particular program/s CPU time: only adequate measure of performance For a given ISA performance increases come from: –increases in clock rate (without adverse CPI affects) –improvements in processor organization that lower CPI –compiler enhancements that lower CPI and/or IC Your workload: the ideal benchmark You should not always believe everything you read!

Happy & Safe Holiday Weekend