Ch1. Fundamentals of Computer Design 2. Performance ECE562/468 Advanced Computer Architecture Prof. Honggang Wang ECE Department University of Massachusetts.

Slides:



Advertisements
Similar presentations
CS1104: Computer Organisation School of Computing National University of Singapore.
Advertisements

Performance What differences do we see in performance? Almost all computers operate correctly (within reason) Most computers implement useful operations.
TU/e Processor Design 5Z032 1 Processor Design 5Z032 The role of Performance Henk Corporaal Eindhoven University of Technology 2009.
CSCE 212 Chapter 8 Storage, Networks, and Other Peripherals Instructor: Jason D. Bakos.
Lecture 7: 9/17/2002CS170 Fall CS170 Computer Organization and Architecture I Ayman Abdel-Hamid Department of Computer Science Old Dominion University.
Ch1. Fundamentals of Computer Design 3. Principles (5) ECE562/468 Advanced Computer Architecture Prof. Honggang Wang ECE Department University of Massachusetts.
Instruction Level Parallelism and Its Exploitation.
CPSC 614 Computer Architecture Lec 2 - Introduction EJ Kim Dept. of Computer Science Texas A&M University Adapted from CS 252 Spring 2006 UC Berkeley Copyright.
CIS629 Fall Lecture Performance Overview Execution time is the best measure of performance: simple, intuitive, straightforward. Two important.
Copyright © 1998 Wanda Kunkle Computer Organization 1 Chapter 2.5 Comparing and Summarizing Performance.
1 The Difference Engine, Charles Babbage Images from Wikipedia (Joe D and Andrew Dunn) Slides courtesy Anselmo Lastra.
Computer ArchitectureFall 2007 © September 17, 2007 Karem Sakallah CS-447– Computer Architecture.
1 Roman Japanese Chinese (compute in hex?). 2 COMP 206: Computer Architecture and Implementation Montek Singh Thu, Jan 22, 2009 Lecture 3: Quantitative.
1 Lecture 11: Digital Design Today’s topics:  Evaluating a system  Intro to boolean functions.
Copyright © 1998 Wanda Kunkle Computer Organization 1 Chapter 2.1 Introduction.
Chapter 4 Assessing and Understanding Performance
1 Introduction Background: CS 3810 or equivalent, based on Hennessy and Patterson’s Computer Organization and Design Text for CS/EE 6810: Hennessy and.
1 Lecture 10: FP, Performance Metrics Today’s topics:  IEEE 754 representations  FP arithmetic  Evaluating a system Reminder: assignment 4 due in a.
CIS429/529 Winter 07 - Performance - 1 Performance Overview Execution time is the best measure of performance: simple, intuitive, straightforward. Two.
1 Chapter 4. 2 Measure, Report, and Summarize Make intelligent choices See through the marketing hype Key to understanding underlying organizational motivation.
COURSE: JUST 3900 INTRODUCTORY STATISTICS FOR CRIMINAL JUSTICE Instructor: Dr. John J. Kerbs, Associate Professor Joint Ph.D. in Social Work and Sociology.
CSE820 Lec 2 - Introduction Rich Enbody Based on slides by David Patterson.
Using Standard Industry Benchmarks Chapter 7 CSE807.
CMSC 611: Advanced Computer Architecture Benchmarking Some material adapted from Mohamed Younis, UMBC CMSC 611 Spr 2003 course slides Some material adapted.
Lecture 2: Technology Trends and Performance Evaluation Performance definition, benchmark, summarizing performance, Amdahl’s law, and CPI.
CS136, Advanced Architecture Introduction to Architecture (continued)
Computer Organization CS224 Fall 2012 Lesson 51. Measuring I/O Performance  I/O performance depends on l Hardware: CPU, memory, controllers, buses l.
1 Computer Performance: Metrics, Measurement, & Evaluation.
CSE 520 Computer Architecture Lec 4 – Quantifying Cost, Energy- Consumption, Performance, and Dependability (Chapter 1) Sandeep K. S. Gupta School of Computing.
Ch2. Instruction-Level Parallelism & Its Exploitation 2. Dynamic Scheduling ECE562/468 Advanced Computer Architecture Prof. Honggang Wang ECE Department.
Computer Performance Computer Engineering Department.
Lecture 2b: Performance Metrics. Performance Metrics Measurable characteristics of a computer system: Count of an event Duration of a time interval Size.
BİL 221 Bilgisayar Yapısı Lab. – 1: Benchmarking.
Memory/Storage Architecture Lab Computer Architecture Performance.
Recap Technology trends Cost/performance Measuring and Reporting Performance What does it mean to say “computer X is faster than computer Y”? E.g. Machine.
Computer Architecture Lec 2 - Introduction. 01/19/10Lec 02-intro 2 Review from last lecture Computer Architecture >> instruction sets Computer Architecture.
1 Introduction Background: CS 3810 or equivalent, based on Hennessy and Patterson’s Computer Organization and Design Text for CS/EE 5810/6810: Hennessy.
10/19/2015Erkay Savas1 Performance Computer Architecture – CS401 Erkay Savas Sabanci University.
1 CS/EE 362 Hardware Fundamentals Lecture 9 (Chapter 2: Hennessy and Patterson) Winter Quarter 1998 Chris Myers.
Describing Data Using Numerical Measures. Topics.
Advanced Computer Networks Topic 2: Characterization of Distributed Systems.
Ch4. Multiprocessors & Thread-Level Parallelism 2. SMP (Symmetric shared-memory Multiprocessors) ECE468/562 Advanced Computer Architecture Prof. Honggang.
CS 5513 Computer Architecture Lecture 2 – More Introduction, Measuring Performance.
Eng. Mohammed Timraz Electronics & Communication Engineer University of Palestine Faculty of Engineering and Urban planning Software Engineering Department.
1 CS/COE0447 Computer Organization & Assembly Language CHAPTER 4 Assessing and Understanding Performance.
Lecture 8: 9/19/2002CS170 Fall CS170 Computer Organization and Architecture I Ayman Abdel-Hamid Department of Computer Science Old Dominion University.
1 Seoul National University Performance. 2 Performance Example Seoul National University Sonata Boeing 727 Speed 100 km/h 1000km/h Seoul to Pusan 10 hours.
CMSC 611: Advanced Computer Architecture Benchmarking Some material adapted from Mohamed Younis, UMBC CMSC 611 Spr 2003 course slides Some material adapted.
KT6213 Lecture 7: Pipelining Computer Organization and Architecture.
1  1998 Morgan Kaufmann Publishers How to measure, report, and summarize performance (suorituskyky, tehokkuus)? What factors determine the performance.
Microprocessor Microarchitecture Introduction Lynn Choi School of Electrical Engineering.
September 10 Performance Read 3.1 through 3.4 for Wednesday Only 3 classes before 1 st Exam!
Today: Standard Deviations & Z-Scores Any questions from last time?
Yiorgos Makris Professor Department of Electrical Engineering University of Texas at Dallas EE (CE) 6304 Computer Architecture Lecture #4 (9/3/15) Course.
Final Exam ECE562/468 Advanced Computer Architecture Prof. Honggang Wang ECE Department University of Massachusetts Dartmouth 285 Old Westport Rd. North.
CMSC 611: Advanced Computer Architecture Performance & Benchmarks Some material adapted from Mohamed Younis, UMBC CMSC 611 Spr 2003 course slides Some.
Jan. 5, 2000Systems Architecture II1 Machine Organization (CS 570) Lecture 2: Performance Evaluation and Benchmarking * Jeremy R. Johnson Wed. Oct. 4,
Ch4. Multiprocessors & Thread-Level Parallelism 4. Syn (Synchronization & Memory Consistency) ECE562 Advanced Computer Architecture Prof. Honggang Wang.
Ch2. Instruction-Level Parallelism & Its Exploitation 2. Dynamic Scheduling ECE562/468 Advanced Computer Architecture Prof. Honggang Wang ECE Department.
June 20, 2001Systems Architecture II1 Systems Architecture II (CS ) Lecture 1: Performance Evaluation and Benchmarking * Jeremy R. Johnson Wed.
Computer Architecture & Operations I
Measuring Performance II and Logic Design
Lecture 2: Performance Evaluation
September 2 Performance Read 3.1 through 3.4 for Tuesday
Ch1. Fundamentals of Computer Design 3. Principles (5)
EE (CE) 6304 Computer Architecture Lecture #4 (8/30/17)
CS2100 Computer Organisation
CMSC 611: Advanced Computer Architecture
CMSC 611: Advanced Computer Architecture
CS2100 Computer Organisation
Presentation transcript:

Ch1. Fundamentals of Computer Design 2. Performance ECE562/468 Advanced Computer Architecture Prof. Honggang Wang ECE Department University of Massachusetts Dartmouth 285 Old Westport Rd. North Dartmouth, MA Slides based on the PowerPoint Presentations created by David Patterson as part of the Instructor Resources for the textbook by Hennessy & Patterson Updated by Honggang Wang.

Administrative Issues (02/012/2015) Project team set-up due Thursday, Feb. 12 – Your group leader must send me an about your group information by 02/12 If you missed some classes, go to the course website for syllabus and 1 st lecture. 2

Review of Lecture #1 In the first lecture, we covered the Technology Trends 1.Define and quantity power 2.Define and quantity relative cost 3.Define and quantity dependability 3

Outline Careful, quantitative comparisons: Performance 1.Compute Mean 2.Benchmark 4

Definition: Performance 5 Performance is in events per sec: –bigger is better If we are primarily concerned with response time: the time between the start and the completion of an event (execution time). " X is n times faster than Y" means:

Performance: What to measure To increase predictability, collections of benchmark applications, called benchmark suites, are popular reliable measure of the performance is the execution time of real program the users would simply compare the execution time of their workloads a popular measure of performance of processors with a variety of applications SPECCPU: popular desktop benchmark suite – CPU only, split between integer and floating point programs – SPECint2000 has 12 integer, SPECfp2000 has 14 integer pgms – SPECCPU2006 to be announced Spring 2006 – SPECSFS (NFS file server) and SPECWeb (WebServer) added as server benchmarks Transaction Processing Council measures server performance and cost- performance for databases ( – TPC-C Complex query for Online Transaction Processing – TPC-H models ad hoc decision support – TPC-W a transactional web benchmark – TPC-App application server and web services benchmark 6

How Summarize Suite Performance (1/5) Arithmetic average of execution time of all pgms? – But they vary by 4X in speed, so some would be more important than others in arithmetic average Could add a weights per program, but how pick up weight? – Different companies want different weights for their products SPECRatio: Normalize execution times to reference computer, yielding a ratio proportional to performance 7

How Summarize Suite Performance (2/5) If program SPECRatio on Computer A is 1.25 times bigger than Computer B, then 8 Note that when comparing 2 computers as a ratio, execution times on the reference computer drop out, so choice of reference computer is irrelevant

How Summarize Suite Performance (3/5) Since ratios, proper mean is geometric mean (SPECRatio unitless, so arithmetic mean meaningless) 9 1. The SPECRatio are multiplied and then the nth root (where n is the count of SPECRatio in the set) of the resulting product is takennth rootproduct 2. indicates the central tendency or typical value of a set of numbers

How Summarize Suite Performance (3/5) Suppose an orange tree yields 100 oranges one year and then 180, 210 and 300 the following years, so the growth is 80%, 16.7% and 42.9% for each year respectively. Using the arithmetic mean calculates a (linear) average growth of 46.5% (80% % % divided by 3).arithmetic mean However, if we start with 100 oranges and let it grow 46.5% each year, the result is 314 oranges, not 300, so the linear average over- states the year-on-year growth. Instead, we can use the geometric mean. Growing with 80% corresponds to multiplying with 1.80, so we take the geometric mean of 1.80, and 1.429, i.e., thus the "average" growth per year is 44.3%. If we start with 100 oranges and let the number grow with 44.3% each year, the result is 300 oranges. 10

How Summarize Suite Performance (4/5) Does a single mean well summarize performance of programs in benchmark suite? Can decide if mean is a good predictor by characterizing variability of distribution using standard deviation Like geometric mean, geometric standard deviation is multiplicative rather than arithmetic Can simply take the logarithm of SPECRatios, compute the standard mean and standard deviation, and then take the exponent to convert back: 11

How Summarize Suite Performance (5/5) Standard deviation is more informative if know distribution has a standard form – bell-shaped normal distribution, whose data are symmetric around mean – lognormal distribution, where logarithms of data--not data itself--are normally distributed (symmetric) on a logarithmic scale For a lognormal distribution, we expect that 68% of samples fall in range 95% of samples fall in range Note: Excel provides functions EXP(), LN(), and STDEV() that make calculating geometric mean and multiplicative standard deviation easy 12

Example Standard Deviation (1/2) GM and multiplicative StDev of SPECfp2000 for Itanium 2 13 Outside 1 StDev

Example Standard Deviation (2/2) GM and multiplicative StDev of SPECfp2000 for AMD Athlon 14 Outside 1 StDev

Comments on Itanium 2 and Athlon Standard deviation of 1.98 for Itanium 2 is much higher-- vs so results will differ more widely from the mean, and therefore are likely less predictable Falling within one standard deviation: – 10 of 14 benchmarks (71%) for Itanium 2 – 11 of 14 benchmarks (78%) for Athlon Thus, the results are quite compatible with a lognormal distribution (expect 68%) 15

Fallacies and Pitfalls (1/2) Fallacies - commonly held misconceptions – When discussing a fallacy, we try to give a counter example. Pitfalls - easily made mistakes. – Often generalizations of principles true in limited context – Show Fallacies and Pitfalls to help you avoid these errors Fallacy: Benchmarks remain valid indefinitely – Once a benchmark becomes popular, tremendous pressure to improve performance by targeted optimizations or by aggressive interpretation of the rules for running the benchmark: “benchmarksmanship.” – 70 benchmarks from the 5 SPEC releases. 70% were dropped from the next release since no longer useful Pitfall: A single point of failure – Rule of thumb for fault tolerant systems: make sure that every component was redundant so that no single component failure could bring down the whole system (e.g, power supply) 16

Fallacies and Pitfalls (2/2) Fallacy: Rated MTTF of disks is 1,200,000 hours or  140 years, so disks practically never fail – manufactures will put thousands of disks in a room, run them for a few months and count the number that fail – But disk lifetime is 5 years  replace a disk every 5 years; on average, 28 replacements wouldn't fail – A better unit: % that fail (1.2M MTTF = 833 FIT) – Fail over lifetime: if had 1000 disks for 5 years = 1000*(5*365*24)*833 /10 9 = 36,485,000 / 10 6 = 37 = 3.7% (37/1000) fail over 5 yr hr MTTF) But this is under pristine conditions lifetime (1.2M – little vibration, narrow temperature range  no power failures Real world: 3% to 6% of SCSI drives fail per year – FIT or 150, ,000 hour MTTF [Gray & van Ingen 05] 17

Summary Quantify and summarize performance – Ratios, Geometric Mean, Multiplicative Standard Deviation 18

19 Things To Do Find your partners for the class project – Feb. 12, Thursday ( me the team information) Check out the class website about – lecture notes – reading assignments Next Topics Fundamentals of Computer Design 3. Principles