Copyright Hesham El-Rewini1 Speedup S = Speed(new) / Speed(old) S = Work/time(new) / Work/time(old) S = time(old) / time(new) S = time(before improvement)

Slides:



Advertisements
Similar presentations
Analyzing Parallel Performance Intel Software College Introduction to Parallel Programming – Part 6.
Advertisements

Performance Analysis of Multiprocessor Architectures
Computer Organization and Architecture 18 th March, 2008.
Example (1) Two computer systems have been tested using three benchmarks. Using the normalized ratio formula and the following tables below, find which.
Copyright © The McGraw-Hill Companies, Inc. Permission required for reproduction or display. Parallel Programming with MPI and OpenMP Michael J. Quinn.
Performance Metrics Parallel Computing - Theory and Practice (2/e) Section 3.6 Michael J. Quinn mcGraw-Hill, Inc., 1994.
Recap.
CIS429.S00: Lec2- 1 Performance Overview Execution time is the best measure of performance: simple, intuitive, straightforward. Two important quantitative.
CS 584 Lecture 11 l Assignment? l Paper Schedule –10 Students –5 Days –Look at the schedule and me your preference. Quickly.
Lecture 5 Today’s Topics and Learning Objectives Quinn Chapter 7 Predict performance of parallel programs Understand barriers to higher performance.
Computer Architecture Lecture 2 Instruction Set Principles.
Chapter 4 Performance. The Role of Performance Example of Performance Measure.
Steve Lantz Computing and Information Science Parallel Performance Week 7 Lecture Notes.
Amdahl's Law.
MisSPECulation Partial and Misleading use of SPEC CPU2000 in Computer Architecture Conferences Dave Kroondyk.
Energy Model for Multiprocess Applications Texas Tech University.
Computer Science 320 Measuring Speedup. What Is Running Time? T(N, K) says that the running time T is a function of the problem size N and the number.
Computer Science and Engineering Copyright by Hesham El-Rewini Advanced Computer Architecture.
“elbowing out” Processors used Speedup Efficiency timeexecution Parallel Processors timeexecution Sequential Efficiency   
INTEL CONFIDENTIAL Predicting Parallel Performance Introduction to Parallel Programming – Part 10.
Flynn’s Taxonomy SISD: Although instruction execution may be pipelined, computers in this category can decode only a single instruction in unit time SIMD:
Amdahl's Law Validity of the single processor approach to achieving large scale computing capabilities Presented By: Mohinderpartap Salooja.
Computer Science and Engineering Parallel and Distributed Processing CSE 8380 February Session 6.
Compiled by Maria Ramila Jimenez
Lecturer: Simon Winberg Lecture 18 Amdahl’s Law (+- 25 min)
Parallel Processing Steve Terpe CS 147. Overview What is Parallel Processing What is Parallel Processing Parallel Processing in Nature Parallel Processing.
Lecture 9 TTH 03:30AM-04:45PM Dr. Jianjun Hu CSCE569 Parallel Computing University of South Carolina Department of.
Scaling Area Under a Curve. Why do parallelism? Speedup – solve a problem faster. Accuracy – solve a problem better. Scaling – solve a bigger problem.
From lecture slides for Computer Organization and Architecture: Designing for Performance, Eighth Edition, Prentice Hall, 2010 CS 211: Computer Architecture.
Performance Lecture notes from MKP, H. H. Lee and S. Yalamanchili.
Parallel Programming with MPI and OpenMP
DO NOW Martina drove miles in 3 hours. At this rate, how far did Martina drive in one hour? Show all work.
Copyright © 2003, SAS Institute Inc. All rights reserved. SAS is a registered trademark or trademark of SAS Institute Inc. in the USA and other countries.
You bike 39 miles in 5 hours. Find the Unit Rate in miles per hour
Performance Enhancement. Performance Enhancement Calculations: Amdahl's Law The performance enhancement possible due to a given design improvement is.
1. Speed, Velocity, & Acceleration 2 You know a car is in motion if you see it in one place 3 then in another place in relation to an object.
Computer Science and Engineering Advanced Computer Architecture CSE 8383 February Session 6.
Computer Science and Engineering Copyright by Hesham El-Rewini Advanced Computer Architecture CSE 8383 May 2, 2006 Session 29.
Scaling Conway’s Game of Life. Why do parallelism? Speedup – solve a problem faster. Accuracy – solve a problem better. Scaling – solve a bigger problem.
Computer Engineering Rabie A. Ramadan Lecture 2. Table of Contents 2 Architecture Development and Styles Performance Measures Amdahl’s Law.
Amdahl’s Law CPS 5401 Fall 2013 Shirley Moore
Computer Science 320 Measuring Sizeup. Speedup vs Sizeup If we add more processors, we should be able to solve a problem of a given size faster If we.
Computer Science and Engineering Copyright by Hesham El-Rewini Advanced Computer Architecture CSE 8383 February Session 7.
Computer Science and Engineering Copyright by Hesham El-Rewini Advanced Computer Architecture CSE 8383 February Session 11.
Chapter 7 Input/Output and Storage Systems. 2 Chapter 7 Objectives Understand how I/O systems work, including I/O methods and architectures. Become familiar.
Matthew Royle Supervisor: Prof Shaun Bangay.  How do we implement OpenCL for CPUs  Differences in parallel architectures  Is our CPU implementation.
Performance. Moore's Law Moore's Law Related Curves.
Speed How many measurements for speed can you think of?
What is *Computer Architecture*
Software Architecture in Practice
Unit Rates with Fractions
Exercise – Given information in the following table:
Introduction to Parallelism.
Chapter 1 Fundamentals of Computer Design
Copyrights apply.
Parallel Computers.
Defining Performance Section /14/2018 9:52 PM.
Parallel Processing Sharing the load.
30 miles in 1 hour 30 mph 30 miles per hour
A B A B C D C D A B A B C D C D A B A B C D C D A B A B C D C D
Quiz Questions Parallel Programming Parallel Computing Potential
By Brandon, Ben, and Lee Parallel Computing.
Performance Cycle time of a computer CPU speed speed = 1 / cycle time
Performance Models And Evaluation
Quiz Questions Parallel Programming Parallel Computing Potential
Quiz Questions Parallel Programming Parallel Computing Potential
Quiz Questions Parallel Programming Parallel Computing Potential
Parent Mr. Mangham has 45 Little Debbies at his house. He needs to take some up to school, 15 miles away, to share with his wonderful students, but his.
Unit 4 Motion & Forces.
Performance Measurement and Analysis
Presentation transcript:

copyright Hesham El-Rewini1 Speedup S = Speed(new) / Speed(old) S = Work/time(new) / Work/time(old) S = time(old) / time(new) S = time(before improvement) / time(after improvement)

copyright Hesham El-Rewini2 Speedup Time (one CPU): T(1) Time (n CPUs): T(n) Speedup: S S = T(1)/T(n)

copyright Hesham El-Rewini3 Amdahl’s Law The performance improvement to be gained from using some faster mode of execution is limited by the fraction of the time the faster mode can be used

copyright Hesham El-Rewini4 20 hours 200 miles A B Walk 4 miles /hour Bike 10 miles / hour Car-1 50 miles / hour Car miles / hour Airplane 600 miles /hour must walk Example

copyright Hesham El-Rewini5 20 hours 200 miles A B Walk 4 miles /hour  = 70 hours S = 1 Bike 10 miles / hour  = 40 hours S = 1.75 Car-1 50 miles / hour  = 24 hours S = 2.9 Car miles / hour  = hours S = 3.2 Airplane 600miles /hour  = hours S = 3.4 must walk Example

copyright Hesham El-Rewini6 Amdahl’s Law (1967)  : The fraction of the program that is naturally serial (1-  ): The fraction of the program that is naturally parallel

copyright Hesham El-Rewini7 S = T(1)/T(N) T(N) = T(1)  + T(1)(1-  ) N S = 1  + (1-  ) N = N  N + (1-  )

copyright Hesham El-Rewini8 Amdahl’s Law