Introduction to Parallel Processing

Slides:



Advertisements
Similar presentations
Prepared 7/28/2011 by T. O’Neil for 3460:677, Fall 2011, The University of Akron.
Advertisements

SE-292 High Performance Computing
Computer Abstractions and Technology
1 Architectural Complexity: Opening the Black Box Methods for Exposing Internal Functionality of Complex Single and Multiple Processor Systems EECC-756.
Parallel Computers Chapter 1
Introduction to MIMD architectures
Room: E-3-31 Phone: Dr Masri Ayob TK 2123 COMPUTER ORGANISATION & ARCHITECTURE Lecture 4: Computer Performance.
Multiprocessors ELEC 6200: Computer Architecture and Design Instructor : Agrawal Name: Nam.

Introduction What is Parallel Algorithms? Why Parallel Algorithms? Evolution and Convergence of Parallel Algorithms Fundamental Design Issues.
Parallel Algorithms - Introduction Advanced Algorithms & Data Structures Lecture Theme 11 Prof. Dr. Th. Ottmann Summer Semester 2006.
 Parallel Computer Architecture Taylor Hearn, Fabrice Bokanya, Beenish Zafar, Mathew Simon, Tong Chen.
CPE 731 Advanced Computer Architecture Multiprocessor Introduction
1 Computer Science, University of Warwick Architecture Classifications A taxonomy of parallel architectures: in 1972, Flynn categorised HPC architectures.
Introduction to Parallel Processing Ch. 12, Pg
Advanced Computer Architectures
Computer performance.
1 Parallel computing and its recent topics. 2 Outline 1. Introduction of parallel processing (1)What is parallel processing (2)Classification of parallel.
High-Performance Computing 12.1: Concurrent Processing.
Department of Computer Science University of the West Indies.
2015/10/14Part-I1 Introduction to Parallel Processing.
Chapter 2 Parallel Architecture. Moore’s Law The number of transistors on a chip doubles every years. – Has been valid for over 40 years – Can’t.
Parallel Computing.
CS591x -Cluster Computing and Parallel Programming
CPS 4150 Computer Organization Fall 2006 Ching-Song Don Wei.
Pipelining and Parallelism Mark Staveley
Winter 2014Parallel Processing, Fundamental ConceptsSlide 1 Part I Fundamental Concepts.
Outline Why this subject? What is High Performance Computing?
Computer Architecture And Organization UNIT-II Flynn’s Classification Of Computer Architectures.
Lecture 3: Computer Architectures
3/12/2013Computer Engg, IIT(BHU)1 INTRODUCTION-1.
3/12/2013Computer Engg, IIT(BHU)1 PARALLEL COMPUTERS- 3.
Computer Architecture Lecture 24 Parallel Processing Ralph Grishman November 2015 NYU.
An Overview of Parallel Processing
LECTURE #1 INTRODUCTON TO PARALLEL COMPUTING. 1.What is parallel computing? 2.Why we need parallel computing? 3.Why parallel computing is more difficult?
CDA-5155 Computer Architecture Principles Fall 2000 Multiprocessor Architectures.
Computer Science and Engineering Parallel and Distributed Processing CSE 8380 April 28, 2005 Session 29.
Classification of parallel computers Limitations of parallel processing.
Lecture 13 Parallel Processing. 2 What is Parallel Computing? Traditionally software has been written for serial computation. Parallel computing is the.
Computer Architecture Furkan Rabee
Computer Organization and Architecture Lecture 1 : Introduction
These slides are based on the book:
Auburn University COMP8330/7330/7336 Advanced Parallel and Distributed Computing Parallel Hardware Dr. Xiao Qin Auburn.
Flynn’s Taxonomy Many attempts have been made to come up with a way to categorize computer architectures. Flynn’s Taxonomy has been the most enduring of.
Introduction to Parallel Computing: MPI, OpenMP and Hybrid Programming
Advanced Architectures
CHAPTER SEVEN PARALLEL PROCESSING © Prepared By: Razif Razali.
Introduction to Parallel Computing
CMSC 611: Advanced Computer Architecture
buses, crossing switch, multistage network.
What Exactly is Parallel Processing?
CS 147 – Parallel Processing
Flynn’s Classification Of Computer Architectures
Architecture & Organization 1
Team 1 Aakanksha Gupta, Solomon Walker, Guanghong Wang
CS775: Computer Architecture
What is Parallel and Distributed computing?
MIMD Multiple instruction, multiple data
Architecture & Organization 1
Chapter 17 Parallel Processing
Multiprocessors - Flynn’s taxonomy (1966)
Introduction to Multiprocessors
Course Description: Parallel Computer Architecture
buses, crossing switch, multistage network.
Overview Parallel Processing Pipelining
AN INTRODUCTION ON PARALLEL PROCESSING
Computer Evolution and Performance
Chapter 4 Multiprocessors
Parallelism Can we make it faster? 8-May-19.
COMPUTER ORGANIZATION AND ARCHITECTURE
Presentation transcript:

Introduction to Parallel Processing 2018/4/19 Part-I

Preface It has allowed hardware performance to continue its exponential growth. This trend is expected to continue in the near future. It has led to unprecedented hardware complexity and almost intolerable development costs. In computer designers' quest for user-friendliness, compactness, simplicity, high per­formance, low cost, and low power, parallel processing plays a key role. High-performance uniprocessors are becoming increasingly complex, expensive, and power-hungry. 2018/4/19 Part-I

Introduction to Parallelism WHY PARALLEL PROCESSING? In the past two decades, the performance of microprocessors has enjoyed an exponential growth. (a factor of 2 every 18 months, Moore's law) Increase in complexity of VLSI chips Introduction of, and improvements in, architectural features Moore's law seems to hold regardless of how one measures processor performance: counting the number of executed instructions per second (IPS), counting the number of floating-point operations per second (FLOPS), or using sophisticated benchmark suites that attempt to measure the processor's performance on real applications. 2018/4/19 Part-I

2018/4/19 Part-I

WHY PARALLEL PROCESSING? (cont’d) physical laws The most easily understood physical limit is that imposed by the finite speed of signal propagation along a wire. This is sometimes referred to as the speed-of-light argument. pipelining and memory-latency-hiding techniques. The speed-of-light argument suggests that once the above limit has been reached, the only path to improved performance is the use of multiple processors. (the same argument can be invoked to conclude that any parallel processor) 2018/4/19 Part-I

2018/4/19 Part-I

2018/4/19 Part-I

WHY PARALLEL PROCESSING? (cont’d) Who needs supercomputers with TFLOPS or PFLOPS performance? The motivations for parallel processing can be summarized as follows: Higher speed, or solving problems faster. Higher throughput, or solving more instances of given problems. Higher computational power, or solving larger problems. speed-up factor This book focuses on the interplay of architectural and algorithmic speed-up techniques. 2018/4/19 Part-I

2018/4/19 Part-I

A MOTIVATING EXAMPLE A major issue in devising a parallel algorithm for a given problem is the way in which the computational load is divided between the multiple processors. Problem: Prime number finding 2018/4/19 Part-I

Prime number finding Single Processor Multiprocessors (a possible solution, share memory) 2018/4/19 Part-I

2018/4/19 Part-I

Prime number finding (cont’d.) Multiprocessors (data parallel approach, distributed memory) 2018/4/19 Part-I

2018/4/19 Part-I

PARALLEL PROCESSING UPS AND DOWNS Parallel processing, in the literal sense of the term, is used in virtually every modern computer. overlap between instruction preparation and execution in a pipelined processor. multiple functional units multitasking very-long-instruction-word (VLIW) computers In this book, the term parallel processing is used in a restricted sense of having multiple (usually identical) processors for the main computation and not for the I/O or other peripheral activities. 2018/4/19 Part-I

The history of parallel processing The history of parallel processing has had its ups and downs with what appears to be a 20-year cycle. commercial 2018/4/19 Part-I

TYPES OF PARALLELISM: A TAXONOMY Parallel computers can be divided into two main categories of control flow and data flow. Control-flow parallel computers are essentially based on the same principles as the sequential or von Neumann computer. Data-flow parallel computers, sometimes referred to as "non-von Neumann“ (DNA computer) In 1966, M. 1. Flynn proposed a four-way classification of computer systems based on the notions of instruction streams and data streams. 2018/4/19 Part-I

2018/4/19 Part-I

2018/4/19 Part-I

Flynn classification SPMD and MPMD CISC, NUMA, PRAM, RISC, and VLIW. The MIMD category includes a wide class of computers. For this reason, in 1988, E. E. Johnson proposed a further classification of such machines based on their memory structure (global or distributed) and the mechanism used for communication/synchronization (shared variables or message passing). SPMD and MPMD CISC, NUMA, PRAM, RISC, and VLIW. 2018/4/19 Part-I

Share memory architecture Uniform memory access Cache only Share memory architecture 2018/4/19 Part-I

Distrusted memory architecture 2018/4/19 Part-I

2018/4/19 Part-I

2018/4/19 Part-I

2018/4/19 Part-I

2018/4/19 Part-I

2018/4/19 Part-I

2018/4/19 Part-I

2018/4/19 Part-I

Multistage interconnection network 2018/4/19 Part-I

2018/4/19 Part-I

ROADBLOCKS TO PARALLEL PROCESSING Grosch's law (computing power is proportional to the square of cost) Minsky's conjecture (speed-up is proportional to the logarithm of the number p of processors) The tyranny of Ie technology (uniprocessors will be just as fast) The tyranny of vector supercomputers (why bother with parallel processors?) The software inertia (billions of dollars worth of existing software) Amdahl's law 2018/4/19 Part-I

2018/4/19 Part-I

EFFECTIVENESS Of PARALLEL PROCESSING 2018/4/19 Part-I

2018/4/19 Part-I