MULTIPROCESSOR OPERATING SYSTEM

Slides:



Advertisements
Similar presentations
Threads, SMP, and Microkernels
Advertisements

© 2009 Fakultas Teknologi Informasi Universitas Budi Luhur Jl. Ciledug Raya Petukangan Utara Jakarta Selatan Website:
Prepared 7/28/2011 by T. O’Neil for 3460:677, Fall 2011, The University of Akron.
Parallel Programming Motivation and terminology – from ACM/IEEE 2013 curricula.
SGI’2000Parallel Programming Tutorial Supercomputers 2 With the acknowledgement of Igor Zacharov and Wolfgang Mertz SGI European Headquarters.
Classification of Distributed Systems Properties of Distributed Systems n motivation: advantages of distributed systems n classification l architecture.
Taxanomy of parallel machines. Taxonomy of parallel machines Memory – Shared mem. – Distributed mem. Control – SIMD – MIMD.
Chapter 4 Threads, SMP, and Microkernels Patricia Roy Manatee Community College, Venice, FL ©2008, Prentice Hall Operating Systems: Internals and Design.
Introduction to MIMD architectures
Background Computer System Architectures Computer System Software.
1 Threads, SMP, and Microkernels Chapter 4. 2 Process: Some Info. Motivation for threads! Two fundamental aspects of a “process”: Resource ownership Scheduling.
2. Multiprocessors Main Structures 2.1 Shared Memory x Distributed Memory Shared-Memory (Global-Memory) Multiprocessor:  All processors can access all.
Multiprocessors ELEC 6200: Computer Architecture and Design Instructor : Agrawal Name: Nam.
Chapter 17 Parallel Processing.
1 Lecture 23: Multiprocessors Today’s topics:  RAID  Multiprocessor taxonomy  Snooping-based cache coherence protocol.
 Parallel Computer Architecture Taylor Hearn, Fabrice Bokanya, Beenish Zafar, Mathew Simon, Tong Chen.
1 Pertemuan 25 Parallel Processing 1 Matakuliah: H0344/Organisasi dan Arsitektur Komputer Tahun: 2005 Versi: 1/1.
Distributed Computing Distributed computing deals with hardware and software systems containing more than one processing element or storage element, concurrent.
4. Multiprocessors Main Structures 4.1 Shared Memory x Distributed Memory Shared-Memory (Global-Memory) Multiprocessor:  All processors can access all.
DISTRIBUTED COMPUTING
Flynn’s Taxonomy of Computer Architectures Source: Wikipedia Michael Flynn 1966 CMPS 5433 – Parallel Processing.
Computer System Architectures Computer System Software
Parallel Computing Basic Concepts Computational Models Synchronous vs. Asynchronous The Flynn Taxonomy Shared versus Distributed Memory Interconnection.
Parallel Processing – Page 1 of 63CSCI 4717 – Computer Architecture CSCI 4717/5717 Computer Architecture Topic: Symmetric Multiprocessors & Clusters Reading:
1 Lecture 20: Parallel and Distributed Systems n Classification of parallel/distributed architectures n SMPs n Distributed systems n Clusters.
Spaces Dr. Samir Tartir Extracted from Principles of Concurrent and Distributed Programming, Second Edition By M. Ben-Ari.
Operating System 4 THREADS, SMP AND MICROKERNELS
What is a Distributed System? n From various textbooks: l “A distributed system is a collection of independent computers that appear to the users of the.
Chapter 4 Threads, SMP, and Microkernels Patricia Roy Manatee Community College, Venice, FL ©2008, Prentice Hall Operating Systems: Internals and Design.
Department of Computer Science University of the West Indies.
CSCI 232© 2005 JW Ryder1 Parallel Processing Large class of techniques used to provide simultaneous data processing tasks Purpose: Increase computational.
Processes and Threads Processes have two characteristics: – Resource ownership - process includes a virtual address space to hold the process image – Scheduling/execution.
1 Multiprocessor and Real-Time Scheduling Chapter 10 Real-Time scheduling will be covered in SYSC3303.
Heterogeneous Multikernel OS Yauhen Klimiankou BSUIR
Lecture 13: Multiprocessors Kai Bu
Advanced Principles of Operating Systems (CE-403).
Flynn’s Architecture. SISD (single instruction and single data stream) SIMD (single instruction and multiple data streams) MISD (Multiple instructions.
Spring 2003CSE P5481 Issues in Multiprocessors Which programming model for interprocessor communication shared memory regular loads & stores message passing.
PARALLEL PROCESSOR- TAXONOMY. CH18 Parallel Processing {Multi-processor, Multi-computer} Multiple Processor Organizations Symmetric Multiprocessors Cache.
Parallel Computing.
Operating System 4 THREADS, SMP AND MICROKERNELS.
Copyright © Clifford Neuman - UNIVERSITY OF SOUTHERN CALIFORNIA - INFORMATION SCIENCES INSTITUTE Advanced Operating Systems Lecture notes Dr.
Outline Why this subject? What is High Performance Computing?
Computer Architecture And Organization UNIT-II Flynn’s Classification Of Computer Architectures.
Lecture 3: Computer Architectures
Parallel Processing Presented by: Wanki Ho CS147, Section 1.
Multiprocessor  Use large number of processor design for workstation or PC market  Has an efficient medium for communication among the processor memory.
An Overview of Parallel Processing
Chapter 4 Threads, SMP, and Microkernels Dave Bremer Otago Polytechnic, N.Z. ©2008, Prentice Hall Operating Systems: Internals and Design Principles, 6/E.
Parallel Computing Presented by Justin Reschke
Page 1 2P13 Week 1. Page 2 Page 3 Page 4 Page 5.
Classification of parallel computers Limitations of parallel processing.
Multi Processing prepared and instructed by Shmuel Wimer Eng. Faculty, Bar-Ilan University June 2016Multi Processing1.
Lecture 13 Parallel Processing. 2 What is Parallel Computing? Traditionally software has been written for serial computation. Parallel computing is the.
Overview Parallel Processing Pipelining
CHAPTER SEVEN PARALLEL PROCESSING © Prepared By: Razif Razali.
CLASSIFICATION OF PARALLEL COMPUTERS
buses, crossing switch, multistage network.
CS 147 – Parallel Processing
Flynn’s Classification Of Computer Architectures
Overview Parallel Processing Pipelining
Different Architectures
Chapter 17 Parallel Processing
Symmetric Multiprocessing (SMP)
buses, crossing switch, multistage network.
Overview Parallel Processing Pipelining
Operating System 4 THREADS, SMP AND MICROKERNELS
CS 286 Computer Organization and Architecture
Presentation transcript:

MULTIPROCESSOR OPERATING SYSTEM MODULE 8

PRESENTORS: JAMES REYES RYANNE NERY P. BACUD

MOTIVATIONS FOR MULTIPROCESSOR SYSTEMS Enhanced Performance. Multiprocessor systems increase system performance Fault Tolerance. A multiprocessor system exhibits graceful performance degradation to processor failures because of the availability of multiple processors.

BASIC MULTIPROCESSOR SYSTEM ARCHITECTURES According to the classification of Flynn, in MIMD (multiple instruction multiple data) architectures, multiple instruction streams operate on different data streams. In the broadest sense, an MIMD architecture qualifies as a full-fledged multiprocessor system. Thus, a multiprocessor system consists of multiple processors, which execute different programs (or different segments of a program) concurrently. The main memory is typically shared by all the processors. Based on whether a memory location can be directly accessed by a processor or not, there are two type s of multiprocessor system

Two types of multiprocessor systems: Tightly coupled Loosely coupled (Based on whether a memory location can be directly accessed by a processor or not, ) NEXT>>

Tightly Coupled In tightly coupled systems, all processors share the same memory address space and all processors can directly access a global main memory. Examples of commercially available tightly coupled systems are Multimax of Encore Corporation, Flex/32 of Flexible Corporation, and FX of Sequent Computers. Tightly coupled systems can use the main memory for interprocessor communication and synchronization. <<BACK

Loosely Coupled Systems In loosely coupled systems, not only is the main memory partitioned and attached to processors, but each processor has its own address space. Therefore, a processor cannot directly access the memory attached to other processors. One example of a loosely coupled systems is Intel’s Hypercube. Loosely coupled systems, on the other hand, use only message passing for interprocessor communication and synchronization. <<BACK

TYPES OF MULTIPROCESSOR SYSTEM UMA (uniform memory access) NUMA (nonuniform memory access) NORMA (no remote memory access) (Based on the vicinity and accessibility of the main memory to the processors) NEXT>>

UMA architectures The main memory is located at a central location such that it is equidistant from all the processors in terms of access (in the absence of conflicts). That is, all the processors have the same access time to the main memory. In addition to this centralized shared memory, processors may also have private memories, where they can cache data for higher performance. Some examples of UMA architectures are Multimax of Encore Corporation, Balance of Sequent, and VAX 8800 of Digital Equipment. <<BACK

NUMA architectures The main memory is physically partitioned and the partitions are attached to the processors. All the processors, however, share the same memory address space. A processor can directly access the memory attached to any other processor, but the time to access its own memory partition. Examples of NUMA architectures are cm* of CMU and Butterfly machine of BBN Laboratories. <<BACK

NORMA Architectures The main memory is physically partitioned and the partitions are attached to the processors. However, a processor cannot directly access the memory of any other processor. The processors must send message over the interconnection network to exchange information. An example of NORMA architecture is Intel’s Hypercube. <<BACK

INTERCONNECTING NETWORKS FOR MULTIPROCESSOR SYSTEMS The interconnection network in multiprocessor systems provides data transfer facility between processors and memory modules for memory access. The design of the interconnection network is the most crucial hardware issue in the design of multiprocessor systems. Generally, circuit switching is issued to establish a connection between processors and memory modules. Thus, during a data transfer, a dedicated path exists between the processor and the memory module. NEXT>>

Various types of interconnection networks include: Bus Cross-bar Switch Multistage Interconnection Network NEXT>>

Bus Processors are connected to memory modules via a bus. <<BACK

Cross-bar Switch A cross bar switch is a matrix (or grid structure) that has a switch at every cross-point. A cross-bar is capable of providing an exclusive connection between any processor-memory pair. <<BACK

Multistage Interconnection Network A multistage interconnection network is a compromise between a bus and a cross-bar switch. A multistage interconnection network permits simultaneous connection between several processor-memory pairs and is more cost-effective than a cross-bar. <<BACK

CACHING Multiprocessor systems commonly use caching to reduce memory access time. Under caching, every processor has a private memory, called cache, in addition to the shared global memory. When a processor needs to fetch a word from a data block in the global memory, it fetches the entire block and saves it in its cache for future use. ADVANTAGE DISADVAN TAGE <<NEXT

Caching has two other advantages First, the traffic over the interconnection network is reduced (because most memory accesses can be satisfied by the cache). Second contention at memory modules is reduced (because different blocks of a memory module can be fetched by several processors and can be accessed concurrently from their respective caches). <<BACK

The Cache Coherence Problem Caching poses s problem when a processor modifies a cached block that is also cached by some other processors currently. Modifying a cached block by a processor invalidates the copies of this block in the cache of other processors because these copies have an outdated value. This is called the cache coherence problem in multiprocessor system. <<BACK

HYPERCUBE ARCHITECTURES Hypercube based architectures have recently emerged as a viable alternative in the design of multiprocessor systems with a large number of processors. In an n-dimensional cube, where each node is connected to n other nodes. In hypercube based architectures, the processors are the nodes of a hypercube and a hypercube edge corresponds to a bi-directional communication link between two processors. Each of the 2n nodes of an n-cube are assigned a unique n-bit address ranging from 0 to 2n – 1 such that the addresses assigned to two adjacent nodes differ only in 1 bit position. The address of a node in the ith dimension of a node differs from that node’s address only in ith bit. The maximum distance between any two nodes in an n-cube is n hops. Thus, the delay characteristics of hypercube architectures grow logarithmically with the number of nodes and these architecture are highly scalable.