1 Lecture : Concurrency: Mutual Exclusion and Synchronization Operating System Spring 2008.

Slides:



Advertisements
Similar presentations
Chapter 5 Concurrency: Mutual Exclusion and Synchronization Operating Systems: Internals and Design Principles, 6/E William Stallings Patricia Roy Manatee.
Advertisements

1 Chapter 5 Concurrency: Mutual Exclusion and Synchronization Principals of Concurrency Mutual Exclusion: Hardware Support Semaphores Readers/Writers Problem.
Concurrency: Mutual Exclusion and Synchronization Chapter 5.
Ch. 7 Process Synchronization (1/2) I Background F Producer - Consumer process :  Compiler, Assembler, Loader, · · · · · · F Bounded buffer.
Chapter 6 Process Synchronization Bernard Chen Spring 2007.
Chapter 6: Process Synchronization
Silberschatz, Galvin and Gagne ©2013 Operating System Concepts – 9 th Edition Chapter 5: Process Synchronization.
5.1 Silberschatz, Galvin and Gagne ©2009 Operating System Concepts with Java – 8 th Edition Chapter 5: CPU Scheduling.
Silberschatz, Galvin and Gagne ©2009 Operating System Concepts – 8 th Edition, Chapter 6: Process Synchronization.
Process Synchronization. Module 6: Process Synchronization Background The Critical-Section Problem Peterson’s Solution Synchronization Hardware Semaphores.
CH7 discussion-review Mahmoud Alhabbash. Q1 What is a Race Condition? How could we prevent that? – Race condition is the situation where several processes.
Chapter 5 Concurrency: Mutual Exclusion and Synchronization Operating Systems: Internals and Design Principles, 6/E William Stallings Patricia Roy Manatee.
1 Concurrency: Mutual Exclusion and Synchronization Chapter 5.
Chapter 5 Concurrency: Mutual Exclusion and Synchronization
Silberschatz, Galvin and Gagne  Operating System Concepts Chapter 7: Process Synchronization Background The Critical-Section Problem Synchronization.
Chapter 6: Process Synchronization. Outline Background Critical-Section Problem Peterson’s Solution Synchronization Hardware Semaphores Classic Problems.
Chapter 6: Process Synchronization. 6.2 Silberschatz, Galvin and Gagne ©2005 Operating System Concepts – 7 th Edition, Feb 8, 2005 Objectives Understand.
Chapter 2.3 : Interprocess Communication
1 Chapter 6: Concurrency: Mutual Exclusion and Synchronization Operating System Spring 2007 Chapter 6 of textbook.
Chapter 6: Synchronization. 6.2 Silberschatz, Galvin and Gagne ©2005 Operating System Principles Module 6: Synchronization 6.1 Background 6.2 The Critical-Section.
Process Synchronization Topics: 1.Background 2.The critical-section problem 3.Semaphores 4.Critical Regions 5.Monitors Topics: 1.Background 2.The critical-section.
Instructor: Umar KalimNUST Institute of Information Technology Operating Systems Process Synchronization.
Chapter 6: Process Synchronization. 6.2 Silberschatz, Galvin and Gagne ©2005 Operating System Concepts – 7 th Edition, Feb 8, 2005 Module 6: Process Synchronization.
Adopted from and based on Textbook: Operating System Concepts – 8th Edition, by Silberschatz, Galvin and Gagne Updated and Modified by Dr. Abdullah Basuhail,
Operating Systems CSE 411 CPU Management Oct Lecture 13 Instructor: Bhuvan Urgaonkar.
Chapter 6: Process Synchronization. 6.2 Silberschatz, Galvin and Gagne ©2005 Operating System Concepts – 7 th Edition, Feb 8, 2005 Background Concurrent.
Silberschatz, Galvin and Gagne ©2009 Operating System Concepts – 8 th Edition, Chapter 6: Process Synchronization.
Concurrency, Mutual Exclusion and Synchronization.
Process Synchronization Continued 7.2 Critical-Section Problem 7.3 Synchronization Hardware 7.4 Semaphores.
6.3 Peterson’s Solution The two processes share two variables: Int turn; Boolean flag[2] The variable turn indicates whose turn it is to enter the critical.
1 Chapter 6: Process Synchronization Background The Critical-Section Problem Peterson’s Solution Special Machine Instructions for Synchronization Semaphores.
Silberschatz, Galvin and Gagne  Operating System Concepts Chapter 7: Process Synchronization Background The Critical-Section Problem Synchronization.
Chap 6 Synchronization. Background Concurrent access to shared data may result in data inconsistency Maintaining data consistency requires mechanisms.
Chapter 6: Synchronization. 6.2 Silberschatz, Galvin and Gagne ©2005 Operating System Concepts Module 6: Synchronization Background The Critical-Section.
Concurrency: Mutual Exclusion and Synchronization Chapter 5.
1 Concurrency: Mutual Exclusion and Synchronization Chapter 5.
Silberschatz, Galvin and Gagne ©2013 Operating System Concepts Essentials – 9 th Edition Chapter 5: Process Synchronization.
1 Lecture 8: Concurrency: Mutual Exclusion and Synchronization(cont.) Advanced Operating System Fall 2009.
Chapter 6: Process Synchronization. 6.2 Silberschatz, Galvin and Gagne ©2005 Operating System Concepts Module 6: Process Synchronization Background The.
Chapter 6: Process Synchronization. Module 6: Process Synchronization Background The Critical-Section Problem Peterson’s Solution Synchronization Hardware.
Background Concurrent access to shared data may result in data inconsistency Maintaining data consistency requires mechanisms to ensure the orderly execution.
1 Lecture 8: Concurrency: Mutual Exclusion and Synchronization Advanced Operating System Fall 2012.
Computer Architecture and Operating Systems CS 3230: Operating System Section Lecture OS-5 Process Synchronization Department of Computer Science and Software.
Synchronicity Introduction to Operating Systems: Module 5.
Operating Systems CSE 411 CPU Management Dec Lecture Instructor: Bhuvan Urgaonkar.
Silberschatz, Galvin and Gagne ©2009 Operating System Concepts – 8 th Edition, Chapter 6: Process Synchronization.
CS4315A. Berrached:CMS:UHD1 Process Synchronization Chapter 8.
Process Synchronization CS 360. Slide 2 CS 360, WSU Vancouver Process Synchronization Background The Critical-Section Problem Synchronization Hardware.
Silberschatz, Galvin and Gagne  2002 Modified for CSCI 399, Royden, Operating System Concepts Operating Systems Lecture 22 Semaphores Classic.
Process Synchronization. Objectives To introduce the critical-section problem, whose solutions can be used to ensure the consistency of shared data To.
Silberschatz, Galvin and Gagne ©2009 Operating System Concepts – 8 th Edition Chapter 6: Process Synchronization.
Chapter 6 Synchronization Dr. Yingwu Zhu. The Problem with Concurrent Execution Concurrent processes (& threads) often access shared data and resources.
6.1 Silberschatz, Galvin and Gagne ©2009 Operating System Concepts – 8 th Edition Chapter 6: Synchronization Background The Critical-Section Problem Peterson’s.
Process Synchronization
Chapter 5: Process Synchronization
Chapter 5: Process Synchronization
Chapter 5: Process Synchronization
Chapter 6: Process Synchronization
Chapter 6: Process Synchronization
Chapter 5: Process Synchronization
Topic 6 (Textbook - Chapter 5) Process Synchronization
Process Synchronization
Module 7a: Classic Synchronization
Lecture 2 Part 2 Process Synchronization
Critical section problem
Grades.
Concurrency: Mutual Exclusion and Process Synchronization
Chapter 6: Process Synchronization
Chapter 6: Synchronization Tools
Lecture 6: Concurrency: Mutual Exclusion and Synchronization
Presentation transcript:

1 Lecture : Concurrency: Mutual Exclusion and Synchronization Operating System Spring 2008

2 Concurrency An OS has many concurrent processes that run in parallel but share common access Race Condition: A situation where several processes access and manipulate the same data concurrently and the outcome of the execution depends on the particular order in which the access takes place.

3 Example for Race condition Suppose a customer wants to book a seat on UAL 56. Ticket agent will check the #-of-seats. If it is greater than 0, he will grab a seat and decrement #-of-seats by 1. UAL 56: #-of-seats=12 Main memory Terminal … Ticket Agent 1 Ticket Agent 2 Ticket Agent n

4 Example for Race condition(cont.) Ticket Agent 1 P1: LOAD #-of-seats P2: DEC 1 P3: STORE #-of-seats Ticket Agent 2 Q1: LOAD #-of-seats Q2: DEC 1 Q3: STORE #-of-seats Ticket Agent 3 R1: LOAD #-of-seats R2: DEC 1 R3: STORE #-of-seats Suppose, initially, #-of-seats=12 Suppose instructions are interleaved as P1,Q1,R1,P2,Q2,R2,P3,Q3,R3 The result would be #-of-seats=11, instead of 9 To solve the above problem, we must make sure that: P1,P2,P3 must be completely executed before we execute Q1 or R1, or Q1,Q2,Q3 must be completely executed before we execute P1 or R1, or R1,R2,R3 must be completely executed before we execute P1 or Q1.

5 Critical Section Problem Goal: To program the processes so that, at any moment of time, at most one of the processes is in its critical section. Prefix 0 CS 0 Suffix 0 P0P0 Prefix 1 CS 1 Suffix 1 P1P1 Prefix n-1 CS n-1 Suffix n-1 P n-1 … Critical section: a segment of code in which the process may be changing common variables, updating a table, writing a file, and so on.

6 Solution to Critical-Section Problem Any facility to provide support for mutual exclusion should meet the following requirements: 1. Mutual exclusion must be enforced: Only one process at a time is allowed into its critical section 2. A process that halts in its noncritical section must do so without interfering with other processes. 3. A process waiting to enter its critical section cannot be delayed infinitely 4. When no process is in a critical section, any process that requests entry to its critical section must be permitted to enter without delay. 5. No assumption are made about the relative process speeds or the number of processors. 6. A process remains inside its critical section for a finite time only.

7 Three Environments 1. There is no central program to coordinate the processes. The processes communicate with each other through global variable. 2. Special hardware instructions 3. There is a central program to coordinate the processes.

8 Three Environments 1. There is no central program to coordinate the processes. The processes communicate with each other through global variable. 2. Special hardware instructions 3. There is a central program to coordinate the processes.

9 1 st Attempt Start with just 2 processes, P 0 and p 1 Global variable turn, initially turn=0 Prefix 0 While (turn  0) do {} CS 0 turn=1 suffix 0 Prefix 1 While (turn  1) do {} CS 1 turn=0 suffix 1 The processes take turn to enter its critical section If turn=0, P 0 enters If turn=1, P 1 enters This solution guarantees mutual exclusion. But the drawback is that the method is not fair, because P0 is priviledged. Worse yet, until P0 executed its CS, P1 is blocked.

10 2 st Attempt Global variable flag[0] and flag[1], initially flag[0] and flag[1] are both false Prefix 0 While (flag[1]) do {} flag[0]=true CS 0 flag[0]=false suffix 0 Prefix 1 While (flag[0]) do {} flag[1]=true CS 1 flag[1]= false suffix 1 If P 0 is in critical section, flag[0] is true; If P 1 is in critical section, flag[1] is true If one process leaves the system, it will not block the other process. However, mutual exclusion is not guaranteed. P 0 executes the while statement and finds that flag[1] is false; P 1 executes the while statement and finds that flag[0] is false. P 0 sets flag[0] to true and enters its critical section; P 1 sets flag[1] to true and enters its critical section.

11 3 st Attempt Global variable flag[0] and flag[1], initially flag[0] and flag[1] are both false Prefix 0 flag[0]=true While (flag[1]) do {} CS 0 flag[0]=false suffix 0 Prefix 1 flag[1]=true While (flag[0]) do {} CS 1 flag[1]= false suffix 1 If P 0 is in critical section, flag[0] is true; If P 1 is in critical section, flag[1] is true Guarantees mutual exclusion. But mutual blocking can occur. P 0 sets flag[0] to be true; P 1 sets flag[1] to be true; Both will be hung in the while loop.

12 4 st Attempt Global variable flag[0] and flag[1], initially flag[0] and flag[1] are both false Prefix 0 L0: flag[0]=true If (flag[1]) then { flag[0]=false; goto L0} CS 0 flag[0]=false suffix 0 Prefix 1 L1: flag[1]=true If (flag[0]) then { flag[1]=false; goto L1}} CS 1 flag[1]= false suffix 1 Guarantees mutual exclusion. mutual blocking can occur if they execute at the same speed.

13 Correct Solution (Dekker’s Alg) – The first correct mutual exclusion alg (early 1960’s) Initially, flag[0]=flag[1]=false; turn=0 Prefix 0 flag[0]=true while (flag[1]) do {if (turn=1){ flag[0]=false; while(turn=1) do{} flag[0]=true; } CS 0 turn=1 flag[0]=false suffix 0 Prefix 1 flag[1]=true while (flag[0]) do {if (turn=0){ flag[1]=false; while(turn=0) do{} flag[1]=true; } CS 1 turn=0 flag[1]=false suffix 1

14 Peterson’s Algorithm for 2 processes – The simplest and most compact mutual exclusion alg. Initially, flag[0]=flag[1]=false Prefix 0 flag[0]=true turn=1 while (flag[1] and turn=1) do{} CS 0 flag[0]=false suffix 0 Prefix 1 flag[1]=true turn=0 while (flag[0] and turn=0) do{} CS 1 flag[1]=false suffix 1

15 Solution for n processes Global Variable 1. Flag[0..n-1] – array of size n. 2. Turn. Initially, Turn=some no. between 0 and n-1 Idle if Pi is outside Csi Want-in if Pi wants to be in CSi in-CS if Pi is in CSi Flag[i]=

16 Solutions for n processes Pi Prefix i Repeat Flag[i]=want-in; j=Turn; while j  i do {if Flag[j]  idle then j=Turn else j=(j+1) mod n} Flag[i]=in-CS j=0 while (j<n) and (j=i or Flag[j]  in-CS) do {j=j+1} Until (j  n) and (Turn=i or Flag[Turn]=idle) Turn=i; CS i j=(Turn+1)mod n While (j  Turn) and (Flag[j]=idle) do{j=(i+1) mod n} Turn=j Flag[i]=idle

17 Three Environments 1. There is no central program to coordinate the processes. The processes communicate with each other through global variable. 2. Special hardware instructions 3. There is a central program to coordinate the processes.

18 Hardware Support Disable interrupt CS Enable interrupt Won’t work if we have multiprocessors

19 Special Machine Instructions Modern machines provide special atomic hardware instructions  Atomic = non-interruptable Either test memory word and set value Or swap contents of two memory words

20 TS – Test and Set Boolean TS(i)= true if i=0; it will also set i to 1 false if i=1 Initially, lock=0 P i Prefix i While(¬ TS(lock)) do {} CS i Lock=0 suffix i It is possible that a process may starve if 2 processes enter the critical section arbitrarily often.

21 Three Environments 1. There is no central program to coordinate the processes. The processes communicate with each other through global variable. 2. Special hardware instructions 3. There is a central program to coordinate the processes.

22 Semaphores A variable that has an integer value upon which 3 operations are defined. Three operations: 1. A semaphore may be initialized to a nonnegative value 2. The wait operation decrements the semaphore value. If the value becomes negative, then the process executing the wait is blocked 3. The signal operation increments the semaphore value. If the value is not positive, then a process blocked by a wait operation is unblocked. Other than these 3 operations, there is no way to inspect or manipulate semaphores.

23 Wait(s) and Signal(s) Wait(s) – is also called P(s) { s=s-1; if (s<0) {place this process in a waiting queue} } Signal(s) – is also called V(s) { s=s+1; if(s  0) {remove a process from the waiting queue} }

24 Semaphore as General Synchronization Tool Counting semaphore – integer value can range over an unrestricted domain Binary semaphore – integer value can range only between 0 and 1; can be simpler to implement Also known as mutex locks Wait B(s) s is a binary semaphore { if s=1 then s=0 else block this process } Signal B(s) { if there is a blocked process then unblock a process else s=1 } Can implement a counting semaphore S as a binary semaphore

25 Note The wait and signal primitives are assumed to be atomic; they cannot be interrupted and each routine can be treated as an indivisible step.

26 Two classical examples Producer and Consumer Problem Readers/Writers Problem

27 Two classical examples Producer and Consumer Problem Readers/Writers Problem

28 Producer and Consumer Problem Producer can only put something in when there is an empty buffer Consumer can only take something out when there is a full buffer Producer and consumer are concurrent processes … 0 N-1 N buffers producer consumer

29 Producer and Consumer Problem(cont.) Global Variable 1. B[0..N-1] – an array of size N (Buffer) 2. P – a semaphore, initialized to N 3. C – a semaphore, initialized to 0 Local Variable 1. In – a ptr(integer) used by the producer, in=0 initially 2. Out – a ptr(integer) used by the consumer, out=0 initially Producer Process producer: produce(w) wait(p) B[in]=w in=(in+1)mod N signal(c) goto producer Consumer Process consumer: wait(c) w=B[out] out=(out+1)mod N signal(p) consume(w) goto consumer W is a local buffer used by the producer to produce W is a local buffer used by the consumer to store the item to be consumed

30 Two classical examples Producer and Consumer Problem Readers/Writers Problem

31 Readers/Writers Problem Suppose a data object is to be shared among several concurrent processes. Some of these processes want only to read the data object, while others want to update (both read and write) Readers – Processes that read only Writers – processes that read and write If a reader process is using the data object, then other reader processes are allowed to use it at the same time. If a writer process is using the data object, then no other process (reader or writer) is allowed to use it simultaneously.

32 Solve Readers/Writers Problem using wait and signal primitives(cont.) Global Variable: Wrt is a binary semaphore, initialized to 1; Wrt is used by both readers and writers For Reader Processes: Mutex is a binary semaphore, initialized to 1;Readcount is an integer variable, initialized to 0 Mutex and readcount used by readers only Reader Processes Wait(mutex) Readcount=readcount+1 If readcount=1 then wait(wrt) Signal(mutex); … Reading is performed … Wait(mutex) Readcount=readcount-1 If readcount=0 then signal(wrt) Signal(mutex) Writer Processes Wait(wrt) … Writing is performed … Signal(wrt)

33 End of lecture 6 Thank you!