Presentation is loading. Please wait.

Presentation is loading. Please wait.

CSE 153 Design of Operating Systems Winter 2015 Lecture 5: Synchronization.

Similar presentations


Presentation on theme: "CSE 153 Design of Operating Systems Winter 2015 Lecture 5: Synchronization."— Presentation transcript:

1 CSE 153 Design of Operating Systems Winter 2015 Lecture 5: Synchronization

2 Administrivia l Homework 1 posted today u Due in class next Tuesday l Hopefully you have started on project 1 by now? u Be prepared with questions for this weeks Lab u Read up on scheduling and synchronization in textbook and start early! u Need to read and understand a lot of code before starting to code u Students typically find this to be the hardest of the three projects CSE 153 – Lecture 5 – Synchronization2

3 Cooperation between Threads l Threads cooperate in multithreaded programs u To share resources, access shared data structures »Threads accessing a memory cache in a Web server u To coordinate their execution »One thread executes relative to another CSE 153 – Lecture 5 – Synchronization3

4 4 Threads: Sharing Data int num_connections = 0; web_server() { while (1) { int sock = accept(); thread_fork(handle_request, sock); } handle_request(int sock) { ++num_connections; Process request close(sock); }

5 CSE 153 – Lecture 5 – Synchronization5 Threads: Cooperation l Threads voluntarily give up the CPU with thread_yield while (1) { printf(“ping\n”); thread_yield(); } while (1) { printf(“pong\n”); thread_yield(); } Ping ThreadPong Thread

6 January 21, 2014CSE 153 – Lecture 5 – Synchronization6 Synchronization l For correctness, we need to control this cooperation u Threads interleave executions arbitrarily and at different rates u Scheduling is not under program control l We control cooperation using synchronization u Synchronization enables us to restrict the possible inter- leavings of thread executions l Discuss in terms of threads, also applies to processes

7 January 21, 2014CSE 153 – Lecture 5 – Synchronization7 Shared Resources We initially focus on coordinating access to shared resources l Basic problem u If two concurrent threads are accessing a shared variable, and that variable is read/modified/written by those threads, then access to the variable must be controlled to avoid erroneous behavior l Over the next couple of lectures, we will look at u Mechanisms to control access to shared resources »Locks, mutexes, semaphores, monitors, condition variables, etc. u Patterns for coordinating accesses to shared resources »Bounded buffer, producer-consumer, etc.

8 CSE 153 – Lecture 5 – Synchronization8 Classic Example l Suppose we have to implement a function to handle withdrawals from a bank account: withdraw (account, amount) { balance = get_balance(account); balance = balance – amount; put_balance(account, balance); return balance; } l Now suppose that you and your father share a bank account with a balance of $1000 l Then you each go to separate ATM machines and simultaneously withdraw $100 from the account

9 CSE 153 – Lecture 5 – Synchronization9 Example Continued l We’ll represent the situation by creating a separate thread for each person to do the withdrawals l These threads run on the same bank machine: l What’s the problem with this implementation? u Think about potential schedules of these two threads withdraw (account, amount) { balance = get_balance(account); balance = balance – amount; put_balance(account, balance); return balance; } withdraw (account, amount) { balance = get_balance(account); balance = balance – amount; put_balance(account, balance); return balance; }

10 CSE 153 – Lecture 5 – Synchronization10 Interleaved Schedules l The problem is that the execution of the two threads can be interleaved: l What is the balance of the account now? balance = get_balance(account); balance = balance – amount; balance = get_balance(account); balance = balance – amount; put_balance(account, balance); Execution sequence seen by CPU Context switch

11 CSE 153 – Lecture 5 – Synchronization11 Shared Resources l Problem: two threads accessed a shared resource u Known as a race condition (memorize this buzzword) l Need mechanisms to control this access u So we can reason about how the program will operate l Our example was updating a shared bank account l Also necessary for synchronizing access to any shared data structure u Buffers, queues, lists, hash tables, etc.

12 CSE 153 – Lecture 5 – Synchronization12 When Are Resources Shared? l Local variables? u Not shared: refer to data on the stack u Each thread has its own stack u Never pass/share/store a pointer to a local variable on the stack for thread T1 to another thread T2 l Global variables and static objects? u Shared: in static data segment, accessible by all threads l Dynamic objects and other heap objects? u Shared: Allocated from heap with malloc/free or new/delete Stack (T1) Code Static Data Heap Stack (T2) Stack (T3) Thread 3 Thread 2 PC (T1) PC (T3) PC (T2) Thread 1

13 CSE 153 – Lecture 5 – Synchronization13 How Interleaved Can It Get?............... get_balance(account); put_balance(account, balance); balance = balance – amount; balance = get_balance(account); balance =................................... How contorted can the interleavings be? l We'll assume that the only atomic operations are reads and writes of individual memory locations u Some architectures don't even give you that! l We'll assume that a context switch can occur at any time l We'll assume that you can delay a thread as long as you like as long as it's not delayed forever

14 CSE 153 – Lecture 5 – Synchronization14 Mutual Exclusion l Mutual exclusion to synchronize access to shared resources u This allows us to have larger atomic blocks u What does atomic mean? l Code that uses mutual called a critical section u Only one thread at a time can execute in the critical section u All other threads are forced to wait on entry u When a thread leaves a critical section, another can enter u Example: sharing an ATM with others l What requirements would you place on a critical section?

15 CSE 153 – Lecture 5 – Synchronization15 Critical Section Requirements Critical sections have the following requirements: 1) Mutual exclusion (mutex) u If one thread is in the critical section, then no other is 2) Progress u A thread in the critical section will eventually leave the critical section u If some thread T is not in the critical section, then T cannot prevent some other thread S from entering the critical section 3) Bounded waiting (no starvation) u If some thread T is waiting on the critical section, then T will eventually enter the critical section 4) Performance u The overhead of entering and exiting the critical section is small with respect to the work being done within it

16 CSE 153 – Lecture 5 – Synchronization16 About Requirements There are three kinds of requirements that we'll use l Safety property: nothing bad happens u Mutex l Liveness property: something good happens u Progress, Bounded Waiting l Performance requirement u Performance l Properties hold for each run, while performance depends on all the runs u Rule of thumb: When designing a concurrent algorithm, worry about safety first (but don't forget liveness!).

17 CSE 153 – Lecture 5 – Synchronization17 Mechanisms For Building Critical Sections l Atomic read/write u Can it be done? l Locks u Primitive, minimal semantics, used to build others l Semaphores u Basic, easy to get the hang of, but hard to program with l Monitors u High-level, requires language support, operations implicit

18 CSE 153 – Lecture 5 – Synchronization18 Mutual Exclusion with Atomic Read/Writes: First Try while (true) { while (turn != 1) ; critical section turn = 2; outside of critical section } while (true) { while (turn != 2) ; critical section turn = 1; outside of critical section } int turn = 1; This is called alternation It satisfies mutex: If blue is in the critical section, then turn == 1 and if yellow is in the critical section then turn == 2 (turn == 1) ≡ (turn != 2) Is there anything wrong with this solution?

19 CSE 153 – Lecture 5 – Synchronization19 Mutual Exclusion with Atomic Read/Writes: First Try while (true) { while (turn != 1) ; critical section turn = 2; outside of critical section } while (true) { while (turn != 2) ; critical section turn = 1; outside of critical section } int turn = 1; This is called alternation It satisfies mutex: If blue is in the critical section, then turn == 1 and if yellow is in the critical section then turn == 2 (turn == 1) ≡ (turn != 2) It violates progress: blue thread could go into an infinite loop outside of the critical section, which will prevent the yellow one from entering

20 CSE 153 – Lecture 5 – Synchronization20 Mutex with Atomic R/W: Peterson's Algorithm while (true) { try1 = true; turn = 2; while (try2 && turn != 1) ; critical section try1 = false; outside of critical section } while (true) { try2 = true; turn = 1; while (try1 && turn != 2) ; critical section try2 = false; outside of critical section } int turn = 1; bool try1 = false, try2 = false; This satisfies all the requirements Here's why...

21 CSE 153 – Lecture 5 – Synchronization21 Mutex with Atomic R/W: Peterson's Algorithm while (true) { {¬ try1 ∧ (turn == 1 ∨ turn == 2) } 1 try1 = true; { try1 ∧ (turn == 1 ∨ turn == 2) } 2 turn = 2; { try1 ∧ (turn == 1 ∨ turn == 2) } 3 while (try2 && turn != 1) ; { try1 ∧ (turn == 1 ∨ ¬ try2 ∨ (try2 ∧ (yellow at 6 or at 7)) } critical section 4 try1 = false; {¬ try1 ∧ (turn == 1 ∨ turn == 2) } outside of critical section } while (true) { {¬ try2 ∧ (turn == 1 ∨ turn == 2) } 5 try2 = true; { try2 ∧ (turn == 1 ∨ turn == 2) } 6 turn = 1; { try2 ∧ (turn == 1 ∨ turn == 2) } 7 while (try1 && turn != 2) ; { try2 ∧ (turn == 2 ∨ ¬ try1 ∨ (try1 ∧ (blue at 2 or at 3)) } critical section 8 try2 = false; {¬ try2 ∧ (turn == 1 ∨ turn == 2) } outside of critical section } int turn = 1; bool try1 = false, try2 = false; (blue at 4) ∧ try1 ∧ (turn == 1 ∨ ¬ try2 ∨ (try2 ∧ (yellow at 6 or at 7)) ∧ (yellow at 8) ∧ try2 ∧ (turn == 2 ∨ ¬ try1 ∨ (try1 ∧ (blue at 2 or at 3))... ⇒ (turn == 1 ∧ turn == 2)

22 CSE 153 – Lecture 5 – Synchronization22 Locks l A lock is an object in memory providing two operations u acquire(): before entering the critical section u release(): after leaving a critical section l Threads pair calls to acquire() and release() u Between acquire()/release(), the thread holds the lock u acquire() does not return until any previous holder releases u What can happen if the calls are not paired?

23 CSE 153 – Lecture 5 – Synchronization23 Using Locks u Why is the “return” outside the critical section? Is this ok? u What happens when a third thread calls acquire? withdraw (account, amount) { acquire(lock); balance = get_balance(account); balance = balance – amount; put_balance(account, balance); release(lock); return balance; } acquire(lock); balance = get_balance(account); balance = balance – amount; balance = get_balance(account); balance = balance – amount; put_balance(account, balance); release(lock); acquire(lock); put_balance(account, balance); release(lock); Critical Section

24 CSE 153 – Lecture 5 – Synchronization24 l How do we implement locks? Here is one attempt: l This is called a spinlock because a thread spins waiting for the lock to be released l Does this work? Implementing Locks (1) struct lock { int held = 0; } void acquire (lock) { while (lock->held); lock->held = 1; } void release (lock) { lock->held = 0; } busy-wait (spin-wait) for lock to be released

25 CSE 153 – Lecture 5 – Synchronization25 Implementing Locks (2) l No. Two independent threads may both notice that a lock has been released and thereby acquire it. struct lock { int held = 0; } void acquire (lock) { while (lock->held); lock->held = 1; } void release (lock) { lock->held = 0; } A context switch can occur here, causing a race condition

26 CSE 153 – Lecture 5 – Synchronization26 Implementing Locks (3) l The problem is that the implementation of locks has critical sections, too l How do we stop the recursion? l The implementation of acquire/release must be atomic u An atomic operation is one which executes as though it could not be interrupted u Code that executes “all or nothing” l How do we make them atomic? l Need help from hardware u Atomic instructions (e.g., test-and-set) u Disable/enable interrupts (prevents context switches)

27 CSE 153 – Lecture 5 – Synchronization27 Atomic Instructions: Test-And-Set l The semantics of test-and-set are: u Record the old value u Set the value to indicate available u Return the old value l Hardware executes it atomically! l When executing test-and-set on “flag” u What is value of flag afterwards if it was initially False? True? u What is the return result if flag was initially False? True? bool test_and_set (bool *flag) { bool old = *flag; *flag = True; return old; }

28 CSE 153 – Lecture 5 – Synchronization28 Using Test-And-Set l Here is our lock implementation with test-and-set: l When will the while return? What is the value of held? struct lock { int held = 0; } void acquire (lock) { while (test-and-set(&lock->held)); } void release (lock) { lock->held = 0; }

29 CSE 153 – Lecture 5 – Synchronization29 Problems with Spinlocks l The problem with spinlocks is that they are wasteful l If a thread is spinning on a lock, then the scheduler thinks that this thread needs CPU and puts it on the ready queue l If N threads are contending for the lock, the thread which holds the lock gets only 1/N’th of the CPU

30 CSE 153 – Lecture 5 – Synchronization30 Disabling Interrupts l Another implementation of acquire/release is to disable interrupts: l Note that there is no state associated with the lock l Can two threads disable interrupts simultaneously? struct lock { } void acquire (lock) { disable interrupts; } void release (lock) { enable interrupts; }

31 CSE 153 – Lecture 5 – Synchronization31 On Disabling Interrupts l Disabling interrupts blocks notification of external events that could trigger a context switch (e.g., timer) l In a “real” system, this is only available to the kernel u Why? l Disabling interrupts is insufficient on a multiprocessor u Back to atomic instructions l Like spinlocks, only want to disable interrupts to implement higher-level synchronization primitives u Don’t want interrupts disabled between acquire and release

32 CSE 153 – Lecture 5 – Synchronization32 Summarize Where We Are l Goal: Use mutual exclusion to protect critical sections of code that access shared resources l Method: Use locks (spinlocks or disable interrupts) l Problem: Critical sections can be long acquire(lock) … Critical section … release(lock) Disabling Interrupts: l Should not disable interrupts for long periods of time l Can miss or delay important events (e.g., timer, I/O) Spinlocks: l Threads waiting to acquire lock spin in test-and-set loop l Wastes CPU cycles l Longer the CS, the longer the spin l Greater the chance for lock holder to be interrupted

33 CSE 153 – Lecture 5 – Synchronization33 Higher-Level Synchronization l Spinlocks and disabling interrupts are useful only for very short and simple critical sections u Wasteful otherwise u These primitives are “primitive” – don’t do anything besides mutual exclusion l Need higher-level synchronization primitives that: u Block waiters u Leave interrupts enabled within the critical section l All synchronization requires atomicity l So we’ll use our “atomic” locks as primitives to implement them

34 l Block waiters, interrupts enabled in critical sections void release (lock) { Disable interrupts; if (Q) remove and unblock a waiting thread; else lock->held = 0; Enable interrupts; } CSE 153 – Lecture 5 – Synchronization34 Implementing Locks (4) struct lock { int held = 0; queue Q; } void acquire (lock) { Disable interrupts; if (lock->held) { put current thread on lock Q; block current thread; } lock->held = 1; Enable interrupts; } acquire(lock) … Critical section … release(lock) Interrupts Enabled Interrupts Disabled

35 CSE 153 – Lecture 5 – Synchronization35 Next time… l Semaphores and monitors u Read Chapter 5.4 – 5.7 l Homework 1 due next Tuesday

36 CSE 153 – Lecture 5 – Synchronization36 Semaphores l Semaphores are data structures that also provide mutual exclusion to critical sections u Block waiters, interrupts enabled within CS u Described by Dijkstra in THE system in 1968 l Semaphores count the number of threads using a critical section (more later) l Semaphores support two operations: u wait(semaphore): decrement, block until semaphore is open »Also P() after the Dutch word for test u signal(semaphore): increment, allow another thread to enter »Also V() after the Dutch word for increment

37 CSE 153 – Lecture 5 – Synchronization37 Blocking in Semaphores l Associated with each semaphore is a queue of waiting processes l When wait() is called by a thread: u If semaphore is open, thread continues u If semaphore is closed, thread blocks on queue l signal() opens the semaphore: u If a thread is waiting on the queue, the thread is unblocked u If no threads are waiting on the queue, the signal is remembered for the next thread »In other words, signal() has “history” »This history uses a counter

38 CSE 153 – Lecture 5 – Synchronization38 Using Semaphores l Use is similar to locks, but semantics are different struct account { double balance; semaphore S; } withdraw (account, amount) { wait(account->S); tmp = account->balance; tmp = tmp - amount; account->balance = tmp; signal(account->S); return tmp; } wait(account->S); tmp = account->balance; tmp = tmp – amount; wait(account->S); account->balance = tmp; signal(account->S); wait(account->S); … signal(account->S); … signal(account->S); Threads block It is undefined which thread runs after a signal


Download ppt "CSE 153 Design of Operating Systems Winter 2015 Lecture 5: Synchronization."

Similar presentations


Ads by Google