Académique Documents
Professionnel Documents
Culture Documents
Chapter 6
Background
Multiprocessing A B C A Multiprogramming
B B C A B C
C B
Maintaining data consistency requires mechanisms to ensure the orderly execution of cooperating processes
Bounded-Buffer
Shared data
#define BUFFER_SIZE 10 typedef struct { ... } item; item buffer[BUFFER_SIZE]; int in = 0; int out = 0; int count = 0;
Producer
while (true) { /* produce an item and put in nextProduced */ while (count == BUFFER_SIZE); // do nothing /* add one item */ buffer [in] = nextProduced; in = (in + 1) % BUFFER_SIZE; count++;
Consumer
while (true) { while (count == 0) ; // do nothing /* consume the item in nextConsumed */ nextConsumed = buffer[out]; out = (out + 1) % BUFFER_SIZE; count--; }
Race Condition
Suppose counter initially set to 5. Execution of counter++ and counter-consecutively should leave the value at 5. Concurrent execution could leave inconsistent data:
S0: producer execute register1 = count {register1 = 5} S1: producer execute register1 = register1 + 1 {register1 = 6} S2: consumer execute register2 = count {register2 = 5} S3: consumer execute register2 = register2 - 1 {register2 = 4} S4: producer execute count = register1 {count = 6 } S5: consumer execute count = register2 {count = 4}
Could end up with counter value of 4, 5 or 6 There is no way to predict the relative speed of process execution, so you cannot guarantee that one will finish before the other.
Assume atomic program instruction execution. Assume a = 7, b = 2 initially. Question: What are the possible ending values for a and b if P1 and P2 execute concurrently?
Race Condition
Race condition: The situation where several processes access and manipulate shared data concurrently. The final value of the shared data depends upon the order of process accessing the data. To prevent race conditions, concurrent processes must be synchronized. e.g. The statements count++; count--; must be performed atomically.
Atomic operation means an operation that completes in its entirety without interruption.
Atomic Operations
To understand a concurrent program, we need to know what the underlying indivisible operations are! Atomic Operation: an operation that always runs to completion or not at all
It is indivisible: it cannot be stopped in the middle and state cannot be modified by someone else in the middle Fundamental building block if no atomic operations, then have no way for threads to work together
On most machines, memory references and assignments (i.e. loads and stores) of words are atomic Many instructions are not atomic
Double-precision floating point store often not atomic VAX and IBM 360 had an instruction to copy a whole array
Correctness Requirements
Threaded programs must work for all interleavings of thread instruction sequences
Bugs of cooperating threads inherently non-deterministic and nonreproducible Really hard to debug unless carefully designed!
Example: Therac-25
Machine for radiation therapy
Software control of electron accelerator and electron beam/ Xray production Software control of dosage
Assume that each process executes at a nonzero speed No assumption concerning relative speed of the N processes
Algorithm 1
Shared variables: int turn; initially turn = 0 turn = i Pi can enter its critical section Process Pi do { while (turn != i) ; critical section turn = j; remainder section } while (1); Satisfies mutual exclusion, but not progress
Algorithm 2
Shared variables boolean flag[2]; initially flag [0] = flag [1] = false. flag [i] = true Pi ready to enter its critical section Process Pi do { flag[i] = true; while (flag[j]) ; critical section flag [i] = false; remainder section } while (1); Satisfies mutual exclusion, but not progress requirement.
Algorithm 3
Combined shared variables of algorithms 1 and 2. Process Pi do { flag [i]:= true; turn = j; while (flag [j] and turn == j) ; // do nothing critical section flag [i] = false; remainder section } while (1); Meets all three requirements; solves the critical-section problem for two processes.
Synchronization Hardware
Many systems provide hardware support for critical section code Uniprocessors could disable interrupts
Currently running code would execute without preemption Generally too inefficient on multiprocessor systems Operating systems using this not broadly scalable Either test memory word and set value Or swap contents of two memory words
We can use these special instructions to solve the critical section problem.
Lock
Suppose we have some sort of implementation of a lock (more in a moment).
Lock.Acquire() wait until lock is free, then grab Lock.Release() Unlock, waking up anyone waiting These must be atomic operations if two threads are waiting for the lock and both see its free, only one succeeds to grab the lock
Once again, section of code between Acquire() and Release() called a Critical Section
We are going to implement various higher-level synchronization primitives using atomic operations
Everything is pretty painful if only atomic primitives are load and store Need to provide primitives useful at user-level
TestAndndSet Instruction
Definition: boolean TestAndSet (boolean *target) { boolean rv = *target; *target = TRUE; return rv: }
critical section
// release lock lock = FALSE; remainder section }
Notes on TestAndSet
The previous algorithm satisfies mutual exclusion.
Swap Instruction
Definition: /*86/
void Swap (boolean *a, boolean *b) { boolean temp = *a; *a = *b; *b = temp: }
Shared Boolean variable lock initialized to FALSE; Each process has a local Boolean variable key. Solution: while (true) { key = TRUE; // acquire lock while ( key == TRUE) Swap (&lock, &key ); // critical section // release lock lock = FALSE;
//
}
remainder section
Satisfying Requirements
The previous algorithm satisfies the three requirements Mutual Exclusion: The first process to execute TestAndSet(lock) when lock is false, will set lock to true so no other process can enter the CS. Progress: When a process exits the CS, it either sets lock to false, or waiting[j] to false, allowing the next process to proceed. Bounded Waiting: When a process exits the CS, it examines all the other processes in the waiting array in a circular order. Any process waiting for CS will have to wait at most n-1 turns.
Simple explanation:
If lock is free, test&set reads 0 and sets value=1, so lock is now busy. It returns 0 so while exits. If lock is busy, test&set reads 1 and sets value=1 (no change). It returns 1, so while loop continues When we set value = 0, someone else can get lock
int guard = 0; int lock = FREE; Acquire() { Release() { // Short busy-wait time // Short busy-wait time while (TestAndSet(guard)); while (TestAndSet(guard)); if anyone on wait queue { if (lock == BUSY) { take thread off wait queue put thread on wait queue; Place on ready queue; go to sleep() & guard = 0; } else { } else { lock = FREE; lock = BUSY; } guard = 0; guard = 0; } } Note: sleep has to be sure to reset the guard variable Why cant we do it just before or just after the sleep?
Semaphores
Semaphores are a kind of generalized lock
First defined by Dijkstra in late 60s Main synchronization primitive used in original UNIX
Definition: a Semaphore has a non-negative integer value and supports the following two operations:
P(): an atomic operation that waits for semaphore to become positive, then decrements it by 1
Think of this as the wait() operation
V(): an atomic operation that increments the semaphore by 1, waking up a waiting P, if any
This of this as the signal() operation
Note that P() stands for proberen (to test) and V stands for verhogen (to increment) in Dutch
Value= 2 0 1
Advantages of Semaphores
Powerful--Can solve a variety of synchronization problems Simple solution for multiple processes.
Bounded-Buffer Problem
Just the producer-consumer problem
Size N buffers, each can hold one item Semaphore mutex initialized to the value 1 Semaphore full initialized to the value 0 Semaphore empty initialized to the value N.
while (true) { // produce an item wait (empty); wait (mutex); // add the item to the buffer signal (mutex); signal (full); }
Readers-Writers Problem
A data set is shared among a number of concurrent processes
Readers only read the data set; they do not perform any updates Writers can both read and write.
Problem allow multiple readers to read at the same time. Only one single writer can access the shared data at the same time.
Shared Data
Data set Semaphore mutex initialized to 1. Semaphore wrt initialized to 1. Integer readcount initialized to 0.
// writing is performed
signal (wrt) ; }
while (true) { wait (mutex) ; readcount ++ ; if (readcount == 1) wait (wrt) ; signal (mutex) // reading is performed
wait (mutex) ; readcount - - ; if (readcount == 0) signal (wrt) ; signal (mutex) ; }
Questions: Which processes does this method favor (readers or writers)? What potential problem does this algorithm have?
Dining-Philosophers Problem
Philosophers are seated around a table. They alternate thinking and eating. One chopstick is placed in between each pair of philosophers. A philosopher needs the chopsticks on both sides to eat. A philosopher can only pick up one chopstick at a time. We want to guarantee that we don't generate deadlock. We want to guarantee that no philosopher starves.
Dining-Philosophers Problem
Shared data
Bowl of rice (data set) Semaphore chopstick [5] initialized to 1
// eat
signal ( chopstick[i] ); signal (chopstick[ (i + 1) % 5] ); // think }
Dining-Philosophers Problem
Deadlock: If all philosophers decide to eat at the same time and pick up the chopstick on their left. None will be able to pick up a chopstick on their right. Possible Solutions: 1)Allow at most 4 philosophers at the table 2)Allow philosopher to pick up chopsticks only if both available. 3)Odd philosophers pick up left first then right. Even philosophers pick up right first then left.
Starvation: A philosopher may never have access to two chopsticks if the philosophers next to him are always eating.
Cleaner idea: Use locks for mutual exclusion and condition variables for scheduling constraints
Monitors
A high-level abstraction that provides a convenient and effective mechanism for process synchronization Definition: a lock and zero or more condition variables for managing concurrent access to shared data
Some languages like Java provide this natively Most others use actual locks and condition variables Only one process may be active within the monitor at a time monitor monitor-name { // shared variable declarations procedure P1 () { . } procedure Pn () {}
Initialization code ( .) { } }
}
Condition Variable: a queue of threads waiting for something inside a critical section
Key idea: make it possible to go to sleep inside critical section by atomically releasing lock at time we go to sleep
RemoveFromQueue() { lock.Acquire(); // Lock shared data item = queue.dequeue();// Get next item or null lock.Release(); // Release Lock return(item); // Might return null }
Condition Variables
How do we change the RemoveFromQueue() routine to wait until something is on the queue?
Could do this by keeping a count of the number of things on the queue (with semaphores), but error prone
Condition Variable: a queue of threads waiting for something inside a critical section
Key idea: allow sleeping inside critical section by atomically releasing lock at time we go to sleep
RemoveFromQueue() { lock.Acquire(); // Get Lock while (queue.isEmpty()) { dataready.wait(&lock); // If nothing, sleep } item = queue.dequeue(); // Get next item lock.Release(); // Release Lock return(item); }
Pthreads Synchronization
Pthreads API is OS-independent It provides:
mutex locks condition variables semaphores
Mutex (lock)
pthread_mutex_t mutex; const pthread_mutexattr_t attr; int status; status = pthread_mutex_init(&mutex,&attr);
Thread i
status = pthread_mutex_destroy(&mutex);
status = pthread_mutex_unlock(&mutex); status = pthread_mutex_lock(&mutex); - block status = pthread_mutex_trylock(&mutex); - unblock
Condition variables
int status; pthread_condition_t cond; const pthread_condattr_t attr; pthread_mutex mutex; status = pthread_cond_init(&cond,&attr); status = pthread_cond_destroy(&cond); status = pthread_cond_signal(&cond); status = pthread_cond_broadcast(&cond);
Condition variables
status = pthread_cond_wait(&cond,&mutex);
wait on a condition variable. First, a thread needs to get the lock of the mutex. Then it could check whether or not a condition is satisfied. If so, it could continue its work and unlock the mutex later. Otherwise, it would wait until some thread use pthread_signal to notify it that the condition is satisfied. It will release the mutex while waiting and automatically regain it when awaken.
Semaphore
int status, pshared; sem_t sem; unsigned int initial_value; status = sem_init(&sem, pshared, initial_value); status = sem_destroy(&sem); status = sem_post(&sem); /* signal*/