This is not the current version of the class.

Synchronization Section 1: Threads and synchronization objects

In the first part of section, we’ll spend some time on some simple thread examples, working through ways of enforcing ordering through atomics, mutexes, and condition variables. Then we turn to the problem set, discussing different coarse-grained and fine-grained synchronization designs. After a discussion of fancy mutexes, there are some poop jokes.

Simultaneously enrolled college students are responsible for attending section live (in-person or via zoom). If you cannot attend any section live in a particular week, you may watch a recording (Canvas > Zoom > Cloud Recordings) instead. Please self-report your live attendance (or that you watched a recording) using this attendance form.

Three threads

Recall that std::thread's constructor takes the name of the function to run on the new thread and any arguments, like std::thread t(func, param1, ...). Given a thread t, t.join() blocks the current thread until thread t terminates. t.detach() sets the t thread free—it can no longer be joined. Each thread object must be detached or joined before it goes out of scope (is destroyed).

QUESTION. What can this code print, and how many threads might run concurrently during its execution?

int main() {
    for (int i = 0; i != 3; ++i) {
        std::thread t(printf, "%d\n", i + 1);
        t.join();
    }
}

QUESTION. What can this code print, and how many threads might run concurrently during its execution?

int main() {
    for (int i = 0; i != 3; ++i) {
        std::thread t(printf, "%d\n", i + 1);
        t.detach();
    }
}

QUESTION. What can this code print? (sample03.cc)

int main() {
    for (int i = 0; i != 3; ++i) {
        std::thread t(printf, "%d\n", i + 1);
    }
}

Counting to three

We now turn to the code in synchs1/counting.cc.

#include <thread>
#include <mutex>
#include <condition_variable>
#include <cstdio>

/* G1 */

void counter(int n) {
    /* C1 */
    printf("%d\n", n + 1);
    /* C2 */
}

int main() {
    /* M1 */
    std::thread t1(counter, 0);
    std::thread t2(counter, 1);
    std::thread t3(counter, 2);
    /* M2 */
    t3.join();
    t2.join();
    t1.join();
}

EXERCISE. What can this code print, and how many threads might run concurrently during its execution?

Counting with mutexes

In the remaining exercises in this section, we will synchronize counting.cc so that it always prints 1\n2\n3\n, in that order. (The resulting code might as well run in a single thread—all useful work is performed by one thread at a time—but that’s OK for our purposes.)

There must be no data races. You should add code only in the marked locations (G1, C1, C2, M1, and/or M2). In particular, you should not reorder the join calls.

This is a synchronization problem! Simultaneously-running threads, like t1, t2, and t3, can execute in any order. Forcing them to execute in a specific order requires coordination using using shared state. But when to safely access shared state requires synchronization.

DISCUSSION. A fundamental question in synchronization is the amount of shared state required (independent of the synchronization objects required to access that state safely). Here are two different models that might work for this problem. Describe, in English, how these models might work. Can you think of any others? Are there practical consequences to the different models?

  1. All threads share access to one int.
  2. There are three bools, where each bool is read by one thread and written by a different thread.

The int might represent a phase that indicates which thread should print next. In the first phase (phase 0), thread t1 should print and then advance the phase to 1. Then, in the second phase, thread t2 should print and advance the phase to 2. And so on. Each thread is waiting for its phase and then advancing the phase.

The bools, on the other hand, might represent explicit permission to run. Each thread waits for its bool to become true, then prints, then sets the next bool to true. The main thread gives t1 permission to print by setting the first bool to true; then, after printing, t1 gives t2 permission to print; and so on.

The phase model (the single int) is more coarse-grained, since all threads are constantly accessing the same state. The permission model (three bools) is relatively more fine-grained, since each thread reads different state. Coarse-grained synchronization is typically more efficient in terms of space. It may be slower than fine-grained synchronization, though, because many threads have conflicting access to the same state. In this simple example, the performance consequences aren’t important, but fine-grained synchronization can matter a lot in more practical, large-scale code.

QUESTION. The following code using std::mutex is suspicious—very likely to be incorrect. Describe why, and suggest how to fix this problem.

std::mutex m;
int phase;
...
m.lock();
while (phase != 1) {
}
m.unlock();

EXERCISE. Change counting.cc to use synchronize using mutexes. Do not use condition variables or atomic variables.

DISCUSSION. Our solutions place the shared state and synchronization objects in global variables. Why is this? Is there a way to make these variables local variables in one thread or another?

EXERCISE. Consider our solutions for this exercise. How can synchronization operations be moved around the code without breaking functionality?

QUESTION. Does your mutex-based solution use blocking or polling for the counter threads?

Counting with condition variables

EXERCISE. Change counting.cc to use synchronize using mutexes and condition variables. Do not use atomics.

QUESTION. Does your condition-variable-based solution use blocking or polling?

QUESTION. Which of these solutions use fine-grained synchronization (as opposed to coarse-grained synchronization)?

Counting with atomics (offline)

OFFLINE EXERCISE. Use atomic variables and no other synchronization objects to ensure that counting.cc always prints 1\n2\n3\n.

OFFLINE QUESTION. Does your atomics-based solution use blocking or polling?

(Recall that blocking is when a thread waits in a non-runnable state. This can happen, for example, if the thread calls a system call like usleep, or waitpid without WNOHANG; the calling thread will not run until the system call completes, leaving the operating system free to do other work. Polling is when a thread waits in a runnable state—it continually loops back and checks to see if it can continue. This can happen, for example, if the thread calls waitpid with WNOHANG, in a loop.)

Bathroom synchronization

Ah, the bathroom. Does anything deserve more careful consideration? Is anything more suitable for poop jokes?

Bathroom 1: Stall privacy

In this exercise, the setting is a bathroom with K stalls numbered 0 through K–1. A number of users enter the bathroom; each has a preferred stall. Each user can be modeled as an independent thread that repeatedly calls the following code:

void user(int preferred_stall) {
    while (true) {
        wait_until_uncomfortable();
        poop_into(preferred_stall);
    }
}

Each user has a preferred stall that never changes, and two users should never poop_into the same stall at the same time.

You may intuitively smell that this bathroom model has a synchronization problem. Answer the following questions without referring to the solutions first.

EXERCISE. What synchronization property is missing?

EXERCISE. How could you show that there was a synchronization problem in philosophy01.cc?

EXERCISE. Provide synchronization using a single mutex, i.e., using coarse-grained locking. Your solution should be correct (it should avoid gross conditions [which are race conditions in the context of a bathroom-themed synchronization problem]), but it need not be efficient. Your solution will consist of (1) a definition for the mutex, which must be a global variable, and (2) new code for user(). Don’t worry too much about syntax.

EXERCISE. Provide synchronization using fine-grained locking. Your solution should be more efficient than the previous solution. Specifically, it should allow the bathroom to achieve full utilization: it should be possible for all K stalls to be in use at the same time. You will need to change both the global mutex—there will be more than one global mutex—and the code for user().

Bathroom 2: Stall indifference

The setting is the same, except that now users have no predetermined preferred stall. In philosophy02.cc, they pick a random stall each time.

void user() {
    ...
    while (true) {
        wait_until_uncomfortable();
        int preferred_stall = pick_stall(randomness);
        poop_into(preferred_stall);
    }
}

EXERCISE. Adapt your solution from Bathroom 1 to work for this variant.

EXERCISE. As written, the user code picks a preferred stall and then waits for that stall to become available. This isn’t very efficient, since multiple users might pick the same stall by chance. Write a synchronization solution where each user chooses any available stall. First, use a single mutex coupled with a global data structure representing stall occupancy.

EXERCISE. Repeat the exercise, but this time use multiple mutexes and no other global data structure.

EXERCISE. Compare these two implementations in terms of the number of std::mutex lock attempts it might take to poop.

EXERCISE. These solutions use polling. Why?

EXERCISE. Implement a blocking implementation. You will need a condition_variable.

Bathroom 3: Pooping philosophers (offline)

To facilitate the communal philosophical life, Plato decreed that philosophy should be performed in a bathroom with K thrones and K curtains, arranged like so:

Philosophers’ bathroom

Each philosopher sits on a throne with two curtains, one on each side. The philosophers normally argue with one another. But sometimes a philosopher’s got to poop, and to poop one needs privacy. A pooping philosopher must first draw the curtains on either side; philosopher X will need curtains X and (X+1) mod K. For instance, here, philosophers 0 and 2 can poop simultaneously, but philosopher 4 cannot poop until philosopher 0 is done (because philosopher 4 needs curtain 0), and philosopher 1 cannot poop until philosophers 0 and 2 are both done:

Philosophers’ bathroom with curtains drawn

EXERCISE. In high-level terms, what is the mutual exclusion problem here—that is, what resources are subject to mutual exclusion?

A philosopher could be described like this:

void philosopher(int x) {
    while (true) {
        argue_until_uncomfortable();
        draw curtains x and (x + 1) % K;
        poop();
        release curtains x and (x + 1) % K;
    }
}

EXERCISE. Write a correctly-synchronized version of this philosopher. Your solution should not suffer from deadlock, and philosophers should block (not poll) while waiting to poop. First, use std::scoped_lock—which can (magically) lock more than one mutex at a time while avoiding deadlock.

EXERCISE. Can you do it without std::scoped_lock?

Fancy mutexes

std::recursive_mutex is like a normal mutex, but a thread that holds a lock may recursively acquire it again (as long as they recursively release it the same number of times). This might be useful if you need to acquire several locks, and sometimes some of those locks might actually be the same lock: with recursive locks, you don't need to keep track of which ones you already hold, to avoid acquiring the same lock twice.

std::shared_mutex is an instantiation of a different kind of lock altogether: a readers-writer lock, or a shared-exclusive lock. The idea behind this kind of lock is that rather than just have one level of access—mutual exclusion—there are two levels of access: shared access and exclusive access. When a thread holds the lock with shared access, that guarantees that all other threads that hold the lock have shared access. When a thread holds the lock with exclusive access, that guarantees that it is the only thread to have any kind of access. We call this a readers-writer lock because the two levels of access often correspond to multiple reader threads, which require the shared data not to change while they're accessing it, but don't modify it themselves; and a single writer thread, which needs to be the only thread with access to the data it's modifying. You do not need the functionality of std::shared_mutex for the problem set, and in fact it is not particularly helpful for this particular application.

std::mutex will likely suffice for this problem set, though std::recursive_mutex may simplify some logic.

std::lock avoids deadlock

C++ also defines some handy tools to acquire locks while avoiding deadlock. Instead of calling .lock() on the mutex itself, you can pass several mutexes as arguments to the function std::lock, which magically acquires the locks in an order that avoids deadlock. For example:

#include <mutex>

std::mutex a;
std::mutex b;

void worker1() {
    std::lock(a, b);
    ...
    a.unlock(); // the order of these doesn't matter,
    b.unlock(); // since unlocking can't cause deadlock
}

void worker2() {
    std::lock(b, a); // look ma, no deadlock!
    ...
    b.unlock();
    a.unlock();
}

...

std::unique_lock and std::scoped_lock unlock for you

The std::scoped_lock and std::unique_lock objects automate some aspects of locking, because they are capable of automatically unlocking. This is really convenient—it’s very easy to forget a call to unlock if a function has multiple return statements in it.

To use a scoped lock, you declare an object at the point where you would normally call lock(). This object is typically called a guard, since it exists to “guard” the lock. You must pass in a reference to the std::mutex that you intend to lock. For example, here’s a version of the increase function from above:

#include <mutex>

int i;
std::mutex m; // protects write access to i

void increase() {
    std::unique_lock guard(m);
    i++;
}

The declaration of guard will block until m can be locked, and then m is automatically unlocked when the function returns and guard goes out of scope (the variable stops existing). This is an instance of an idiom called Resource acquisition is initialization (RAII), which is the idea that an object's resources are automatically acquired and released by its constructor and destructor, which are triggered by scope.

Here’s how std::scoped_lock (which is the multi-lock version of std::unique_lock) would simplify the code above:

#include <mutex>

std::mutex a;
std::mutex b;

void worker1() {
    std::scoped_lock guard(a, b);
    ...
}

void worker2() {
    std::scoped_lock guard(b, a);
    ...
}

...

The std::unique_lock has .lock(), .try_lock(), and .unlock() member functions, in addition to the initialization/deinitialization style, and these can be useful sometimes. The std::scoped_lock has no member functions at all: you can only use it by declaring it and then letting it go out of scope later.