mirror of
https://github.com/LadybirdBrowser/ladybird.git
synced 2024-11-22 15:40:19 +00:00
dcfec8bfbe
tl;dr: This fixes deadlocks that would occur in most applications in Serenity when SMP was enabled. As an optimization to `pthread_mutex_unlock()`, if only one thread holds a mutex lock, it will avoid calling `futex()` to wake, since no threads are actually waiting in `futex()`. If a thread manages to synchronously unlock and relock the same mutex, the state will be set to indicate that a wake is not needed, regardless of whether any other threads are waiting for the lock. This should be fine, as a thread that is waiting will set the mutex to needing a wake and check if it is unlocked to unblock itself in the same atomic operation. However, when `pthread_mutex_lock()` was called with only one thread holding the mutex lock, instead of telling `futex()` to atomically sleep the thread if the state is set to indicate that a wake is needed, the first wait would check if the state was set to indicate a wake was _not_ needed. This means it is possible for the call to `futex()` to wait when any subsequent call to `pthread_mutex_unlock()` would not call `futex()` to wake, causing it to wait forever. After that, any other thread that tries to take the lock will see that the lock is taken and also `futex()` wait. Despite the fact that these other threads would set the state to needing a wake, there will be no unblocked thread holding the lock to actually wake them. By making it wait only if the state indicates to other threads that a wake is needed, heavily contended mutexes should no longer cause deadlocks. Most applications would encounter these deadlocks due to the mutex used by `malloc()`, some sooner than others. The worst offenders (other than Ladybird) were most likely VideoPlayer and SoundPlayer.
204 lines
6.5 KiB
C++
204 lines
6.5 KiB
C++
/*
|
|
* Copyright (c) 2021, the SerenityOS developers.
|
|
*
|
|
* SPDX-License-Identifier: BSD-2-Clause
|
|
*/
|
|
|
|
#include <AK/Atomic.h>
|
|
#include <AK/NeverDestroyed.h>
|
|
#include <AK/Types.h>
|
|
#include <AK/Vector.h>
|
|
#include <bits/pthread_integration.h>
|
|
#include <errno.h>
|
|
#include <pthread.h>
|
|
#include <sched.h>
|
|
#include <serenity.h>
|
|
#include <unistd.h>
|
|
|
|
namespace {
|
|
|
|
// Most programs don't need this, no need to incur an extra mutex lock/unlock on them
|
|
static Atomic<bool> g_did_touch_atfork { false };
|
|
static pthread_mutex_t g_atfork_list_mutex __PTHREAD_MUTEX_INITIALIZER;
|
|
static NeverDestroyed<Vector<void (*)(void), 4>> g_atfork_prepare_list;
|
|
static NeverDestroyed<Vector<void (*)(void), 4>> g_atfork_child_list;
|
|
static NeverDestroyed<Vector<void (*)(void), 4>> g_atfork_parent_list;
|
|
|
|
}
|
|
|
|
extern "C" {
|
|
void __pthread_fork_prepare(void)
|
|
{
|
|
if (!g_did_touch_atfork.load())
|
|
return;
|
|
|
|
pthread_mutex_lock(&g_atfork_list_mutex);
|
|
for (auto entry : g_atfork_prepare_list.get())
|
|
entry();
|
|
pthread_mutex_unlock(&g_atfork_list_mutex);
|
|
}
|
|
|
|
void __pthread_fork_child(void)
|
|
{
|
|
if (!g_did_touch_atfork.load())
|
|
return;
|
|
|
|
pthread_mutex_lock(&g_atfork_list_mutex);
|
|
for (auto entry : g_atfork_child_list.get())
|
|
entry();
|
|
pthread_mutex_unlock(&g_atfork_list_mutex);
|
|
}
|
|
|
|
void __pthread_fork_parent(void)
|
|
{
|
|
if (!g_did_touch_atfork.load())
|
|
return;
|
|
|
|
pthread_mutex_lock(&g_atfork_list_mutex);
|
|
for (auto entry : g_atfork_parent_list.get())
|
|
entry();
|
|
pthread_mutex_unlock(&g_atfork_list_mutex);
|
|
}
|
|
|
|
void __pthread_fork_atfork_register_prepare(void (*func)(void))
|
|
{
|
|
g_did_touch_atfork.store(true);
|
|
|
|
pthread_mutex_lock(&g_atfork_list_mutex);
|
|
g_atfork_prepare_list->append(func);
|
|
pthread_mutex_unlock(&g_atfork_list_mutex);
|
|
}
|
|
|
|
void __pthread_fork_atfork_register_parent(void (*func)(void))
|
|
{
|
|
g_did_touch_atfork.store(true);
|
|
|
|
pthread_mutex_lock(&g_atfork_list_mutex);
|
|
g_atfork_parent_list->append(func);
|
|
pthread_mutex_unlock(&g_atfork_list_mutex);
|
|
}
|
|
|
|
void __pthread_fork_atfork_register_child(void (*func)(void))
|
|
{
|
|
g_did_touch_atfork.store(true);
|
|
|
|
pthread_mutex_lock(&g_atfork_list_mutex);
|
|
g_atfork_child_list->append(func);
|
|
pthread_mutex_unlock(&g_atfork_list_mutex);
|
|
}
|
|
|
|
// https://pubs.opengroup.org/onlinepubs/009695399/functions/pthread_self.html
|
|
int pthread_self()
|
|
{
|
|
return gettid();
|
|
}
|
|
|
|
static constexpr u32 MUTEX_UNLOCKED = 0;
|
|
static constexpr u32 MUTEX_LOCKED_NO_NEED_TO_WAKE = 1;
|
|
static constexpr u32 MUTEX_LOCKED_NEED_TO_WAKE = 2;
|
|
|
|
// https://pubs.opengroup.org/onlinepubs/009695399/functions/pthread_mutex_init.html
|
|
int pthread_mutex_init(pthread_mutex_t* mutex, pthread_mutexattr_t const* attributes)
|
|
{
|
|
mutex->lock = 0;
|
|
mutex->owner = 0;
|
|
mutex->level = 0;
|
|
mutex->type = attributes ? attributes->type : __PTHREAD_MUTEX_NORMAL;
|
|
return 0;
|
|
}
|
|
|
|
// https://pubs.opengroup.org/onlinepubs/009695399/functions/pthread_mutex_trylock.html
|
|
int pthread_mutex_trylock(pthread_mutex_t* mutex)
|
|
{
|
|
u32 expected = MUTEX_UNLOCKED;
|
|
bool exchanged = AK::atomic_compare_exchange_strong(&mutex->lock, expected, MUTEX_LOCKED_NO_NEED_TO_WAKE, AK::memory_order_acquire);
|
|
|
|
if (exchanged) [[likely]] {
|
|
if (mutex->type == __PTHREAD_MUTEX_RECURSIVE)
|
|
AK::atomic_store(&mutex->owner, pthread_self(), AK::memory_order_relaxed);
|
|
mutex->level = 0;
|
|
return 0;
|
|
} else if (mutex->type == __PTHREAD_MUTEX_RECURSIVE) {
|
|
pthread_t owner = AK::atomic_load(&mutex->owner, AK::memory_order_relaxed);
|
|
if (owner == pthread_self()) {
|
|
// We already own the mutex!
|
|
mutex->level++;
|
|
return 0;
|
|
}
|
|
}
|
|
return EBUSY;
|
|
}
|
|
|
|
// https://pubs.opengroup.org/onlinepubs/009695399/functions/pthread_mutex_lock.html
|
|
int pthread_mutex_lock(pthread_mutex_t* mutex)
|
|
{
|
|
// Fast path: attempt to claim the mutex without waiting.
|
|
u32 value = MUTEX_UNLOCKED;
|
|
bool exchanged = AK::atomic_compare_exchange_strong(&mutex->lock, value, MUTEX_LOCKED_NO_NEED_TO_WAKE, AK::memory_order_acquire);
|
|
if (exchanged) [[likely]] {
|
|
if (mutex->type == __PTHREAD_MUTEX_RECURSIVE)
|
|
AK::atomic_store(&mutex->owner, pthread_self(), AK::memory_order_relaxed);
|
|
mutex->level = 0;
|
|
return 0;
|
|
} else if (mutex->type == __PTHREAD_MUTEX_RECURSIVE) {
|
|
pthread_t owner = AK::atomic_load(&mutex->owner, AK::memory_order_relaxed);
|
|
if (owner == pthread_self()) {
|
|
// We already own the mutex!
|
|
mutex->level++;
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
// Slow path: wait, record the fact that we're going to wait, and always
|
|
// remember to wake the next thread up once we release the mutex.
|
|
if (value != MUTEX_LOCKED_NEED_TO_WAKE)
|
|
value = AK::atomic_exchange(&mutex->lock, MUTEX_LOCKED_NEED_TO_WAKE, AK::memory_order_acquire);
|
|
|
|
while (value != MUTEX_UNLOCKED) {
|
|
futex_wait(&mutex->lock, MUTEX_LOCKED_NEED_TO_WAKE, nullptr, 0, false);
|
|
value = AK::atomic_exchange(&mutex->lock, MUTEX_LOCKED_NEED_TO_WAKE, AK::memory_order_acquire);
|
|
}
|
|
|
|
if (mutex->type == __PTHREAD_MUTEX_RECURSIVE)
|
|
AK::atomic_store(&mutex->owner, pthread_self(), AK::memory_order_relaxed);
|
|
mutex->level = 0;
|
|
return 0;
|
|
}
|
|
|
|
int __pthread_mutex_lock_pessimistic_np(pthread_mutex_t* mutex)
|
|
{
|
|
// Same as pthread_mutex_lock(), but always set MUTEX_LOCKED_NEED_TO_WAKE,
|
|
// and also don't bother checking for already owning the mutex recursively,
|
|
// because we know we don't. Used in the condition variable implementation.
|
|
u32 value = AK::atomic_exchange(&mutex->lock, MUTEX_LOCKED_NEED_TO_WAKE, AK::memory_order_acquire);
|
|
while (value != MUTEX_UNLOCKED) {
|
|
futex_wait(&mutex->lock, value, nullptr, 0, false);
|
|
value = AK::atomic_exchange(&mutex->lock, MUTEX_LOCKED_NEED_TO_WAKE, AK::memory_order_acquire);
|
|
}
|
|
|
|
if (mutex->type == __PTHREAD_MUTEX_RECURSIVE)
|
|
AK::atomic_store(&mutex->owner, pthread_self(), AK::memory_order_relaxed);
|
|
mutex->level = 0;
|
|
return 0;
|
|
}
|
|
|
|
// https://pubs.opengroup.org/onlinepubs/009695399/functions/pthread_mutex_unlock.html
|
|
int pthread_mutex_unlock(pthread_mutex_t* mutex)
|
|
{
|
|
if (mutex->type == __PTHREAD_MUTEX_RECURSIVE && mutex->level > 0) {
|
|
mutex->level--;
|
|
return 0;
|
|
}
|
|
|
|
if (mutex->type == __PTHREAD_MUTEX_RECURSIVE)
|
|
AK::atomic_store(&mutex->owner, 0, AK::memory_order_relaxed);
|
|
|
|
u32 value = AK::atomic_exchange(&mutex->lock, MUTEX_UNLOCKED, AK::memory_order_release);
|
|
if (value == MUTEX_LOCKED_NEED_TO_WAKE) [[unlikely]] {
|
|
int rc = futex_wake(&mutex->lock, 1, false);
|
|
VERIFY(rc >= 0);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
}
|