mirror of
https://github.com/LadybirdBrowser/ladybird.git
synced 2024-11-22 23:50:19 +00:00
cee9528168
Let's be explicit about what kind of lock this is meant to be.
230 lines
5.7 KiB
C++
230 lines
5.7 KiB
C++
/*
|
|
* Copyright (c) 2018-2020, Andreas Kling <kling@serenityos.org>
|
|
*
|
|
* SPDX-License-Identifier: BSD-2-Clause
|
|
*/
|
|
|
|
#pragma once
|
|
|
|
#include <AK/Assertions.h>
|
|
#include <AK/Atomic.h>
|
|
#include <AK/HashMap.h>
|
|
#include <AK/Types.h>
|
|
#include <Kernel/Forward.h>
|
|
#include <Kernel/LockMode.h>
|
|
#include <Kernel/WaitQueue.h>
|
|
|
|
namespace Kernel {
|
|
|
|
class Mutex {
|
|
friend class Thread;
|
|
|
|
AK_MAKE_NONCOPYABLE(Mutex);
|
|
AK_MAKE_NONMOVABLE(Mutex);
|
|
|
|
public:
|
|
using Mode = LockMode;
|
|
|
|
Mutex(const char* name = nullptr)
|
|
: m_name(name)
|
|
{
|
|
}
|
|
~Mutex() = default;
|
|
|
|
#if LOCK_DEBUG
|
|
void lock(Mode mode = Mode::Exclusive, const SourceLocation& location = SourceLocation::current());
|
|
void restore_lock(Mode, u32, const SourceLocation& location = SourceLocation::current());
|
|
#else
|
|
void lock(Mode = Mode::Exclusive);
|
|
void restore_lock(Mode, u32);
|
|
#endif
|
|
|
|
void unlock();
|
|
[[nodiscard]] Mode force_unlock_if_locked(u32&);
|
|
[[nodiscard]] bool is_locked() const
|
|
{
|
|
ScopedSpinLock lock(m_lock);
|
|
return m_mode != Mode::Unlocked;
|
|
}
|
|
[[nodiscard]] bool own_lock() const
|
|
{
|
|
ScopedSpinLock lock(m_lock);
|
|
if (m_mode == Mode::Exclusive)
|
|
return m_holder == Thread::current();
|
|
if (m_mode == Mode::Shared)
|
|
return m_shared_holders.contains(Thread::current());
|
|
return false;
|
|
}
|
|
|
|
[[nodiscard]] const char* name() const { return m_name; }
|
|
|
|
static const char* mode_to_string(Mode mode)
|
|
{
|
|
switch (mode) {
|
|
case Mode::Unlocked:
|
|
return "unlocked";
|
|
case Mode::Exclusive:
|
|
return "exclusive";
|
|
case Mode::Shared:
|
|
return "shared";
|
|
default:
|
|
return "invalid";
|
|
}
|
|
}
|
|
|
|
private:
|
|
typedef IntrusiveList<Thread, RawPtr<Thread>, &Thread::m_blocked_threads_list_node> BlockedThreadList;
|
|
|
|
ALWAYS_INLINE BlockedThreadList& thread_list_for_mode(Mode mode)
|
|
{
|
|
VERIFY(mode == Mode::Exclusive || mode == Mode::Shared);
|
|
return mode == Mode::Exclusive ? m_blocked_threads_list_exclusive : m_blocked_threads_list_shared;
|
|
}
|
|
|
|
void block(Thread&, Mode, ScopedSpinLock<SpinLock<u8>>&, u32);
|
|
void unblock_waiters(Mode);
|
|
|
|
const char* m_name { nullptr };
|
|
Mode m_mode { Mode::Unlocked };
|
|
|
|
// When locked exclusively, only the thread already holding the lock can
|
|
// lock it again. When locked in shared mode, any thread can do that.
|
|
u32 m_times_locked { 0 };
|
|
|
|
// One of the threads that hold this lock, or nullptr. When locked in shared
|
|
// mode, this is stored on best effort basis: nullptr value does *not* mean
|
|
// the lock is unlocked, it just means we don't know which threads hold it.
|
|
// When locked exclusively, this is always the one thread that holds the
|
|
// lock.
|
|
RefPtr<Thread> m_holder;
|
|
HashMap<Thread*, u32> m_shared_holders;
|
|
|
|
BlockedThreadList m_blocked_threads_list_exclusive;
|
|
BlockedThreadList m_blocked_threads_list_shared;
|
|
|
|
mutable SpinLock<u8> m_lock;
|
|
};
|
|
|
|
class Locker {
|
|
public:
|
|
#if LOCK_DEBUG
|
|
ALWAYS_INLINE explicit Locker(Mutex& l, Mutex::Mode mode = Mutex::Mode::Exclusive, const SourceLocation& location = SourceLocation::current())
|
|
#else
|
|
ALWAYS_INLINE explicit Locker(Mutex& l, Mutex::Mode mode = Mutex::Mode::Exclusive)
|
|
#endif
|
|
: m_lock(l)
|
|
{
|
|
#if LOCK_DEBUG
|
|
m_lock.lock(mode, location);
|
|
#else
|
|
m_lock.lock(mode);
|
|
#endif
|
|
}
|
|
|
|
ALWAYS_INLINE ~Locker()
|
|
{
|
|
if (m_locked)
|
|
unlock();
|
|
}
|
|
ALWAYS_INLINE void unlock()
|
|
{
|
|
VERIFY(m_locked);
|
|
m_locked = false;
|
|
m_lock.unlock();
|
|
}
|
|
|
|
#if LOCK_DEBUG
|
|
ALWAYS_INLINE void lock(Mutex::Mode mode = Mutex::Mode::Exclusive, const SourceLocation& location = SourceLocation::current())
|
|
#else
|
|
ALWAYS_INLINE void lock(Mutex::Mode mode = Mutex::Mode::Exclusive)
|
|
#endif
|
|
{
|
|
VERIFY(!m_locked);
|
|
m_locked = true;
|
|
|
|
#if LOCK_DEBUG
|
|
m_lock.lock(mode, location);
|
|
#else
|
|
m_lock.lock(mode);
|
|
#endif
|
|
}
|
|
|
|
Mutex& get_lock() { return m_lock; }
|
|
const Mutex& get_lock() const { return m_lock; }
|
|
|
|
private:
|
|
Mutex& m_lock;
|
|
bool m_locked { true };
|
|
};
|
|
|
|
template<typename T>
|
|
class Lockable {
|
|
public:
|
|
Lockable() = default;
|
|
Lockable(T&& resource)
|
|
: m_resource(move(resource))
|
|
{
|
|
}
|
|
[[nodiscard]] Mutex& lock() { return m_lock; }
|
|
[[nodiscard]] T& resource() { return m_resource; }
|
|
|
|
[[nodiscard]] T lock_and_copy()
|
|
{
|
|
Locker locker(m_lock);
|
|
return m_resource;
|
|
}
|
|
|
|
private:
|
|
T m_resource;
|
|
Mutex m_lock;
|
|
};
|
|
|
|
class ScopedLockRelease {
|
|
AK_MAKE_NONCOPYABLE(ScopedLockRelease);
|
|
|
|
public:
|
|
ScopedLockRelease& operator=(ScopedLockRelease&&) = delete;
|
|
|
|
ScopedLockRelease(Mutex& lock)
|
|
: m_lock(&lock)
|
|
, m_previous_mode(lock.force_unlock_if_locked(m_previous_recursions))
|
|
{
|
|
}
|
|
|
|
ScopedLockRelease(ScopedLockRelease&& from)
|
|
: m_lock(exchange(from.m_lock, nullptr))
|
|
, m_previous_mode(exchange(from.m_previous_mode, Mutex::Mode::Unlocked))
|
|
, m_previous_recursions(exchange(from.m_previous_recursions, 0))
|
|
{
|
|
}
|
|
|
|
~ScopedLockRelease()
|
|
{
|
|
if (m_lock && m_previous_mode != Mutex::Mode::Unlocked)
|
|
m_lock->restore_lock(m_previous_mode, m_previous_recursions);
|
|
}
|
|
|
|
void restore_lock()
|
|
{
|
|
VERIFY(m_lock);
|
|
if (m_previous_mode != Mutex::Mode::Unlocked) {
|
|
m_lock->restore_lock(m_previous_mode, m_previous_recursions);
|
|
m_previous_mode = Mutex::Mode::Unlocked;
|
|
m_previous_recursions = 0;
|
|
}
|
|
}
|
|
|
|
void do_not_restore()
|
|
{
|
|
VERIFY(m_lock);
|
|
m_previous_mode = Mutex::Mode::Unlocked;
|
|
m_previous_recursions = 0;
|
|
}
|
|
|
|
private:
|
|
Mutex* m_lock;
|
|
Mutex::Mode m_previous_mode;
|
|
u32 m_previous_recursions;
|
|
};
|
|
|
|
}
|