2019-07-29 10:00:14 +00:00
|
|
|
#include <Kernel/Lock.h>
|
2019-12-01 10:57:20 +00:00
|
|
|
#include <Kernel/Thread.h>
|
2019-07-29 10:00:14 +00:00
|
|
|
|
|
|
|
void Lock::lock()
|
|
|
|
{
|
2019-08-22 07:22:30 +00:00
|
|
|
ASSERT(!Scheduler::is_active());
|
2019-07-29 10:00:14 +00:00
|
|
|
if (!are_interrupts_enabled()) {
|
|
|
|
kprintf("Interrupts disabled when trying to take Lock{%s}\n", m_name);
|
|
|
|
dump_backtrace();
|
|
|
|
hang();
|
|
|
|
}
|
|
|
|
for (;;) {
|
2019-10-12 17:17:34 +00:00
|
|
|
bool expected = false;
|
|
|
|
if (m_lock.compare_exchange_strong(expected, true, AK::memory_order_acq_rel)) {
|
2019-07-29 10:00:14 +00:00
|
|
|
if (!m_holder || m_holder == current) {
|
|
|
|
m_holder = current;
|
|
|
|
++m_level;
|
2019-10-12 17:17:34 +00:00
|
|
|
m_lock.store(false, AK::memory_order_release);
|
2019-07-29 10:00:14 +00:00
|
|
|
return;
|
|
|
|
}
|
2019-10-12 17:17:34 +00:00
|
|
|
m_lock.store(false, AK::memory_order_release);
|
2019-12-01 14:54:47 +00:00
|
|
|
current->wait_on(m_queue, m_holder, m_name);
|
2019-07-29 10:00:14 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
void Lock::unlock()
|
|
|
|
{
|
|
|
|
for (;;) {
|
2019-10-12 17:17:34 +00:00
|
|
|
bool expected = false;
|
|
|
|
if (m_lock.compare_exchange_strong(expected, true, AK::memory_order_acq_rel)) {
|
2019-07-29 10:00:14 +00:00
|
|
|
ASSERT(m_holder == current);
|
|
|
|
ASSERT(m_level);
|
|
|
|
--m_level;
|
|
|
|
if (m_level) {
|
2019-10-12 17:17:34 +00:00
|
|
|
m_lock.store(false, AK::memory_order_release);
|
2019-07-29 10:00:14 +00:00
|
|
|
return;
|
|
|
|
}
|
|
|
|
m_holder = nullptr;
|
2019-12-01 10:57:20 +00:00
|
|
|
m_queue.wake_one();
|
2019-10-12 17:17:34 +00:00
|
|
|
m_lock.store(false, AK::memory_order_release);
|
2019-07-29 10:00:14 +00:00
|
|
|
return;
|
|
|
|
}
|
2019-12-01 10:57:20 +00:00
|
|
|
// I don't know *who* is using "m_lock", so just yield.
|
|
|
|
Scheduler::yield();
|
2019-07-29 10:00:14 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
bool Lock::unlock_if_locked()
|
|
|
|
{
|
|
|
|
for (;;) {
|
2019-10-12 17:17:34 +00:00
|
|
|
bool expected = false;
|
|
|
|
if (m_lock.compare_exchange_strong(expected, true, AK::memory_order_acq_rel)) {
|
2019-07-29 10:00:14 +00:00
|
|
|
if (m_level == 0) {
|
2019-10-12 17:17:34 +00:00
|
|
|
m_lock.store(false, AK::memory_order_release);
|
2019-07-29 10:00:14 +00:00
|
|
|
return false;
|
|
|
|
}
|
2019-09-14 17:35:07 +00:00
|
|
|
if (m_holder != current) {
|
2019-10-12 17:17:34 +00:00
|
|
|
m_lock.store(false, AK::memory_order_release);
|
2019-09-14 17:35:07 +00:00
|
|
|
return false;
|
|
|
|
}
|
2019-07-29 10:00:14 +00:00
|
|
|
ASSERT(m_level);
|
|
|
|
--m_level;
|
|
|
|
if (m_level) {
|
2019-10-12 17:17:34 +00:00
|
|
|
m_lock.store(false, AK::memory_order_release);
|
2019-07-29 10:00:14 +00:00
|
|
|
return false;
|
|
|
|
}
|
|
|
|
m_holder = nullptr;
|
2019-10-12 17:17:34 +00:00
|
|
|
m_lock.store(false, AK::memory_order_release);
|
2019-12-01 10:57:20 +00:00
|
|
|
m_queue.wake_one();
|
2019-07-29 10:00:14 +00:00
|
|
|
return true;
|
|
|
|
}
|
2019-12-01 10:57:20 +00:00
|
|
|
// I don't know *who* is using "m_lock", so just yield.
|
|
|
|
Scheduler::yield();
|
2019-07-29 10:00:14 +00:00
|
|
|
}
|
|
|
|
}
|