mirror of
https://github.com/LadybirdBrowser/ladybird.git
synced 2024-11-23 08:00:20 +00:00
a6a439243f
This step would ideally not have been necessary (increases amount of refactoring and templates necessary, which in turn increases build times), but it gives us a couple of nice properties: - SpinlockProtected inside Singleton (a very common combination) can now obtain any lock rank just via the template parameter. It was not previously possible to do this with SingletonInstanceCreator magic. - SpinlockProtected's lock rank is now mandatory; this is the majority of cases and allows us to see where we're still missing proper ranks. - The type already informs us what lock rank a lock has, which aids code readability and (possibly, if gdb cooperates) lock mismatch debugging. - The rank of a lock can no longer be dynamic, which is not something we wanted in the first place (or made use of). Locks randomly changing their rank sounds like a disaster waiting to happen. - In some places, we might be able to statically check that locks are taken in the right order (with the right lock rank checking implementation) as rank information is fully statically known. This refactoring even more exposes the fact that Mutex has no lock rank capabilites, which is not fixed here.
232 lines
7.1 KiB
C++
232 lines
7.1 KiB
C++
/*
|
|
* Copyright (c) 2018-2020, Andreas Kling <kling@serenityos.org>
|
|
*
|
|
* SPDX-License-Identifier: BSD-2-Clause
|
|
*/
|
|
|
|
#include <AK/Singleton.h>
|
|
#include <AK/Time.h>
|
|
#include <Kernel/Scheduler.h>
|
|
#include <Kernel/Sections.h>
|
|
#include <Kernel/Time/TimeManagement.h>
|
|
#include <Kernel/TimerQueue.h>
|
|
|
|
namespace Kernel {
|
|
|
|
static Singleton<TimerQueue> s_the;
|
|
static Spinlock<LockRank::None> g_timerqueue_lock {};
|
|
|
|
Time Timer::remaining() const
|
|
{
|
|
return m_remaining;
|
|
}
|
|
|
|
Time Timer::now(bool is_firing) const
|
|
{
|
|
// NOTE: If is_firing is true then TimePrecision::Precise isn't really useful here.
|
|
// We already have a quite precise time stamp because we just updated the time in the
|
|
// interrupt handler. In those cases, just use coarse timestamps.
|
|
auto clock_id = m_clock_id;
|
|
if (is_firing) {
|
|
switch (clock_id) {
|
|
case CLOCK_MONOTONIC:
|
|
clock_id = CLOCK_MONOTONIC_COARSE;
|
|
break;
|
|
case CLOCK_MONOTONIC_RAW:
|
|
// TODO: use a special CLOCK_MONOTONIC_RAW_COARSE like mechanism here
|
|
break;
|
|
case CLOCK_REALTIME:
|
|
clock_id = CLOCK_REALTIME_COARSE;
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
}
|
|
return TimeManagement::the().current_time(clock_id);
|
|
}
|
|
|
|
TimerQueue& TimerQueue::the()
|
|
{
|
|
return *s_the;
|
|
}
|
|
|
|
UNMAP_AFTER_INIT TimerQueue::TimerQueue()
|
|
{
|
|
m_ticks_per_second = TimeManagement::the().ticks_per_second();
|
|
}
|
|
|
|
bool TimerQueue::add_timer_without_id(NonnullLockRefPtr<Timer> timer, clockid_t clock_id, Time const& deadline, Function<void()>&& callback)
|
|
{
|
|
if (deadline <= TimeManagement::the().current_time(clock_id))
|
|
return false;
|
|
|
|
// Because timer handlers can execute on any processor and there is
|
|
// a race between executing a timer handler and cancel_timer() this
|
|
// *must* be a LockRefPtr<Timer>. Otherwise, calling cancel_timer() could
|
|
// inadvertently cancel another timer that has been created between
|
|
// returning from the timer handler and a call to cancel_timer().
|
|
timer->setup(clock_id, deadline, move(callback));
|
|
|
|
SpinlockLocker lock(g_timerqueue_lock);
|
|
timer->m_id = 0; // Don't generate a timer id
|
|
add_timer_locked(move(timer));
|
|
return true;
|
|
}
|
|
|
|
TimerId TimerQueue::add_timer(NonnullLockRefPtr<Timer>&& timer)
|
|
{
|
|
SpinlockLocker lock(g_timerqueue_lock);
|
|
|
|
timer->m_id = ++m_timer_id_count;
|
|
VERIFY(timer->m_id != 0); // wrapped
|
|
auto id = timer->m_id;
|
|
add_timer_locked(move(timer));
|
|
return id;
|
|
}
|
|
|
|
void TimerQueue::add_timer_locked(NonnullLockRefPtr<Timer> timer)
|
|
{
|
|
Time timer_expiration = timer->m_expires;
|
|
|
|
timer->clear_cancelled();
|
|
timer->clear_callback_finished();
|
|
timer->set_in_use();
|
|
|
|
auto& queue = queue_for_timer(*timer);
|
|
if (queue.list.is_empty()) {
|
|
queue.list.append(timer.leak_ref());
|
|
queue.next_timer_due = timer_expiration;
|
|
} else {
|
|
Timer* following_timer = nullptr;
|
|
for (auto& t : queue.list) {
|
|
if (t.m_expires > timer_expiration) {
|
|
following_timer = &t;
|
|
break;
|
|
}
|
|
}
|
|
if (following_timer) {
|
|
bool next_timer_needs_update = queue.list.first() == following_timer;
|
|
queue.list.insert_before(*following_timer, timer.leak_ref());
|
|
if (next_timer_needs_update)
|
|
queue.next_timer_due = timer_expiration;
|
|
} else {
|
|
queue.list.append(timer.leak_ref());
|
|
}
|
|
}
|
|
}
|
|
|
|
bool TimerQueue::cancel_timer(Timer& timer, bool* was_in_use)
|
|
{
|
|
bool in_use = timer.is_in_use();
|
|
if (was_in_use)
|
|
*was_in_use = in_use;
|
|
|
|
// If the timer isn't in use, the cancellation is a no-op.
|
|
if (!in_use) {
|
|
VERIFY(!timer.m_list_node.is_in_list());
|
|
return false;
|
|
}
|
|
|
|
bool did_already_run = timer.set_cancelled();
|
|
auto& timer_queue = queue_for_timer(timer);
|
|
if (!did_already_run) {
|
|
timer.clear_in_use();
|
|
|
|
SpinlockLocker lock(g_timerqueue_lock);
|
|
if (timer_queue.list.contains(timer)) {
|
|
// The timer has not fired, remove it
|
|
VERIFY(timer.ref_count() > 1);
|
|
remove_timer_locked(timer_queue, timer);
|
|
return true;
|
|
}
|
|
|
|
// The timer was queued to execute but hasn't had a chance
|
|
// to run. In this case, it should still be in m_timers_executing
|
|
// and we don't need to spin. It still holds a reference
|
|
// that will be dropped when it does get a chance to run,
|
|
// but since we called set_cancelled it will only drop its reference
|
|
VERIFY(m_timers_executing.contains(timer));
|
|
m_timers_executing.remove(timer);
|
|
return true;
|
|
}
|
|
|
|
// At this point the deferred call is queued and is being executed
|
|
// on another processor. We need to wait until it's complete!
|
|
while (!timer.is_callback_finished())
|
|
Processor::wait_check();
|
|
|
|
return false;
|
|
}
|
|
|
|
void TimerQueue::remove_timer_locked(Queue& queue, Timer& timer)
|
|
{
|
|
bool was_next_timer = (queue.list.first() == &timer);
|
|
queue.list.remove(timer);
|
|
auto now = timer.now(false);
|
|
if (timer.m_expires > now)
|
|
timer.m_remaining = timer.m_expires - now;
|
|
|
|
if (was_next_timer)
|
|
update_next_timer_due(queue);
|
|
// Whenever we remove a timer that was still queued (but hasn't been
|
|
// fired) we added a reference to it. So, when removing it from the
|
|
// queue we need to drop that reference.
|
|
timer.unref();
|
|
}
|
|
|
|
void TimerQueue::fire()
|
|
{
|
|
SpinlockLocker lock(g_timerqueue_lock);
|
|
|
|
auto fire_timers = [&](Queue& queue) {
|
|
auto* timer = queue.list.first();
|
|
VERIFY(timer);
|
|
VERIFY(queue.next_timer_due == timer->m_expires);
|
|
|
|
while (timer && timer->now(true) > timer->m_expires) {
|
|
queue.list.remove(*timer);
|
|
|
|
m_timers_executing.append(*timer);
|
|
|
|
update_next_timer_due(queue);
|
|
|
|
lock.unlock();
|
|
|
|
// Defer executing the timer outside of the irq handler
|
|
Processor::deferred_call_queue([this, timer]() {
|
|
// Check if we were cancelled in between being triggered
|
|
// by the timer irq handler and now. If so, just drop
|
|
// our reference and don't execute the callback.
|
|
if (!timer->set_cancelled()) {
|
|
timer->m_callback();
|
|
SpinlockLocker lock(g_timerqueue_lock);
|
|
m_timers_executing.remove(*timer);
|
|
}
|
|
timer->clear_in_use();
|
|
timer->set_callback_finished();
|
|
// Drop the reference we added when queueing the timer
|
|
timer->unref();
|
|
});
|
|
|
|
lock.lock();
|
|
timer = queue.list.first();
|
|
}
|
|
};
|
|
|
|
if (!m_timer_queue_monotonic.list.is_empty())
|
|
fire_timers(m_timer_queue_monotonic);
|
|
if (!m_timer_queue_realtime.list.is_empty())
|
|
fire_timers(m_timer_queue_realtime);
|
|
}
|
|
|
|
void TimerQueue::update_next_timer_due(Queue& queue)
|
|
{
|
|
VERIFY(g_timerqueue_lock.is_locked());
|
|
|
|
if (auto* next_timer = queue.list.first())
|
|
queue.next_timer_due = next_timer->m_expires;
|
|
else
|
|
queue.next_timer_due = {};
|
|
}
|
|
|
|
}
|