1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162636465666768697071727374757677787980818283848586878889909192939495969798991001011021031041051061071081091101111121131141151161171181191201211221231241251261271281291301311321331341351361371381391401411421431441451461471481491501511521531541551561571581591601611621631641651661671681691701711721731741751761771781791801811821831841851861871881891901911921931941951961971981992002012022032042052062072082092102112122132142152162172182192202212222232242252262272282292302312322332342352362372382392402412422432442452462472482492502512522532542552562572582592602612622632642652662672682692702712722732742752762772782792802812822832842852862872882892902912922932942952962972982993003013023033043053063073083093103113123133143153163173183193203213223233243253263273283293303313323333343353363373383393403413423433443453463473483493503513523533543553563573583593603613623633643653663673683693703713723733743753763773783793803813823833843853863873883893903913923933943953963973983994004014024034044054064074084094104114124134144154164174184194204214224234244254264274284294304314324334344354364374384394404414424434444454464474484494504514524534544554564574584594604614624634644654664674684694704714724734744754764774784794804814824834844854864874884894904914924934944954964974984995005015025035045055065075085095105115125135145155165175185195205215225235245255265275285295305315325335345355365375385395405415425435445455465475485495505515525535545555565575585595605615625635645655665675685695705715725735745755765775785795805815825835845855865875885895905915925935945955965975985996006016026036046056066076086096106116126136146156166176186196206216226236246256266276286296306316326336346356366376386396406416426436446456466476486496506516526536546556566576586596606616626636646656666676686696706716726736746756766776786796806816826836846856866876886896906916926936946956966976986997007017027037047057067077087097107117127137147157167177187197207217227237247257267277287297307317327337347357367377387397407417427437447457467477487497507517527537547557567577587597607617627637647657667677687697707717727737747757767777787797807817827837847857867877887897907917927937947957967977987998008018028038048058068078088098108118128138148158168178188198208218228238248258268278288298308318328338348358368378388398408418428438448458468478488498508518528538548558568578588598608618628638648658668678688698708718728738748758768778788798808818828838848858868878888898908918928938948958968978988999009019029039049059069079089099109119129139149159169179189199209219229239249259269279289299309319329339349359369379389399409419429439449459469479489499509519529539549559569579589599609619629639649659669679689699709719729739749759769779789799809819829839849859869879889899909919929939949959969979989991000100110021003100410051006100710081009101010111012101310141015101610171018101910201021102210231024102510261027102810291030103110321033103410351036103710381039104010411042104310441045104610471048104910501051105210531054105510561057105810591060106110621063106410651066106710681069107010711072107310741075107610771078107910801081108210831084108510861087108810891090109110921093109410951096109710981099110011011102110311041105110611071108110911101111111211131114111511161117111811191120112111221123112411251126112711281129113011311132113311341135113611371138113911401141114211431144114511461147114811491150115111521153115411551156115711581159116011611162116311641165116611671168116911701171117211731174117511761177117811791180118111821183118411851186118711881189119011911192119311941195119611971198119912001201120212031204120512061207120812091210121112121213121412151216121712181219122012211222122312241225122612271228122912301231123212331234123512361237123812391240124112421243124412451246124712481249125012511252125312541255125612571258125912601261126212631264126512661267126812691270127112721273127412751276127712781279128012811282128312841285128612871288128912901291129212931294129512961297129812991300130113021303130413051306130713081309131013111312131313141315131613171318131913201321132213231324132513261327132813291330133113321333133413351336133713381339134013411342134313441345134613471348134913501351135213531354135513561357135813591360136113621363136413651366136713681369137013711372137313741375137613771378137913801381138213831384138513861387138813891390139113921393139413951396139713981399140014011402140314041405140614071408140914101411141214131414141514161417141814191420142114221423142414251426142714281429143014311432 |
- /*
- * Copyright (c) 2018-2021, Andreas Kling <kling@serenityos.org>
- *
- * SPDX-License-Identifier: BSD-2-Clause
- */
- #pragma once
- #include <AK/Concepts.h>
- #include <AK/EnumBits.h>
- #include <AK/HashMap.h>
- #include <AK/IntrusiveList.h>
- #include <AK/Optional.h>
- #include <AK/OwnPtr.h>
- #ifdef LOCK_DEBUG
- # include <AK/SourceLocation.h>
- #endif
- #include <AK/String.h>
- #include <AK/Time.h>
- #include <AK/Vector.h>
- #include <AK/WeakPtr.h>
- #include <AK/Weakable.h>
- #include <Kernel/Arch/x86/RegisterState.h>
- #include <Kernel/Arch/x86/SafeMem.h>
- #include <Kernel/Debug.h>
- #include <Kernel/FileSystem/InodeIdentifier.h>
- #include <Kernel/Forward.h>
- #include <Kernel/KResult.h>
- #include <Kernel/KString.h>
- #include <Kernel/Locking/LockMode.h>
- #include <Kernel/Memory/VirtualRange.h>
- #include <Kernel/Scheduler.h>
- #include <Kernel/TimerQueue.h>
- #include <Kernel/UnixTypes.h>
- #include <LibC/fd_set.h>
- #include <LibC/signal_numbers.h>
- namespace Kernel {
- namespace Memory {
- extern RecursiveSpinLock s_mm_lock;
- }
- enum class DispatchSignalResult {
- Deferred = 0,
- Yield,
- Terminate,
- Continue
- };
- struct SignalActionData {
- VirtualAddress handler_or_sigaction;
- u32 mask { 0 };
- int flags { 0 };
- };
- struct ThreadSpecificData {
- ThreadSpecificData* self;
- };
- #define THREAD_PRIORITY_MIN 1
- #define THREAD_PRIORITY_LOW 10
- #define THREAD_PRIORITY_NORMAL 30
- #define THREAD_PRIORITY_HIGH 50
- #define THREAD_PRIORITY_MAX 99
- #define THREAD_AFFINITY_DEFAULT 0xffffffff
- struct ThreadRegisters {
- #if ARCH(I386)
- FlatPtr ss;
- FlatPtr gs;
- FlatPtr fs;
- FlatPtr es;
- FlatPtr ds;
- FlatPtr edi;
- FlatPtr esi;
- FlatPtr ebp;
- FlatPtr esp;
- FlatPtr ebx;
- FlatPtr edx;
- FlatPtr ecx;
- FlatPtr eax;
- FlatPtr eip;
- FlatPtr esp0;
- FlatPtr ss0;
- #else
- FlatPtr rdi;
- FlatPtr rsi;
- FlatPtr rbp;
- FlatPtr rsp;
- FlatPtr rbx;
- FlatPtr rdx;
- FlatPtr rcx;
- FlatPtr rax;
- FlatPtr r8;
- FlatPtr r9;
- FlatPtr r10;
- FlatPtr r11;
- FlatPtr r12;
- FlatPtr r13;
- FlatPtr r14;
- FlatPtr r15;
- FlatPtr rip;
- FlatPtr rsp0;
- #endif
- FlatPtr cs;
- #if ARCH(I386)
- FlatPtr eflags;
- #else
- FlatPtr rflags;
- #endif
- FlatPtr cr3;
- FlatPtr ip() const
- {
- #if ARCH(I386)
- return eip;
- #else
- return rip;
- #endif
- }
- FlatPtr sp() const
- {
- #if ARCH(I386)
- return esp;
- #else
- return rsp;
- #endif
- }
- };
- class Thread
- : public RefCounted<Thread>
- , public Weakable<Thread> {
- AK_MAKE_NONCOPYABLE(Thread);
- AK_MAKE_NONMOVABLE(Thread);
- friend class Mutex;
- friend class Process;
- friend class ProtectedProcessBase;
- friend class Scheduler;
- friend struct ThreadReadyQueue;
- static SpinLock<u8> g_tid_map_lock;
- static HashMap<ThreadID, Thread*>* g_tid_map;
- public:
- inline static Thread* current()
- {
- return Processor::current_thread();
- }
- static void initialize();
- static KResultOr<NonnullRefPtr<Thread>> try_create(NonnullRefPtr<Process>);
- ~Thread();
- static RefPtr<Thread> from_tid(ThreadID);
- static void finalize_dying_threads();
- ThreadID tid() const { return m_tid; }
- ProcessID pid() const;
- void set_priority(u32 p) { m_priority = p; }
- u32 priority() const { return m_priority; }
- void detach()
- {
- ScopedSpinLock lock(m_lock);
- m_is_joinable = false;
- }
- [[nodiscard]] bool is_joinable() const
- {
- ScopedSpinLock lock(m_lock);
- return m_is_joinable;
- }
- Process& process() { return m_process; }
- const Process& process() const { return m_process; }
- // NOTE: This returns a null-terminated string.
- StringView name() const
- {
- // NOTE: Whoever is calling this needs to be holding our lock while reading the name.
- VERIFY(m_lock.own_lock());
- return m_name ? m_name->view() : StringView {};
- }
- void set_name(OwnPtr<KString> name)
- {
- ScopedSpinLock lock(m_lock);
- m_name = move(name);
- }
- void finalize();
- enum State : u8 {
- Invalid = 0,
- Runnable,
- Running,
- Dying,
- Dead,
- Stopped,
- Blocked
- };
- class [[nodiscard]] BlockResult {
- public:
- enum Type {
- WokeNormally,
- NotBlocked,
- InterruptedBySignal,
- InterruptedByDeath,
- InterruptedByTimeout,
- };
- BlockResult() = delete;
- BlockResult(Type type)
- : m_type(type)
- {
- }
- bool operator==(Type type) const
- {
- return m_type == type;
- }
- bool operator!=(Type type) const
- {
- return m_type != type;
- }
- [[nodiscard]] bool was_interrupted() const
- {
- switch (m_type) {
- case InterruptedBySignal:
- case InterruptedByDeath:
- return true;
- default:
- return false;
- }
- }
- [[nodiscard]] bool timed_out() const
- {
- return m_type == InterruptedByTimeout;
- }
- private:
- Type m_type;
- };
- class BlockTimeout {
- public:
- BlockTimeout()
- : m_infinite(true)
- {
- }
- explicit BlockTimeout(bool is_absolute, const Time* time, const Time* start_time = nullptr, clockid_t clock_id = CLOCK_MONOTONIC_COARSE);
- const Time& absolute_time() const { return m_time; }
- const Time* start_time() const { return !m_infinite ? &m_start_time : nullptr; }
- clockid_t clock_id() const { return m_clock_id; }
- bool is_infinite() const { return m_infinite; }
- bool should_block() const { return m_infinite || m_should_block; };
- private:
- Time m_time {};
- Time m_start_time {};
- clockid_t m_clock_id { CLOCK_MONOTONIC_COARSE };
- bool m_infinite { false };
- bool m_should_block { false };
- };
- class BlockCondition;
- class Blocker {
- public:
- enum class Type {
- Unknown = 0,
- File,
- Futex,
- Plan9FS,
- Join,
- Queue,
- Routing,
- Sleep,
- Wait
- };
- virtual ~Blocker();
- virtual StringView state_string() const = 0;
- virtual bool should_block() { return true; }
- virtual Type blocker_type() const = 0;
- virtual const BlockTimeout& override_timeout(const BlockTimeout& timeout) { return timeout; }
- virtual bool can_be_interrupted() const { return true; }
- virtual void not_blocking(bool) = 0;
- virtual void was_unblocked(bool did_timeout)
- {
- if (did_timeout) {
- ScopedSpinLock lock(m_lock);
- m_did_timeout = true;
- }
- }
- void set_interrupted_by_death()
- {
- ScopedSpinLock lock(m_lock);
- do_set_interrupted_by_death();
- }
- void set_interrupted_by_signal(u8 signal)
- {
- ScopedSpinLock lock(m_lock);
- do_set_interrupted_by_signal(signal);
- }
- u8 was_interrupted_by_signal() const
- {
- ScopedSpinLock lock(m_lock);
- return do_get_interrupted_by_signal();
- }
- virtual Thread::BlockResult block_result()
- {
- ScopedSpinLock lock(m_lock);
- if (m_was_interrupted_by_death)
- return Thread::BlockResult::InterruptedByDeath;
- if (m_was_interrupted_by_signal != 0)
- return Thread::BlockResult::InterruptedBySignal;
- if (m_did_timeout)
- return Thread::BlockResult::InterruptedByTimeout;
- return Thread::BlockResult::WokeNormally;
- }
- void begin_blocking(Badge<Thread>);
- BlockResult end_blocking(Badge<Thread>, bool);
- protected:
- void do_set_interrupted_by_death()
- {
- m_was_interrupted_by_death = true;
- }
- void do_set_interrupted_by_signal(u8 signal)
- {
- VERIFY(signal != 0);
- m_was_interrupted_by_signal = signal;
- }
- void do_clear_interrupted_by_signal()
- {
- m_was_interrupted_by_signal = 0;
- }
- u8 do_get_interrupted_by_signal() const
- {
- return m_was_interrupted_by_signal;
- }
- [[nodiscard]] bool was_interrupted() const
- {
- return m_was_interrupted_by_death || m_was_interrupted_by_signal != 0;
- }
- void unblock_from_blocker()
- {
- RefPtr<Thread> thread;
- {
- ScopedSpinLock lock(m_lock);
- if (m_is_blocking) {
- m_is_blocking = false;
- VERIFY(m_blocked_thread);
- thread = m_blocked_thread;
- }
- }
- if (thread)
- thread->unblock_from_blocker(*this);
- }
- bool set_block_condition(BlockCondition&, void* = nullptr);
- void set_block_condition_raw_locked(BlockCondition* block_condition)
- {
- m_block_condition = block_condition;
- }
- mutable RecursiveSpinLock m_lock;
- private:
- BlockCondition* m_block_condition { nullptr };
- void* m_block_data { nullptr };
- Thread* m_blocked_thread { nullptr };
- u8 m_was_interrupted_by_signal { 0 };
- bool m_is_blocking { false };
- bool m_was_interrupted_by_death { false };
- bool m_did_timeout { false };
- };
- class BlockCondition {
- AK_MAKE_NONCOPYABLE(BlockCondition);
- AK_MAKE_NONMOVABLE(BlockCondition);
- public:
- BlockCondition() = default;
- virtual ~BlockCondition()
- {
- ScopedSpinLock lock(m_lock);
- VERIFY(m_blockers.is_empty());
- }
- bool add_blocker(Blocker& blocker, void* data)
- {
- ScopedSpinLock lock(m_lock);
- if (!should_add_blocker(blocker, data))
- return false;
- m_blockers.append({ &blocker, data });
- return true;
- }
- void remove_blocker(Blocker& blocker, void* data)
- {
- ScopedSpinLock lock(m_lock);
- // NOTE: it's possible that the blocker is no longer present
- m_blockers.remove_first_matching([&](auto& info) {
- return info.blocker == &blocker && info.data == data;
- });
- }
- bool is_empty() const
- {
- ScopedSpinLock lock(m_lock);
- return is_empty_locked();
- }
- protected:
- template<typename UnblockOne>
- bool unblock(UnblockOne unblock_one)
- {
- ScopedSpinLock lock(m_lock);
- return do_unblock(unblock_one);
- }
- template<typename UnblockOne>
- bool do_unblock(UnblockOne unblock_one)
- {
- VERIFY(m_lock.is_locked());
- bool stop_iterating = false;
- bool did_unblock = false;
- for (size_t i = 0; i < m_blockers.size() && !stop_iterating;) {
- auto& info = m_blockers[i];
- if (unblock_one(*info.blocker, info.data, stop_iterating)) {
- m_blockers.remove(i);
- did_unblock = true;
- continue;
- }
- i++;
- }
- return did_unblock;
- }
- bool is_empty_locked() const
- {
- VERIFY(m_lock.is_locked());
- return m_blockers.is_empty();
- }
- virtual bool should_add_blocker(Blocker&, void*) { return true; }
- struct BlockerInfo {
- Blocker* blocker;
- void* data;
- };
- Vector<BlockerInfo, 4> do_take_blockers(size_t count)
- {
- if (m_blockers.size() <= count)
- return move(m_blockers);
- size_t move_count = (count <= m_blockers.size()) ? count : m_blockers.size();
- VERIFY(move_count > 0);
- Vector<BlockerInfo, 4> taken_blockers;
- taken_blockers.ensure_capacity(move_count);
- for (size_t i = 0; i < move_count; i++)
- taken_blockers.append(m_blockers.take(i));
- m_blockers.remove(0, move_count);
- return taken_blockers;
- }
- void do_append_blockers(Vector<BlockerInfo, 4>&& blockers_to_append)
- {
- if (blockers_to_append.is_empty())
- return;
- if (m_blockers.is_empty()) {
- m_blockers = move(blockers_to_append);
- return;
- }
- m_blockers.ensure_capacity(m_blockers.size() + blockers_to_append.size());
- for (size_t i = 0; i < blockers_to_append.size(); i++)
- m_blockers.append(blockers_to_append.take(i));
- blockers_to_append.clear();
- }
- mutable SpinLock<u8> m_lock;
- private:
- Vector<BlockerInfo, 4> m_blockers;
- };
- friend class JoinBlocker;
- class JoinBlocker final : public Blocker {
- public:
- explicit JoinBlocker(Thread& joinee, KResult& try_join_result, void*& joinee_exit_value);
- virtual Type blocker_type() const override { return Type::Join; }
- virtual StringView state_string() const override { return "Joining"sv; }
- virtual bool can_be_interrupted() const override { return false; }
- virtual bool should_block() override { return !m_join_error && m_should_block; }
- virtual void not_blocking(bool) override;
- bool unblock(void*, bool);
- private:
- NonnullRefPtr<Thread> m_joinee;
- void*& m_joinee_exit_value;
- bool m_join_error { false };
- bool m_did_unblock { false };
- bool m_should_block { true };
- };
- class QueueBlocker : public Blocker {
- public:
- explicit QueueBlocker(WaitQueue&, StringView block_reason = {});
- virtual ~QueueBlocker();
- virtual Type blocker_type() const override { return Type::Queue; }
- virtual StringView state_string() const override { return m_block_reason.is_null() ? m_block_reason : "Queue"sv; }
- virtual void not_blocking(bool) override { }
- virtual bool should_block() override
- {
- return m_should_block;
- }
- bool unblock();
- protected:
- StringView m_block_reason;
- bool m_should_block { true };
- bool m_did_unblock { false };
- };
- class FutexBlocker : public Blocker {
- public:
- explicit FutexBlocker(FutexQueue&, u32);
- virtual ~FutexBlocker();
- virtual Type blocker_type() const override { return Type::Futex; }
- virtual StringView state_string() const override { return "Futex"sv; }
- virtual void not_blocking(bool) override { }
- virtual bool should_block() override
- {
- return m_should_block;
- }
- u32 bitset() const { return m_bitset; }
- void begin_requeue()
- {
- // We need to hold the lock until we moved it over
- m_relock_flags = m_lock.lock();
- }
- void finish_requeue(FutexQueue&);
- bool unblock_bitset(u32 bitset);
- bool unblock(bool force = false);
- protected:
- u32 m_bitset;
- u32 m_relock_flags { 0 };
- bool m_should_block { true };
- bool m_did_unblock { false };
- };
- class FileBlocker : public Blocker {
- public:
- enum class BlockFlags : u16 {
- None = 0,
- Read = 1 << 0,
- Write = 1 << 1,
- ReadPriority = 1 << 2,
- Accept = 1 << 3,
- Connect = 1 << 4,
- SocketFlags = Accept | Connect,
- WriteNotOpen = 1 << 5,
- WriteError = 1 << 6,
- WriteHangUp = 1 << 7,
- ReadHangUp = 1 << 8,
- Exception = WriteNotOpen | WriteError | WriteHangUp | ReadHangUp,
- };
- virtual Type blocker_type() const override { return Type::File; }
- virtual bool should_block() override
- {
- return m_should_block;
- }
- virtual bool unblock(bool, void*) = 0;
- protected:
- bool m_should_block { true };
- };
- class FileDescriptionBlocker : public FileBlocker {
- public:
- const FileDescription& blocked_description() const;
- virtual bool unblock(bool, void*) override;
- virtual void not_blocking(bool) override;
- protected:
- explicit FileDescriptionBlocker(FileDescription&, BlockFlags, BlockFlags&);
- private:
- NonnullRefPtr<FileDescription> m_blocked_description;
- const BlockFlags m_flags;
- BlockFlags& m_unblocked_flags;
- bool m_did_unblock { false };
- };
- class AcceptBlocker final : public FileDescriptionBlocker {
- public:
- explicit AcceptBlocker(FileDescription&, BlockFlags&);
- virtual StringView state_string() const override { return "Accepting"sv; }
- };
- class ConnectBlocker final : public FileDescriptionBlocker {
- public:
- explicit ConnectBlocker(FileDescription&, BlockFlags&);
- virtual StringView state_string() const override { return "Connecting"sv; }
- };
- class WriteBlocker final : public FileDescriptionBlocker {
- public:
- explicit WriteBlocker(FileDescription&, BlockFlags&);
- virtual StringView state_string() const override { return "Writing"sv; }
- virtual const BlockTimeout& override_timeout(const BlockTimeout&) override;
- private:
- BlockTimeout m_timeout;
- };
- class ReadBlocker final : public FileDescriptionBlocker {
- public:
- explicit ReadBlocker(FileDescription&, BlockFlags&);
- virtual StringView state_string() const override { return "Reading"sv; }
- virtual const BlockTimeout& override_timeout(const BlockTimeout&) override;
- private:
- BlockTimeout m_timeout;
- };
- class SleepBlocker final : public Blocker {
- public:
- explicit SleepBlocker(const BlockTimeout&, Time* = nullptr);
- virtual StringView state_string() const override { return "Sleeping"sv; }
- virtual Type blocker_type() const override { return Type::Sleep; }
- virtual const BlockTimeout& override_timeout(const BlockTimeout&) override;
- virtual void not_blocking(bool) override;
- virtual void was_unblocked(bool) override;
- virtual Thread::BlockResult block_result() override;
- private:
- void calculate_remaining();
- BlockTimeout m_deadline;
- Time* m_remaining;
- };
- class SelectBlocker final : public FileBlocker {
- public:
- struct FDInfo {
- NonnullRefPtr<FileDescription> description;
- BlockFlags block_flags { BlockFlags::None };
- BlockFlags unblocked_flags { BlockFlags::None };
- };
- typedef Vector<FDInfo, FD_SETSIZE> FDVector;
- SelectBlocker(FDVector& fds);
- virtual ~SelectBlocker();
- virtual bool unblock(bool, void*) override;
- virtual void not_blocking(bool) override;
- virtual void was_unblocked(bool) override;
- virtual StringView state_string() const override { return "Selecting"sv; }
- private:
- size_t collect_unblocked_flags();
- FDVector& m_fds;
- bool m_did_unblock { false };
- };
- class WaitBlocker final : public Blocker {
- public:
- enum class UnblockFlags {
- Terminated,
- Stopped,
- Continued,
- Disowned
- };
- WaitBlocker(int wait_options, idtype_t id_type, pid_t id, KResultOr<siginfo_t>& result);
- virtual StringView state_string() const override { return "Waiting"sv; }
- virtual Type blocker_type() const override { return Type::Wait; }
- virtual bool should_block() override { return m_should_block; }
- virtual void not_blocking(bool) override;
- virtual void was_unblocked(bool) override;
- bool unblock(Process& process, UnblockFlags flags, u8 signal, bool from_add_blocker);
- bool is_wait() const { return !(m_wait_options & WNOWAIT); }
- private:
- void do_was_disowned();
- void do_set_result(const siginfo_t&);
- const int m_wait_options;
- const idtype_t m_id_type;
- const pid_t m_waitee_id;
- KResultOr<siginfo_t>& m_result;
- RefPtr<Process> m_waitee;
- RefPtr<ProcessGroup> m_waitee_group;
- bool m_did_unblock { false };
- bool m_error { false };
- bool m_got_sigchild { false };
- bool m_should_block;
- };
- class WaitBlockCondition final : public BlockCondition {
- friend class WaitBlocker;
- public:
- WaitBlockCondition(Process& process)
- : m_process(process)
- {
- }
- void disowned_by_waiter(Process&);
- bool unblock(Process&, WaitBlocker::UnblockFlags, u8);
- void try_unblock(WaitBlocker&);
- void finalize();
- protected:
- virtual bool should_add_blocker(Blocker&, void*) override;
- private:
- struct ProcessBlockInfo {
- NonnullRefPtr<Process> process;
- WaitBlocker::UnblockFlags flags;
- u8 signal;
- bool was_waited { false };
- explicit ProcessBlockInfo(NonnullRefPtr<Process>&&, WaitBlocker::UnblockFlags, u8);
- ~ProcessBlockInfo();
- };
- Process& m_process;
- Vector<ProcessBlockInfo, 2> m_processes;
- bool m_finalized { false };
- };
- template<typename AddBlockerHandler>
- KResult try_join(AddBlockerHandler add_blocker)
- {
- if (Thread::current() == this)
- return EDEADLK;
- ScopedSpinLock lock(m_lock);
- if (!m_is_joinable || state() == Dead)
- return EINVAL;
- add_blocker();
- // From this point on the thread is no longer joinable by anyone
- // else. It also means that if the join is timed, it becomes
- // detached when a timeout happens.
- m_is_joinable = false;
- return KSuccess;
- }
- void did_schedule() { ++m_times_scheduled; }
- u32 times_scheduled() const { return m_times_scheduled; }
- void resume_from_stopped();
- [[nodiscard]] bool should_be_stopped() const;
- [[nodiscard]] bool is_stopped() const { return m_state == Stopped; }
- [[nodiscard]] bool is_blocked() const { return m_state == Blocked; }
- [[nodiscard]] bool is_in_block() const
- {
- ScopedSpinLock lock(m_block_lock);
- return m_in_block;
- }
- u32 cpu() const { return m_cpu.load(AK::MemoryOrder::memory_order_consume); }
- void set_cpu(u32 cpu) { m_cpu.store(cpu, AK::MemoryOrder::memory_order_release); }
- u32 affinity() const { return m_cpu_affinity; }
- void set_affinity(u32 affinity) { m_cpu_affinity = affinity; }
- RegisterState& get_register_dump_from_stack();
- const RegisterState& get_register_dump_from_stack() const { return const_cast<Thread*>(this)->get_register_dump_from_stack(); }
- DebugRegisterState& debug_register_state() { return m_debug_register_state; }
- const DebugRegisterState& debug_register_state() const { return m_debug_register_state; }
- ThreadRegisters& regs() { return m_regs; }
- ThreadRegisters const& regs() const { return m_regs; }
- State state() const { return m_state; }
- StringView state_string() const;
- VirtualAddress thread_specific_data() const { return m_thread_specific_data; }
- size_t thread_specific_region_size() const;
- size_t thread_specific_region_alignment() const;
- ALWAYS_INLINE void yield_if_stopped()
- {
- // If some thread stopped us, we need to yield to someone else
- // We check this when entering/exiting a system call. A thread
- // may continue to execute in user land until the next timer
- // tick or entering the next system call, or if it's in kernel
- // mode then we will intercept prior to returning back to user
- // mode.
- ScopedSpinLock lock(m_lock);
- while (state() == Thread::Stopped) {
- lock.unlock();
- // We shouldn't be holding the big lock here
- yield_assuming_not_holding_big_lock();
- lock.lock();
- }
- }
- void block(Kernel::Mutex&, ScopedSpinLock<SpinLock<u8>>&, u32);
- template<typename BlockerType, class... Args>
- [[nodiscard]] BlockResult block(const BlockTimeout& timeout, Args&&... args)
- {
- VERIFY(!Processor::current().in_irq());
- VERIFY(this == Thread::current());
- ScopedCritical critical;
- VERIFY(!Memory::s_mm_lock.own_lock());
- ScopedSpinLock block_lock(m_block_lock);
- // We need to hold m_block_lock so that nobody can unblock a blocker as soon
- // as it is constructed and registered elsewhere
- m_in_block = true;
- BlockerType blocker(forward<Args>(args)...);
- ScopedSpinLock scheduler_lock(g_scheduler_lock);
- // Relaxed semantics are fine for timeout_unblocked because we
- // synchronize on the spin locks already.
- Atomic<bool, AK::MemoryOrder::memory_order_relaxed> timeout_unblocked(false);
- bool timer_was_added = false;
- {
- switch (state()) {
- case Thread::Stopped:
- // It's possible that we were requested to be stopped!
- break;
- case Thread::Running:
- VERIFY(m_blocker == nullptr);
- break;
- default:
- VERIFY_NOT_REACHED();
- }
- m_blocker = &blocker;
- if (!blocker.should_block()) {
- // Don't block if the wake condition is already met
- blocker.not_blocking(false);
- m_blocker = nullptr;
- m_in_block = false;
- return BlockResult::NotBlocked;
- }
- auto& block_timeout = blocker.override_timeout(timeout);
- if (!block_timeout.is_infinite()) {
- // Process::kill_all_threads may be called at any time, which will mark all
- // threads to die. In that case
- timer_was_added = TimerQueue::the().add_timer_without_id(*m_block_timer, block_timeout.clock_id(), block_timeout.absolute_time(), [&]() {
- VERIFY(!Processor::current().in_irq());
- VERIFY(!g_scheduler_lock.own_lock());
- VERIFY(!m_block_lock.own_lock());
- // NOTE: this may execute on the same or any other processor!
- ScopedSpinLock scheduler_lock(g_scheduler_lock);
- ScopedSpinLock block_lock(m_block_lock);
- if (m_blocker && timeout_unblocked.exchange(true) == false)
- unblock();
- });
- if (!timer_was_added) {
- // Timeout is already in the past
- blocker.not_blocking(true);
- m_blocker = nullptr;
- m_in_block = false;
- return BlockResult::InterruptedByTimeout;
- }
- }
- blocker.begin_blocking({});
- set_state(Thread::Blocked);
- }
- scheduler_lock.unlock();
- block_lock.unlock();
- dbgln_if(THREAD_DEBUG, "Thread {} blocking on {} ({}) -->", *this, &blocker, blocker.state_string());
- bool did_timeout = false;
- u32 lock_count_to_restore = 0;
- auto previous_locked = unlock_process_if_locked(lock_count_to_restore);
- for (;;) {
- // Yield to the scheduler, and wait for us to resume unblocked.
- VERIFY(!g_scheduler_lock.own_lock());
- VERIFY(Processor::current().in_critical());
- yield_assuming_not_holding_big_lock();
- VERIFY(Processor::current().in_critical());
- ScopedSpinLock block_lock2(m_block_lock);
- if (should_be_stopped() || state() == Stopped) {
- dbgln("Thread should be stopped, current state: {}", state_string());
- set_state(Thread::Blocked);
- continue;
- }
- if (m_blocker && !m_blocker->can_be_interrupted() && !m_should_die) {
- block_lock2.unlock();
- dbgln("Thread should not be unblocking, current state: {}", state_string());
- set_state(Thread::Blocked);
- continue;
- }
- // Prevent the timeout from unblocking this thread if it happens to
- // be in the process of firing already
- did_timeout |= timeout_unblocked.exchange(true);
- if (m_blocker) {
- // Remove ourselves...
- VERIFY(m_blocker == &blocker);
- m_blocker = nullptr;
- }
- dbgln_if(THREAD_DEBUG, "<-- Thread {} unblocked from {} ({})", *this, &blocker, blocker.state_string());
- m_in_block = false;
- break;
- }
- if (blocker.was_interrupted_by_signal()) {
- ScopedSpinLock scheduler_lock(g_scheduler_lock);
- ScopedSpinLock lock(m_lock);
- dispatch_one_pending_signal();
- }
- // Notify the blocker that we are no longer blocking. It may need
- // to clean up now while we're still holding m_lock
- auto result = blocker.end_blocking({}, did_timeout); // calls was_unblocked internally
- if (timer_was_added && !did_timeout) {
- // Cancel the timer while not holding any locks. This allows
- // the timer function to complete before we remove it
- // (e.g. if it's on another processor)
- TimerQueue::the().cancel_timer(*m_block_timer);
- }
- if (previous_locked != LockMode::Unlocked) {
- // NOTE: this may trigger another call to Thread::block(), so
- // we need to do this after we're all done and restored m_in_block!
- relock_process(previous_locked, lock_count_to_restore);
- }
- return result;
- }
- u32 unblock_from_lock(Kernel::Mutex&);
- void unblock_from_blocker(Blocker&);
- void unblock(u8 signal = 0);
- template<class... Args>
- Thread::BlockResult wait_on(WaitQueue& wait_queue, const Thread::BlockTimeout& timeout, Args&&... args)
- {
- VERIFY(this == Thread::current());
- return block<Thread::QueueBlocker>(timeout, wait_queue, forward<Args>(args)...);
- }
- BlockResult sleep(clockid_t, const Time&, Time* = nullptr);
- BlockResult sleep(const Time& duration, Time* remaining_time = nullptr)
- {
- return sleep(CLOCK_MONOTONIC_COARSE, duration, remaining_time);
- }
- BlockResult sleep_until(clockid_t, const Time&);
- BlockResult sleep_until(const Time& duration)
- {
- return sleep_until(CLOCK_MONOTONIC_COARSE, duration);
- }
- // Tell this thread to unblock if needed,
- // gracefully unwind the stack and die.
- void set_should_die();
- [[nodiscard]] bool should_die() const { return m_should_die; }
- void die_if_needed();
- void exit(void* = nullptr);
- void update_time_scheduled(u64, bool, bool);
- bool tick();
- void set_ticks_left(u32 t) { m_ticks_left = t; }
- u32 ticks_left() const { return m_ticks_left; }
- FlatPtr kernel_stack_base() const { return m_kernel_stack_base; }
- FlatPtr kernel_stack_top() const { return m_kernel_stack_top; }
- void set_state(State, u8 = 0);
- [[nodiscard]] bool is_initialized() const { return m_initialized; }
- void set_initialized(bool initialized) { m_initialized = initialized; }
- void send_urgent_signal_to_self(u8 signal);
- void send_signal(u8 signal, Process* sender);
- u32 update_signal_mask(u32 signal_mask);
- u32 signal_mask_block(sigset_t signal_set, bool block);
- u32 signal_mask() const;
- void clear_signals();
- KResultOr<u32> peek_debug_register(u32 register_index);
- KResult poke_debug_register(u32 register_index, u32 data);
- void set_dump_backtrace_on_finalization() { m_dump_backtrace_on_finalization = true; }
- DispatchSignalResult dispatch_one_pending_signal();
- DispatchSignalResult try_dispatch_one_pending_signal(u8 signal);
- DispatchSignalResult dispatch_signal(u8 signal);
- void check_dispatch_pending_signal();
- [[nodiscard]] bool has_unmasked_pending_signals() const { return m_have_any_unmasked_pending_signals.load(AK::memory_order_consume); }
- [[nodiscard]] bool should_ignore_signal(u8 signal) const;
- [[nodiscard]] bool has_signal_handler(u8 signal) const;
- u32 pending_signals() const;
- u32 pending_signals_for_state() const;
- FPUState& fpu_state() { return m_fpu_state; }
- KResult make_thread_specific_region(Badge<Process>);
- unsigned syscall_count() const { return m_syscall_count; }
- void did_syscall() { ++m_syscall_count; }
- unsigned inode_faults() const { return m_inode_faults; }
- void did_inode_fault() { ++m_inode_faults; }
- unsigned zero_faults() const { return m_zero_faults; }
- void did_zero_fault() { ++m_zero_faults; }
- unsigned cow_faults() const { return m_cow_faults; }
- void did_cow_fault() { ++m_cow_faults; }
- unsigned file_read_bytes() const { return m_file_read_bytes; }
- unsigned file_write_bytes() const { return m_file_write_bytes; }
- void did_file_read(unsigned bytes)
- {
- m_file_read_bytes += bytes;
- }
- void did_file_write(unsigned bytes)
- {
- m_file_write_bytes += bytes;
- }
- unsigned unix_socket_read_bytes() const { return m_unix_socket_read_bytes; }
- unsigned unix_socket_write_bytes() const { return m_unix_socket_write_bytes; }
- void did_unix_socket_read(unsigned bytes)
- {
- m_unix_socket_read_bytes += bytes;
- }
- void did_unix_socket_write(unsigned bytes)
- {
- m_unix_socket_write_bytes += bytes;
- }
- unsigned ipv4_socket_read_bytes() const { return m_ipv4_socket_read_bytes; }
- unsigned ipv4_socket_write_bytes() const { return m_ipv4_socket_write_bytes; }
- void did_ipv4_socket_read(unsigned bytes)
- {
- m_ipv4_socket_read_bytes += bytes;
- }
- void did_ipv4_socket_write(unsigned bytes)
- {
- m_ipv4_socket_write_bytes += bytes;
- }
- void set_active(bool active) { m_is_active = active; }
- u32 saved_critical() const { return m_saved_critical; }
- void save_critical(u32 critical) { m_saved_critical = critical; }
- [[nodiscard]] bool is_active() const { return m_is_active; }
- [[nodiscard]] bool is_finalizable() const
- {
- // We can't finalize as long as this thread is still running
- // Note that checking for Running state here isn't sufficient
- // as the thread may not be in Running state but switching out.
- // m_is_active is set to false once the context switch is
- // complete and the thread is not executing on any processor.
- if (m_is_active.load(AK::memory_order_acquire))
- return false;
- // We can't finalize until the thread is either detached or
- // a join has started. We can't make m_is_joinable atomic
- // because that would introduce a race in try_join.
- ScopedSpinLock lock(m_lock);
- return !m_is_joinable;
- }
- RefPtr<Thread> clone(Process&);
- template<IteratorFunction<Thread&> Callback>
- static IterationDecision for_each_in_state(State, Callback);
- template<IteratorFunction<Thread&> Callback>
- static IterationDecision for_each(Callback);
- template<VoidFunction<Thread&> Callback>
- static IterationDecision for_each_in_state(State, Callback);
- template<VoidFunction<Thread&> Callback>
- static IterationDecision for_each(Callback);
- static constexpr u32 default_kernel_stack_size = 65536;
- static constexpr u32 default_userspace_stack_size = 1 * MiB;
- u64 time_in_user() const { return m_total_time_scheduled_user; }
- u64 time_in_kernel() const { return m_total_time_scheduled_kernel; }
- enum class PreviousMode : u8 {
- KernelMode = 0,
- UserMode
- };
- PreviousMode previous_mode() const { return m_previous_mode; }
- bool set_previous_mode(PreviousMode mode)
- {
- if (m_previous_mode == mode)
- return false;
- m_previous_mode = mode;
- return true;
- }
- TrapFrame*& current_trap() { return m_current_trap; }
- RecursiveSpinLock& get_lock() const { return m_lock; }
- #if LOCK_DEBUG
- void holding_lock(Mutex& lock, int refs_delta, const SourceLocation& location)
- {
- VERIFY(refs_delta != 0);
- m_holding_locks.fetch_add(refs_delta, AK::MemoryOrder::memory_order_relaxed);
- ScopedSpinLock list_lock(m_holding_locks_lock);
- if (refs_delta > 0) {
- bool have_existing = false;
- for (size_t i = 0; i < m_holding_locks_list.size(); i++) {
- auto& info = m_holding_locks_list[i];
- if (info.lock == &lock) {
- have_existing = true;
- info.count += refs_delta;
- break;
- }
- }
- if (!have_existing)
- m_holding_locks_list.append({ &lock, location, 1 });
- } else {
- VERIFY(refs_delta < 0);
- bool found = false;
- for (size_t i = 0; i < m_holding_locks_list.size(); i++) {
- auto& info = m_holding_locks_list[i];
- if (info.lock == &lock) {
- VERIFY(info.count >= (unsigned)-refs_delta);
- info.count -= (unsigned)-refs_delta;
- if (info.count == 0)
- m_holding_locks_list.remove(i);
- found = true;
- break;
- }
- }
- VERIFY(found);
- }
- }
- u32 lock_count() const
- {
- return m_holding_locks.load(AK::MemoryOrder::memory_order_relaxed);
- }
- #endif
- bool is_handling_page_fault() const
- {
- return m_handling_page_fault;
- }
- void set_handling_page_fault(bool b) { m_handling_page_fault = b; }
- void set_idle_thread() { m_is_idle_thread = true; }
- bool is_idle_thread() const { return m_is_idle_thread; }
- ALWAYS_INLINE u32 enter_profiler()
- {
- return m_nested_profiler_calls.fetch_add(1, AK::MemoryOrder::memory_order_acq_rel);
- }
- ALWAYS_INLINE u32 leave_profiler()
- {
- return m_nested_profiler_calls.fetch_sub(1, AK::MemoryOrder::memory_order_acquire);
- }
- bool is_profiling_suppressed() const { return m_is_profiling_suppressed; }
- void set_profiling_suppressed() { m_is_profiling_suppressed = true; }
- InodeIndex global_procfs_inode_index() const { return m_global_procfs_inode_index; }
- String backtrace();
- private:
- Thread(NonnullRefPtr<Process>, NonnullOwnPtr<Memory::Region>, NonnullRefPtr<Timer>, OwnPtr<KString>);
- IntrusiveListNode<Thread> m_process_thread_list_node;
- int m_runnable_priority { -1 };
- friend class WaitQueue;
- class JoinBlockCondition : public BlockCondition {
- public:
- void thread_did_exit(void* exit_value)
- {
- ScopedSpinLock lock(m_lock);
- VERIFY(!m_thread_did_exit);
- m_thread_did_exit = true;
- m_exit_value.store(exit_value, AK::MemoryOrder::memory_order_release);
- do_unblock_joiner();
- }
- void thread_finalizing()
- {
- ScopedSpinLock lock(m_lock);
- do_unblock_joiner();
- }
- void* exit_value() const
- {
- VERIFY(m_thread_did_exit);
- return m_exit_value.load(AK::MemoryOrder::memory_order_acquire);
- }
- void try_unblock(JoinBlocker& blocker)
- {
- ScopedSpinLock lock(m_lock);
- if (m_thread_did_exit)
- blocker.unblock(exit_value(), false);
- }
- protected:
- virtual bool should_add_blocker(Blocker& b, void*) override
- {
- VERIFY(b.blocker_type() == Blocker::Type::Join);
- auto& blocker = static_cast<JoinBlocker&>(b);
- // NOTE: m_lock is held already!
- if (m_thread_did_exit) {
- blocker.unblock(exit_value(), true);
- return false;
- }
- return true;
- }
- private:
- void do_unblock_joiner()
- {
- do_unblock([&](Blocker& b, void*, bool&) {
- VERIFY(b.blocker_type() == Blocker::Type::Join);
- auto& blocker = static_cast<JoinBlocker&>(b);
- return blocker.unblock(exit_value(), false);
- });
- }
- Atomic<void*> m_exit_value { nullptr };
- bool m_thread_did_exit { false };
- };
- LockMode unlock_process_if_locked(u32&);
- void relock_process(LockMode, u32);
- void reset_fpu_state();
- mutable RecursiveSpinLock m_lock;
- mutable RecursiveSpinLock m_block_lock;
- NonnullRefPtr<Process> m_process;
- ThreadID m_tid { -1 };
- ThreadRegisters m_regs;
- DebugRegisterState m_debug_register_state {};
- TrapFrame* m_current_trap { nullptr };
- u32 m_saved_critical { 1 };
- IntrusiveListNode<Thread> m_ready_queue_node;
- Atomic<u32> m_cpu { 0 };
- u32 m_cpu_affinity { THREAD_AFFINITY_DEFAULT };
- Optional<u64> m_last_time_scheduled;
- u64 m_total_time_scheduled_user { 0 };
- u64 m_total_time_scheduled_kernel { 0 };
- u32 m_ticks_left { 0 };
- u32 m_times_scheduled { 0 };
- u32 m_ticks_in_user { 0 };
- u32 m_ticks_in_kernel { 0 };
- u32 m_pending_signals { 0 };
- u32 m_signal_mask { 0 };
- FlatPtr m_kernel_stack_base { 0 };
- FlatPtr m_kernel_stack_top { 0 };
- OwnPtr<Memory::Region> m_kernel_stack_region;
- VirtualAddress m_thread_specific_data;
- Optional<Memory::VirtualRange> m_thread_specific_range;
- Array<SignalActionData, NSIG> m_signal_action_data;
- Blocker* m_blocker { nullptr };
- Kernel::Mutex* m_blocking_lock { nullptr };
- u32 m_lock_requested_count { 0 };
- IntrusiveListNode<Thread> m_blocked_threads_list_node;
- #if LOCK_DEBUG
- struct HoldingLockInfo {
- Mutex* lock;
- SourceLocation source_location;
- unsigned count;
- };
- Atomic<u32> m_holding_locks { 0 };
- SpinLock<u8> m_holding_locks_lock;
- Vector<HoldingLockInfo> m_holding_locks_list;
- #endif
- JoinBlockCondition m_join_condition;
- Atomic<bool, AK::MemoryOrder::memory_order_relaxed> m_is_active { false };
- bool m_is_joinable { true };
- bool m_handling_page_fault { false };
- PreviousMode m_previous_mode { PreviousMode::KernelMode }; // We always start out in kernel mode
- unsigned m_syscall_count { 0 };
- unsigned m_inode_faults { 0 };
- unsigned m_zero_faults { 0 };
- unsigned m_cow_faults { 0 };
- unsigned m_file_read_bytes { 0 };
- unsigned m_file_write_bytes { 0 };
- unsigned m_unix_socket_read_bytes { 0 };
- unsigned m_unix_socket_write_bytes { 0 };
- unsigned m_ipv4_socket_read_bytes { 0 };
- unsigned m_ipv4_socket_write_bytes { 0 };
- FPUState m_fpu_state {};
- State m_state { Invalid };
- OwnPtr<KString> m_name;
- u32 m_priority { THREAD_PRIORITY_NORMAL };
- State m_stop_state { Invalid };
- bool m_dump_backtrace_on_finalization { false };
- bool m_should_die { false };
- bool m_initialized { false };
- bool m_in_block { false };
- bool m_is_idle_thread { false };
- Atomic<bool> m_have_any_unmasked_pending_signals { false };
- Atomic<u32> m_nested_profiler_calls { 0 };
- RefPtr<Timer> m_block_timer;
- // Note: This is needed so when we generate thread stack inodes for ProcFS, we know that
- // we assigned a global Inode index to it so we can use it later
- InodeIndex m_global_procfs_inode_index;
- bool m_is_profiling_suppressed { false };
- void yield_and_release_relock_big_lock();
- void yield_assuming_not_holding_big_lock();
- void drop_thread_count(bool);
- public:
- using ListInProcess = IntrusiveList<Thread, RawPtr<Thread>, &Thread::m_process_thread_list_node>;
- };
- AK_ENUM_BITWISE_OPERATORS(Thread::FileBlocker::BlockFlags);
- template<IteratorFunction<Thread&> Callback>
- inline IterationDecision Thread::for_each(Callback callback)
- {
- ScopedSpinLock lock(g_tid_map_lock);
- for (auto& it : *g_tid_map) {
- IterationDecision decision = callback(*it.value);
- if (decision != IterationDecision::Continue)
- return decision;
- }
- return IterationDecision::Continue;
- }
- template<IteratorFunction<Thread&> Callback>
- inline IterationDecision Thread::for_each_in_state(State state, Callback callback)
- {
- ScopedSpinLock lock(g_tid_map_lock);
- for (auto& it : *g_tid_map) {
- auto& thread = *it.value;
- if (thread.state() != state)
- continue;
- IterationDecision decision = callback(thread);
- if (decision != IterationDecision::Continue)
- return decision;
- }
- return IterationDecision::Continue;
- }
- template<VoidFunction<Thread&> Callback>
- inline IterationDecision Thread::for_each(Callback callback)
- {
- ScopedSpinLock lock(g_tid_map_lock);
- for (auto& it : *g_tid_map)
- callback(*it.value);
- return IterationDecision::Continue;
- }
- template<VoidFunction<Thread&> Callback>
- inline IterationDecision Thread::for_each_in_state(State state, Callback callback)
- {
- return for_each_in_state(state, [&](auto& thread) {
- callback(thread);
- return IterationDecision::Continue;
- });
- }
- }
- template<>
- struct AK::Formatter<Kernel::Thread> : AK::Formatter<FormatString> {
- void format(FormatBuilder&, const Kernel::Thread&);
- };
|