123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320 |
- /*
- * Copyright (c) 2020, the SerenityOS developers.
- *
- * SPDX-License-Identifier: BSD-2-Clause
- */
- #include <Kernel/Arch/x86/RegisterState.h>
- #include <Kernel/Arch/x86/SafeMem.h>
- #define CODE_SECTION(section_name) __attribute__((section(section_name)))
- extern "C" u8* start_of_safemem_text;
- extern "C" u8* end_of_safemem_text;
- extern "C" u8* safe_memcpy_ins_1;
- extern "C" u8* safe_memcpy_1_faulted;
- extern "C" u8* safe_memcpy_ins_2;
- extern "C" u8* safe_memcpy_2_faulted;
- extern "C" u8* safe_strnlen_ins;
- extern "C" u8* safe_strnlen_faulted;
- extern "C" u8* safe_memset_ins_1;
- extern "C" u8* safe_memset_1_faulted;
- extern "C" u8* safe_memset_ins_2;
- extern "C" u8* safe_memset_2_faulted;
- extern "C" u8* start_of_safemem_atomic_text;
- extern "C" u8* end_of_safemem_atomic_text;
- extern "C" u8* safe_atomic_fetch_add_relaxed_ins;
- extern "C" u8* safe_atomic_fetch_add_relaxed_faulted;
- extern "C" u8* safe_atomic_exchange_relaxed_ins;
- extern "C" u8* safe_atomic_exchange_relaxed_faulted;
- extern "C" u8* safe_atomic_load_relaxed_ins;
- extern "C" u8* safe_atomic_load_relaxed_faulted;
- extern "C" u8* safe_atomic_store_relaxed_ins;
- extern "C" u8* safe_atomic_store_relaxed_faulted;
- extern "C" u8* safe_atomic_compare_exchange_relaxed_ins;
- extern "C" u8* safe_atomic_compare_exchange_relaxed_faulted;
- namespace Kernel {
- CODE_SECTION(".text.safemem")
- NEVER_INLINE bool safe_memcpy(void* dest_ptr, const void* src_ptr, size_t n, void*& fault_at)
- {
- fault_at = nullptr;
- size_t dest = (size_t)dest_ptr;
- size_t src = (size_t)src_ptr;
- size_t remainder;
- // FIXME: Support starting at an unaligned address.
- if (!(dest & 0x3) && !(src & 0x3) && n >= 12) {
- size_t size_ts = n / sizeof(size_t);
- asm volatile(
- "safe_memcpy_ins_1: \n"
- #if ARCH(I386)
- "rep movsl \n"
- #else
- "rep movsq \n"
- #endif
- "safe_memcpy_1_faulted: \n" // handle_safe_access_fault() set edx/rdx to the fault address!
- : "=S"(src),
- "=D"(dest),
- "=c"(remainder),
- [fault_at] "=d"(fault_at)
- : "S"(src),
- "D"(dest),
- "c"(size_ts)
- : "memory");
- if (remainder != 0)
- return false; // fault_at is already set!
- n -= size_ts * sizeof(size_t);
- if (n == 0) {
- fault_at = nullptr;
- return true;
- }
- }
- asm volatile(
- "safe_memcpy_ins_2: \n"
- "rep movsb \n"
- "safe_memcpy_2_faulted: \n" // handle_safe_access_fault() set edx/rdx to the fault address!
- : "=c"(remainder),
- [fault_at] "=d"(fault_at)
- : "S"(src),
- "D"(dest),
- "c"(n)
- : "memory");
- if (remainder != 0)
- return false; // fault_at is already set!
- fault_at = nullptr;
- return true;
- }
- CODE_SECTION(".text.safemem")
- NEVER_INLINE ssize_t safe_strnlen(const char* str, size_t max_n, void*& fault_at)
- {
- ssize_t count = 0;
- fault_at = nullptr;
- asm volatile(
- "1: \n"
- "test %[max_n], %[max_n] \n"
- "je 2f \n"
- "dec %[max_n] \n"
- "safe_strnlen_ins: \n"
- "cmpb $0,(%[str], %[count], 1) \n"
- "je 2f \n"
- "inc %[count] \n"
- "jmp 1b \n"
- "safe_strnlen_faulted: \n" // handle_safe_access_fault() set edx/rdx to the fault address!
- "xor %[count_on_error], %[count_on_error] \n"
- "dec %[count_on_error] \n" // return -1 on fault
- "2:"
- : [count_on_error] "=c"(count),
- [fault_at] "=d"(fault_at)
- : [str] "b"(str),
- [count] "c"(count),
- [max_n] "d"(max_n));
- if (count >= 0)
- fault_at = nullptr;
- return count;
- }
- CODE_SECTION(".text.safemem")
- NEVER_INLINE bool safe_memset(void* dest_ptr, int c, size_t n, void*& fault_at)
- {
- fault_at = nullptr;
- size_t dest = (size_t)dest_ptr;
- size_t remainder;
- // FIXME: Support starting at an unaligned address.
- if (!(dest & 0x3) && n >= 12) {
- size_t size_ts = n / sizeof(size_t);
- size_t expanded_c = (u8)c;
- expanded_c |= expanded_c << 8;
- expanded_c |= expanded_c << 16;
- asm volatile(
- "safe_memset_ins_1: \n"
- #if ARCH(I386)
- "rep stosl \n"
- #else
- "rep stosq \n"
- #endif
- "safe_memset_1_faulted: \n" // handle_safe_access_fault() set edx/rdx to the fault address!
- : "=D"(dest),
- "=c"(remainder),
- [fault_at] "=d"(fault_at)
- : "D"(dest),
- "a"(expanded_c),
- "c"(size_ts)
- : "memory");
- if (remainder != 0)
- return false; // fault_at is already set!
- n -= size_ts * sizeof(size_t);
- if (n == 0) {
- fault_at = nullptr;
- return true;
- }
- }
- asm volatile(
- "safe_memset_ins_2: \n"
- "rep stosb \n"
- "safe_memset_2_faulted: \n" // handle_safe_access_fault() set edx/rdx to the fault address!
- : "=D"(dest),
- "=c"(remainder),
- [fault_at] "=d"(fault_at)
- : "D"(dest),
- "c"(n),
- "a"(c)
- : "memory");
- if (remainder != 0)
- return false; // fault_at is already set!
- fault_at = nullptr;
- return true;
- }
- CODE_SECTION(".text.safemem.atomic")
- NEVER_INLINE Optional<u32> safe_atomic_fetch_add_relaxed(volatile u32* var, u32 val)
- {
- u32 result;
- bool error;
- asm volatile(
- "xor %[error], %[error] \n"
- "safe_atomic_fetch_add_relaxed_ins: \n"
- "lock xadd %[result], %[var] \n"
- "safe_atomic_fetch_add_relaxed_faulted: \n"
- : [error] "=d"(error), [result] "=a"(result), [var] "=m"(*var)
- : [val] "a"(val)
- : "memory");
- if (error)
- return {};
- return result;
- }
- CODE_SECTION(".text.safemem.atomic")
- NEVER_INLINE Optional<u32> safe_atomic_exchange_relaxed(volatile u32* var, u32 val)
- {
- u32 result;
- bool error;
- asm volatile(
- "xor %[error], %[error] \n"
- "safe_atomic_exchange_relaxed_ins: \n"
- "xchg %[val], %[var] \n"
- "safe_atomic_exchange_relaxed_faulted: \n"
- : [error] "=d"(error), "=a"(result), [var] "=m"(*var)
- : [val] "a"(val)
- : "memory");
- if (error)
- return {};
- return result;
- }
- CODE_SECTION(".text.safemem.atomic")
- NEVER_INLINE Optional<u32> safe_atomic_load_relaxed(volatile u32* var)
- {
- u32 result;
- bool error;
- asm volatile(
- "xor %[error], %[error] \n"
- "safe_atomic_load_relaxed_ins: \n"
- "mov (%[var]), %[result] \n"
- "safe_atomic_load_relaxed_faulted: \n"
- : [error] "=d"(error), [result] "=c"(result)
- : [var] "b"(var)
- : "memory");
- if (error)
- return {};
- return result;
- }
- CODE_SECTION(".text.safemem.atomic")
- NEVER_INLINE bool safe_atomic_store_relaxed(volatile u32* var, u32 val)
- {
- bool error;
- asm volatile(
- "xor %[error], %[error] \n"
- "safe_atomic_store_relaxed_ins: \n"
- "xchg %[val], %[var] \n"
- "safe_atomic_store_relaxed_faulted: \n"
- : [error] "=d"(error), [var] "=m"(*var)
- : [val] "r"(val)
- : "memory");
- return !error;
- }
- CODE_SECTION(".text.safemem.atomic")
- NEVER_INLINE Optional<bool> safe_atomic_compare_exchange_relaxed(volatile u32* var, u32& expected, u32 val)
- {
- // NOTE: accessing expected is NOT protected as it should always point
- // to a valid location in kernel memory!
- bool error;
- bool did_exchange;
- asm volatile(
- "xor %[error], %[error] \n"
- "safe_atomic_compare_exchange_relaxed_ins: \n"
- "lock cmpxchg %[val], %[var] \n"
- "safe_atomic_compare_exchange_relaxed_faulted: \n"
- : [error] "=d"(error), "=a"(expected), [var] "=m"(*var), "=@ccz"(did_exchange)
- : "a"(expected), [val] "b"(val)
- : "memory");
- if (error)
- return {};
- return did_exchange;
- }
- bool handle_safe_access_fault(RegisterState& regs, FlatPtr fault_address)
- {
- FlatPtr ip = regs.ip();
- ;
- if (ip >= (FlatPtr)&start_of_safemem_text && ip < (FlatPtr)&end_of_safemem_text) {
- // If we detect that the fault happened in safe_memcpy() safe_strnlen(),
- // or safe_memset() then resume at the appropriate _faulted label
- if (ip == (FlatPtr)&safe_memcpy_ins_1)
- ip = (FlatPtr)&safe_memcpy_1_faulted;
- else if (ip == (FlatPtr)&safe_memcpy_ins_2)
- ip = (FlatPtr)&safe_memcpy_2_faulted;
- else if (ip == (FlatPtr)&safe_strnlen_ins)
- ip = (FlatPtr)&safe_strnlen_faulted;
- else if (ip == (FlatPtr)&safe_memset_ins_1)
- ip = (FlatPtr)&safe_memset_1_faulted;
- else if (ip == (FlatPtr)&safe_memset_ins_2)
- ip = (FlatPtr)&safe_memset_2_faulted;
- else
- return false;
- #if ARCH(I386)
- regs.eip = ip;
- regs.edx = fault_address;
- #else
- regs.rip = ip;
- regs.rdx = fault_address;
- #endif
- return true;
- }
- if (ip >= (FlatPtr)&start_of_safemem_atomic_text && ip < (FlatPtr)&end_of_safemem_atomic_text) {
- // If we detect that a fault happened in one of the atomic safe_
- // functions, resume at the appropriate _faulted label and set
- // the edx/rdx register to 1 to indicate an error
- if (ip == (FlatPtr)&safe_atomic_fetch_add_relaxed_ins)
- ip = (FlatPtr)&safe_atomic_fetch_add_relaxed_faulted;
- else if (ip == (FlatPtr)&safe_atomic_exchange_relaxed_ins)
- ip = (FlatPtr)&safe_atomic_exchange_relaxed_faulted;
- else if (ip == (FlatPtr)&safe_atomic_load_relaxed_ins)
- ip = (FlatPtr)&safe_atomic_load_relaxed_faulted;
- else if (ip == (FlatPtr)&safe_atomic_store_relaxed_ins)
- ip = (FlatPtr)&safe_atomic_store_relaxed_faulted;
- else if (ip == (FlatPtr)&safe_atomic_compare_exchange_relaxed_ins)
- ip = (FlatPtr)&safe_atomic_compare_exchange_relaxed_faulted;
- else
- return false;
- #if ARCH(I386)
- regs.eip = ip;
- regs.edx = 1;
- #else
- regs.rip = ip;
- regs.rdx = 1;
- #endif
- return true;
- }
- return false;
- }
- }
|