12345678910111213141516171819202122232425262728293031323334353637383940414243444546474849505152535455565758596061626364656667686970717273747576777879808182838485868788899091929394 |
- /*
- * Copyright (c) 2021-2023, Andreas Kling <kling@serenityos.org>
- *
- * SPDX-License-Identifier: BSD-2-Clause
- */
- #include <AK/Platform.h>
- #include <AK/Random.h>
- #include <AK/Vector.h>
- #include <LibJS/Heap/BlockAllocator.h>
- #include <LibJS/Heap/HeapBlock.h>
- #include <sys/mman.h>
- #ifdef HAS_ADDRESS_SANITIZER
- # include <sanitizer/asan_interface.h>
- #endif
- // FIXME: Implement MADV_FREE and/or MADV_DONTNEED on SerenityOS.
- #if defined(AK_OS_SERENITY) || defined(AK_OS_GNU_HURD) || (!defined(MADV_FREE) && !defined(MADV_DONTNEED))
- # define USE_FALLBACK_BLOCK_DEALLOCATION
- #endif
- namespace JS {
- // NOTE: If this changes, we need to update the mmap() code to ensure correct alignment.
- static_assert(HeapBlock::block_size == 4096);
- BlockAllocator::~BlockAllocator()
- {
- for (auto* block : m_blocks) {
- ASAN_UNPOISON_MEMORY_REGION(block, HeapBlock::block_size);
- if (munmap(block, HeapBlock::block_size) < 0) {
- perror("munmap");
- VERIFY_NOT_REACHED();
- }
- }
- }
- void* BlockAllocator::allocate_block([[maybe_unused]] char const* name)
- {
- if (!m_blocks.is_empty()) {
- // To reduce predictability, take a random block from the cache.
- size_t random_index = get_random_uniform(m_blocks.size());
- auto* block = m_blocks.unstable_take(random_index);
- ASAN_UNPOISON_MEMORY_REGION(block, HeapBlock::block_size);
- #ifdef AK_OS_SERENITY
- if (set_mmap_name(block, HeapBlock::block_size, name) < 0) {
- perror("set_mmap_name");
- VERIFY_NOT_REACHED();
- }
- #endif
- return block;
- }
- #ifdef AK_OS_SERENITY
- auto* block = (HeapBlock*)serenity_mmap(nullptr, HeapBlock::block_size, PROT_READ | PROT_WRITE, MAP_ANONYMOUS | MAP_RANDOMIZED | MAP_PRIVATE, 0, 0, HeapBlock::block_size, name);
- #else
- auto* block = (HeapBlock*)mmap(nullptr, HeapBlock::block_size, PROT_READ | PROT_WRITE, MAP_ANONYMOUS | MAP_PRIVATE, 0, 0);
- #endif
- VERIFY(block != MAP_FAILED);
- return block;
- }
- void BlockAllocator::deallocate_block(void* block)
- {
- VERIFY(block);
- #if defined(USE_FALLBACK_BLOCK_DEALLOCATION)
- // If we can't use any of the nicer techniques, unmap and remap the block to return the physical pages while keeping the VM.
- if (munmap(block, HeapBlock::block_size) < 0) {
- perror("munmap");
- VERIFY_NOT_REACHED();
- }
- if (mmap(block, HeapBlock::block_size, PROT_READ | PROT_WRITE, MAP_ANONYMOUS | MAP_PRIVATE | MAP_FIXED, 0, 0) != block) {
- perror("mmap");
- VERIFY_NOT_REACHED();
- }
- #elif defined(MADV_FREE)
- if (madvise(block, HeapBlock::block_size, MADV_FREE) < 0) {
- perror("madvise(MADV_FREE)");
- VERIFY_NOT_REACHED();
- }
- #elif defined(MADV_DONTNEED)
- if (madvise(block, HeapBlock::block_size, MADV_DONTNEED) < 0) {
- perror("madvise(MADV_DONTNEED)");
- VERIFY_NOT_REACHED();
- }
- #endif
- ASAN_POISON_MEMORY_REGION(block, HeapBlock::block_size);
- m_blocks.append(block);
- }
- }
|