SharedInodeVMObject.cpp 2.8 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172
  1. /*
  2. * Copyright (c) 2020, Andreas Kling <kling@serenityos.org>
  3. *
  4. * SPDX-License-Identifier: BSD-2-Clause
  5. */
  6. #include <Kernel/FileSystem/Inode.h>
  7. #include <Kernel/Locking/Spinlock.h>
  8. #include <Kernel/Memory/SharedInodeVMObject.h>
  9. namespace Kernel::Memory {
  10. ErrorOr<NonnullLockRefPtr<SharedInodeVMObject>> SharedInodeVMObject::try_create_with_inode(Inode& inode)
  11. {
  12. if (inode.size() == 0)
  13. return EINVAL;
  14. return try_create_with_inode_and_range(inode, 0, inode.size());
  15. }
  16. ErrorOr<NonnullLockRefPtr<SharedInodeVMObject>> SharedInodeVMObject::try_create_with_inode_and_range(Inode& inode, u64 offset, size_t range_size)
  17. {
  18. // Note: To ensure further allocation of a Region with this VMObject will not complain
  19. // on "smaller" VMObject than the requested Region, we simply take the max size between both values.
  20. auto size = max(inode.size(), (offset + range_size));
  21. VERIFY(size > 0);
  22. if (auto shared_vmobject = inode.shared_vmobject())
  23. return shared_vmobject.release_nonnull();
  24. auto new_physical_pages = TRY(VMObject::try_create_physical_pages(size));
  25. auto dirty_pages = TRY(Bitmap::create(new_physical_pages.size(), false));
  26. auto vmobject = TRY(adopt_nonnull_lock_ref_or_enomem(new (nothrow) SharedInodeVMObject(inode, move(new_physical_pages), move(dirty_pages))));
  27. TRY(vmobject->inode().set_shared_vmobject(*vmobject));
  28. return vmobject;
  29. }
  30. ErrorOr<NonnullLockRefPtr<VMObject>> SharedInodeVMObject::try_clone()
  31. {
  32. auto new_physical_pages = TRY(this->try_clone_physical_pages());
  33. auto dirty_pages = TRY(Bitmap::create(new_physical_pages.size(), false));
  34. return adopt_nonnull_lock_ref_or_enomem<VMObject>(new (nothrow) SharedInodeVMObject(*this, move(new_physical_pages), move(dirty_pages)));
  35. }
  36. SharedInodeVMObject::SharedInodeVMObject(Inode& inode, FixedArray<RefPtr<PhysicalPage>>&& new_physical_pages, Bitmap dirty_pages)
  37. : InodeVMObject(inode, move(new_physical_pages), move(dirty_pages))
  38. {
  39. }
  40. SharedInodeVMObject::SharedInodeVMObject(SharedInodeVMObject const& other, FixedArray<RefPtr<PhysicalPage>>&& new_physical_pages, Bitmap dirty_pages)
  41. : InodeVMObject(other, move(new_physical_pages), move(dirty_pages))
  42. {
  43. }
  44. ErrorOr<void> SharedInodeVMObject::sync(off_t offset_in_pages, size_t pages)
  45. {
  46. SpinlockLocker locker(m_lock);
  47. size_t highest_page_to_flush = min(page_count(), offset_in_pages + pages);
  48. for (size_t page_index = offset_in_pages; page_index < highest_page_to_flush; ++page_index) {
  49. auto& physical_page = m_physical_pages[page_index];
  50. if (!physical_page)
  51. continue;
  52. u8 page_buffer[PAGE_SIZE];
  53. MM.copy_physical_page(*physical_page, page_buffer);
  54. TRY(m_inode->write_bytes(page_index * PAGE_SIZE, PAGE_SIZE, UserOrKernelBuffer::for_kernel_buffer(page_buffer), nullptr));
  55. }
  56. return {};
  57. }
  58. }