#pragma once #include #include #include #include #include #include class Inode; class VMObject; enum class PageFaultResponse { ShouldCrash, Continue, }; class Region final : public InlineLinkedListNode { friend class MemoryManager; MAKE_SLAB_ALLOCATED(Region) public: enum Access { Read = 1, Write = 2, Execute = 4, }; static NonnullOwnPtr create_user_accessible(const Range&, const StringView& name, u8 access); static NonnullOwnPtr create_user_accessible(const Range&, NonnullRefPtr, size_t offset_in_vmobject, const StringView& name, u8 access); static NonnullOwnPtr create_user_accessible(const Range&, NonnullRefPtr, const StringView& name, u8 access); static NonnullOwnPtr create_kernel_only(const Range&, const StringView& name, u8 access); ~Region(); const Range& range() const { return m_range; } VirtualAddress vaddr() const { return m_range.base(); } size_t size() const { return m_range.size(); } bool is_readable() const { return m_access & Access::Read; } bool is_writable() const { return m_access & Access::Write; } bool is_executable() const { return m_access & Access::Execute; } const String& name() const { return m_name; } unsigned access() const { return m_access; } void set_name(const String& name) { m_name = name; } const VMObject& vmobject() const { return *m_vmobject; } VMObject& vmobject() { return *m_vmobject; } bool is_shared() const { return m_shared; } void set_shared(bool shared) { m_shared = shared; } bool is_stack() const { return m_stack; } void set_stack(bool stack) { m_stack = stack; } bool is_user_accessible() const { return m_user_accessible; } PageFaultResponse handle_fault(const PageFault&); NonnullOwnPtr clone(); bool contains(VirtualAddress vaddr) const { return m_range.contains(vaddr); } bool contains(const Range& range) const { return m_range.contains(range); } unsigned page_index_from_address(VirtualAddress vaddr) const { return (vaddr - m_range.base()).get() / PAGE_SIZE; } size_t first_page_index() const { return m_offset_in_vmo / PAGE_SIZE; } size_t last_page_index() const { return (first_page_index() + page_count()) - 1; } size_t page_count() const { return size() / PAGE_SIZE; } size_t offset_in_vmobject() const { return m_offset_in_vmo; } int commit(); size_t amount_resident() const; size_t amount_shared() const; bool should_cow(size_t page_index) const; void set_should_cow(size_t page_index, bool); void set_writable(bool b) { if (b) m_access |= Access::Read; else m_access &= ~Access::Write; } void map(PageDirectory&); enum class ShouldDeallocateVirtualMemoryRange { No, Yes, }; void unmap(ShouldDeallocateVirtualMemoryRange = ShouldDeallocateVirtualMemoryRange::Yes); void remap(); void remap_page(size_t index); // For InlineLinkedListNode Region* m_next { nullptr }; Region* m_prev { nullptr }; // NOTE: These are public so we can make<> them. Region(const Range&, const String&, u8 access); Region(const Range&, NonnullRefPtr, size_t offset_in_vmo, const String&, u8 access); Region(const Range&, NonnullRefPtr, const String&, u8 access); private: Bitmap& ensure_cow_map() const; PageFaultResponse handle_cow_fault(size_t page_index); PageFaultResponse handle_inode_fault(size_t page_index); PageFaultResponse handle_zero_fault(size_t page_index); RefPtr m_page_directory; Range m_range; size_t m_offset_in_vmo { 0 }; NonnullRefPtr m_vmobject; String m_name; u8 m_access { 0 }; bool m_shared { false }; bool m_user_accessible { false }; bool m_stack { false }; mutable OwnPtr m_cow_map; };