diff options
author | Andreas Kling <kling@serenityos.org> | 2021-08-05 18:58:33 +0200 |
---|---|---|
committer | Andreas Kling <kling@serenityos.org> | 2021-08-06 00:37:47 +0200 |
commit | f35108fc315ebd14acc4a8ce628568fced137e8a (patch) | |
tree | 8271f4bbe687eb8c6baa2968afacb403e4af3a9c | |
parent | 27100126c0cba7a7a1196f837dcacefdb059f664 (diff) | |
download | serenity-f35108fc315ebd14acc4a8ce628568fced137e8a.zip |
Kernel: Simplify PageDirectory allocation failure
This patch gets rid of the "valid" bit in PageDirectory since it was
only used to communicate an allocation failure during construction.
We now do all the work in the static factory functions instead of in the
constructor, which allows us to simply return nullptr instead of an
"invalid" PageDirectory.
-rw-r--r-- | Kernel/VM/MemoryManager.cpp | 2 | ||||
-rw-r--r-- | Kernel/VM/PageDirectory.cpp | 99 | ||||
-rw-r--r-- | Kernel/VM/PageDirectory.h | 14 | ||||
-rw-r--r-- | Kernel/VM/Space.cpp | 2 |
4 files changed, 60 insertions, 57 deletions
diff --git a/Kernel/VM/MemoryManager.cpp b/Kernel/VM/MemoryManager.cpp index 5e769fa6eb..239b95ad39 100644 --- a/Kernel/VM/MemoryManager.cpp +++ b/Kernel/VM/MemoryManager.cpp @@ -392,7 +392,7 @@ UNMAP_AFTER_INIT void MemoryManager::initialize_physical_pages() m_used_memory_ranges.append({ UsedMemoryRangeType::PhysicalPages, m_physical_pages_region->lower(), m_physical_pages_region->upper() }); // Create the bare page directory. This is not a fully constructed page directory and merely contains the allocators! - m_kernel_page_directory = PageDirectory::create_kernel_page_directory(); + m_kernel_page_directory = PageDirectory::must_create_kernel_page_directory(); // Allocate a virtual address range for our array auto range = m_kernel_page_directory->range_allocator().allocate_anywhere(physical_page_array_pages * PAGE_SIZE); diff --git a/Kernel/VM/PageDirectory.cpp b/Kernel/VM/PageDirectory.cpp index f599e7717b..b49c6f3987 100644 --- a/Kernel/VM/PageDirectory.cpp +++ b/Kernel/VM/PageDirectory.cpp @@ -31,77 +31,73 @@ RefPtr<PageDirectory> PageDirectory::find_by_cr3(FlatPtr cr3) return cr3_map().get(cr3).value_or({}); } -UNMAP_AFTER_INIT PageDirectory::PageDirectory() +UNMAP_AFTER_INIT NonnullRefPtr<PageDirectory> PageDirectory::must_create_kernel_page_directory() { + auto directory = adopt_ref_if_nonnull(new (nothrow) PageDirectory).release_nonnull(); + // make sure this starts in a new page directory to make MemoryManager::initialize_physical_pages() happy FlatPtr start_of_range = ((FlatPtr)end_of_kernel_image & ~(FlatPtr)0x1fffff) + 0x200000; - m_range_allocator.initialize_with_range(VirtualAddress(start_of_range), KERNEL_PD_END - start_of_range); - m_identity_range_allocator.initialize_with_range(VirtualAddress(FlatPtr(0x00000000)), 0x00200000); -} + directory->m_range_allocator.initialize_with_range(VirtualAddress(start_of_range), KERNEL_PD_END - start_of_range); + directory->m_identity_range_allocator.initialize_with_range(VirtualAddress(FlatPtr(0x00000000)), 0x00200000); -UNMAP_AFTER_INIT void PageDirectory::allocate_kernel_directory() -{ - // Adopt the page tables already set up by boot.S -#if ARCH(X86_64) - dmesgln("MM: boot_pml4t @ {}", boot_pml4t); - m_pml4t = PhysicalPage::create(boot_pml4t, MayReturnToFreeList::No); -#endif - dmesgln("MM: boot_pdpt @ {}", boot_pdpt); - dmesgln("MM: boot_pd0 @ {}", boot_pd0); - dmesgln("MM: boot_pd_kernel @ {}", boot_pd_kernel); - m_directory_table = PhysicalPage::create(boot_pdpt, MayReturnToFreeList::No); - m_directory_pages[0] = PhysicalPage::create(boot_pd0, MayReturnToFreeList::No); - m_directory_pages[(kernel_mapping_base >> 30) & 0x1ff] = PhysicalPage::create(boot_pd_kernel, MayReturnToFreeList::No); + return directory; } -PageDirectory::PageDirectory(const RangeAllocator* parent_range_allocator) +RefPtr<PageDirectory> PageDirectory::try_create_for_userspace(RangeAllocator const* parent_range_allocator) { constexpr FlatPtr userspace_range_base = 0x00800000; - FlatPtr userspace_range_ceiling = USER_RANGE_CEILING; + FlatPtr const userspace_range_ceiling = USER_RANGE_CEILING; + + auto directory = adopt_ref_if_nonnull(new (nothrow) PageDirectory); + if (!directory) + return {}; - ScopedSpinLock lock(s_mm_lock); if (parent_range_allocator) { - m_range_allocator.initialize_from_parent(*parent_range_allocator); + directory->m_range_allocator.initialize_from_parent(*parent_range_allocator); } else { size_t random_offset = (get_fast_random<u8>() % 32 * MiB) & PAGE_MASK; u32 base = userspace_range_base + random_offset; - m_range_allocator.initialize_with_range(VirtualAddress(base), userspace_range_ceiling - base); + directory->m_range_allocator.initialize_with_range(VirtualAddress(base), userspace_range_ceiling - base); } - // Set up a userspace page directory + // NOTE: Take the MM lock since we need it for quickmap. + ScopedSpinLock lock(s_mm_lock); + #if ARCH(X86_64) - m_pml4t = MM.allocate_user_physical_page(); - if (!m_pml4t) - return; + directory->m_pml4t = MM.allocate_user_physical_page(); + if (!directory->m_pml4t) + return {}; #endif - m_directory_table = MM.allocate_user_physical_page(); - if (!m_directory_table) - return; + + directory->m_directory_table = MM.allocate_user_physical_page(); + if (!directory->m_directory_table) + return {}; auto kernel_pd_index = (kernel_mapping_base >> 30) & 0x1ffu; for (size_t i = 0; i < kernel_pd_index; i++) { - m_directory_pages[i] = MM.allocate_user_physical_page(); - if (!m_directory_pages[i]) - return; + directory->m_directory_pages[i] = MM.allocate_user_physical_page(); + if (!directory->m_directory_pages[i]) + return {}; } + // Share the top 1 GiB of kernel-only mappings (>=kernel_mapping_base) - m_directory_pages[kernel_pd_index] = MM.kernel_page_directory().m_directory_pages[kernel_pd_index]; + directory->m_directory_pages[kernel_pd_index] = MM.kernel_page_directory().m_directory_pages[kernel_pd_index]; #if ARCH(X86_64) { - auto& table = *(PageDirectoryPointerTable*)MM.quickmap_page(*m_pml4t); - table.raw[0] = (FlatPtr)m_directory_table->paddr().as_ptr() | 7; + auto& table = *(PageDirectoryPointerTable*)MM.quickmap_page(*directory->m_pml4t); + table.raw[0] = (FlatPtr)directory->m_directory_table->paddr().as_ptr() | 7; MM.unquickmap_page(); } #endif { - auto& table = *(PageDirectoryPointerTable*)MM.quickmap_page(*m_directory_table); + auto& table = *(PageDirectoryPointerTable*)MM.quickmap_page(*directory->m_directory_table); for (size_t i = 0; i < sizeof(m_directory_pages) / sizeof(m_directory_pages[0]); i++) { - if (m_directory_pages[i]) { + if (directory->m_directory_pages[i]) { #if ARCH(I386) - table.raw[i] = (FlatPtr)m_directory_pages[i]->paddr().as_ptr() | 1; + table.raw[i] = (FlatPtr)directory->m_directory_pages[i]->paddr().as_ptr() | 1; #else - table.raw[i] = (FlatPtr)m_directory_pages[i]->paddr().as_ptr() | 7; + table.raw[i] = (FlatPtr)directory->m_directory_pages[i]->paddr().as_ptr() | 7; #endif } } @@ -136,13 +132,30 @@ PageDirectory::PageDirectory(const RangeAllocator* parent_range_allocator) PageDirectoryEntry buffer; auto* kernel_pd = MM.quickmap_pd(MM.kernel_page_directory(), 0); memcpy(&buffer, kernel_pd, sizeof(PageDirectoryEntry)); - auto* new_pd = MM.quickmap_pd(*this, 0); + auto* new_pd = MM.quickmap_pd(*directory, 0); memcpy(new_pd, &buffer, sizeof(PageDirectoryEntry)); - // If we got here, we successfully created it. Set m_space now - m_valid = true; + cr3_map().set(directory->cr3(), directory.ptr()); + return directory; +} + +PageDirectory::PageDirectory() +{ +} - cr3_map().set(cr3(), this); +UNMAP_AFTER_INIT void PageDirectory::allocate_kernel_directory() +{ + // Adopt the page tables already set up by boot.S +#if ARCH(X86_64) + dmesgln("MM: boot_pml4t @ {}", boot_pml4t); + m_pml4t = PhysicalPage::create(boot_pml4t, MayReturnToFreeList::No); +#endif + dmesgln("MM: boot_pdpt @ {}", boot_pdpt); + dmesgln("MM: boot_pd0 @ {}", boot_pd0); + dmesgln("MM: boot_pd_kernel @ {}", boot_pd_kernel); + m_directory_table = PhysicalPage::create(boot_pdpt, MayReturnToFreeList::No); + m_directory_pages[0] = PhysicalPage::create(boot_pd0, MayReturnToFreeList::No); + m_directory_pages[(kernel_mapping_base >> 30) & 0x1ff] = PhysicalPage::create(boot_pd_kernel, MayReturnToFreeList::No); } PageDirectory::~PageDirectory() diff --git a/Kernel/VM/PageDirectory.h b/Kernel/VM/PageDirectory.h index d66c19dd91..28832056ea 100644 --- a/Kernel/VM/PageDirectory.h +++ b/Kernel/VM/PageDirectory.h @@ -19,14 +19,8 @@ class PageDirectory : public RefCounted<PageDirectory> { friend class MemoryManager; public: - static RefPtr<PageDirectory> create_for_userspace(const RangeAllocator* parent_range_allocator = nullptr) - { - auto page_directory = adopt_ref(*new PageDirectory(parent_range_allocator)); - if (!page_directory->is_valid()) - return {}; - return page_directory; - } - static NonnullRefPtr<PageDirectory> create_kernel_page_directory() { return adopt_ref(*new PageDirectory); } + static RefPtr<PageDirectory> try_create_for_userspace(RangeAllocator const* parent_range_allocator = nullptr); + static NonnullRefPtr<PageDirectory> must_create_kernel_page_directory(); static RefPtr<PageDirectory> find_by_cr3(FlatPtr); ~PageDirectory(); @@ -47,8 +41,6 @@ public: RangeAllocator& identity_range_allocator() { return m_identity_range_allocator; } - bool is_valid() const { return m_valid; } - Space* space() { return m_space; } const Space* space() const { return m_space; } @@ -57,7 +49,6 @@ public: RecursiveSpinLock& get_lock() { return m_lock; } private: - explicit PageDirectory(const RangeAllocator* parent_range_allocator); PageDirectory(); Space* m_space { nullptr }; @@ -74,7 +65,6 @@ private: #endif HashMap<FlatPtr, RefPtr<PhysicalPage>> m_page_tables; RecursiveSpinLock m_lock; - bool m_valid { false }; }; } diff --git a/Kernel/VM/Space.cpp b/Kernel/VM/Space.cpp index fb89991906..9196bfe78b 100644 --- a/Kernel/VM/Space.cpp +++ b/Kernel/VM/Space.cpp @@ -17,7 +17,7 @@ namespace Kernel { OwnPtr<Space> Space::try_create(Process& process, Space const* parent) { - auto page_directory = PageDirectory::create_for_userspace(parent ? &parent->page_directory().range_allocator() : nullptr); + auto page_directory = PageDirectory::try_create_for_userspace(parent ? &parent->page_directory().range_allocator() : nullptr); if (!page_directory) return {}; auto space = adopt_own_if_nonnull(new (nothrow) Space(process, page_directory.release_nonnull())); |