/* * Copyright (c) 2020, The SerenityOS developers. * All rights reserved. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions are met: * * 1. Redistributions of source code must retain the above copyright notice, this * list of conditions and the following disclaimer. * * 2. Redistributions in binary form must reproduce the above copyright notice, * this list of conditions and the following disclaimer in the documentation * and/or other materials provided with the distribution. * * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE * DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR * SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER * CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, * OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. */ #pragma once #include #include #include #include #include namespace Kernel { template class Heap { AK_MAKE_NONCOPYABLE(Heap); struct AllocationHeader { size_t allocation_size_in_chunks; u8 data[0]; }; static size_t calculate_chunks(size_t memory_size) { return (sizeof(u8) * memory_size) / (sizeof(u8) * CHUNK_SIZE + 1); } public: Heap(u8* memory, size_t memory_size) : m_total_chunks(calculate_chunks(memory_size)) , m_chunks(memory) , m_bitmap(Bitmap::wrap(memory + m_total_chunks * CHUNK_SIZE, m_total_chunks)) { // To keep the alignment of the memory passed in, place the bitmap // at the end of the memory block. ASSERT(m_total_chunks * CHUNK_SIZE + (m_total_chunks + 7) / 8 <= memory_size); } ~Heap() { } static size_t calculate_memory_for_bytes(size_t bytes) { size_t needed_chunks = (sizeof(AllocationHeader) + bytes + CHUNK_SIZE - 1) / CHUNK_SIZE; return needed_chunks * CHUNK_SIZE + (needed_chunks + 7) / 8; } void* allocate(size_t size) { // We need space for the AllocationHeader at the head of the block. size_t real_size = size + sizeof(AllocationHeader); size_t chunks_needed = (real_size + CHUNK_SIZE - 1) / CHUNK_SIZE; if (chunks_needed > free_chunks()) return nullptr; Optional first_chunk; // Choose the right politic for allocation. constexpr u32 best_fit_threshold = 128; if (chunks_needed < best_fit_threshold) { first_chunk = m_bitmap.find_first_fit(chunks_needed); } else { first_chunk = m_bitmap.find_best_fit(chunks_needed); } if (!first_chunk.has_value()) return nullptr; auto* a = (AllocationHeader*)(m_chunks + (first_chunk.value() * CHUNK_SIZE)); u8* ptr = a->data; a->allocation_size_in_chunks = chunks_needed; m_bitmap.set_range(first_chunk.value(), chunks_needed, true); m_allocated_chunks += chunks_needed; if constexpr (HEAP_SCRUB_BYTE_ALLOC != 0) { __builtin_memset(ptr, HEAP_SCRUB_BYTE_ALLOC, (chunks_needed * CHUNK_SIZE) - sizeof(AllocationHeader)); } return ptr; } void deallocate(void* ptr) { if (!ptr) return; auto* a = (AllocationHeader*)((((u8*)ptr) - sizeof(AllocationHeader))); ASSERT((u8*)a >= m_chunks && (u8*)ptr < m_chunks + m_total_chunks * CHUNK_SIZE); ASSERT((u8*)a + a->allocation_size_in_chunks * CHUNK_SIZE <= m_chunks + m_total_chunks * CHUNK_SIZE); FlatPtr start = ((FlatPtr)a - (FlatPtr)m_chunks) / CHUNK_SIZE; m_bitmap.set_range(start, a->allocation_size_in_chunks, false); ASSERT(m_allocated_chunks >= a->allocation_size_in_chunks); m_allocated_chunks -= a->allocation_size_in_chunks; if constexpr (HEAP_SCRUB_BYTE_FREE != 0) { __builtin_memset(a, HEAP_SCRUB_BYTE_FREE, a->allocation_size_in_chunks * CHUNK_SIZE); } } template void* reallocate(void* ptr, size_t new_size, MainHeap& h) { if (!ptr) return h.allocate(new_size); auto* a = (AllocationHeader*)((((u8*)ptr) - sizeof(AllocationHeader))); ASSERT((u8*)a >= m_chunks && (u8*)ptr < m_chunks + m_total_chunks * CHUNK_SIZE); ASSERT((u8*)a + a->allocation_size_in_chunks * CHUNK_SIZE <= m_chunks + m_total_chunks * CHUNK_SIZE); size_t old_size = a->allocation_size_in_chunks * CHUNK_SIZE; if (old_size == new_size) return ptr; auto* new_ptr = h.allocate(new_size); if (new_ptr) __builtin_memcpy(new_ptr, ptr, min(old_size, new_size)); deallocate(ptr); return new_ptr; } void* reallocate(void* ptr, size_t new_size) { return reallocate(ptr, new_size, *this); } bool contains(const void* ptr) const { const auto* a = (const AllocationHeader*)((((const u8*)ptr) - sizeof(AllocationHeader))); if ((const u8*)a < m_chunks) return false; if ((const u8*)ptr >= m_chunks + m_total_chunks * CHUNK_SIZE) return false; return true; } u8* memory() const { return m_chunks; } size_t total_chunks() const { return m_total_chunks; } size_t total_bytes() const { return m_total_chunks * CHUNK_SIZE; } size_t free_chunks() const { return m_total_chunks - m_allocated_chunks; }; size_t free_bytes() const { return free_chunks() * CHUNK_SIZE; } size_t allocated_chunks() const { return m_allocated_chunks; } size_t allocated_bytes() const { return m_allocated_chunks * CHUNK_SIZE; } private: size_t m_total_chunks { 0 }; size_t m_allocated_chunks { 0 }; u8* m_chunks { nullptr }; Bitmap m_bitmap; }; template struct ExpandableHeapTraits { static bool add_memory(ExpandHeap& expand, size_t allocation_request) { return expand.add_memory(allocation_request); } static bool remove_memory(ExpandHeap& expand, void* memory) { return expand.remove_memory(memory); } }; struct DefaultExpandHeap { bool add_memory(size_t) { // Requires explicit implementation return false; } bool remove_memory(void*) { return false; } }; template class ExpandableHeap { AK_MAKE_NONCOPYABLE(ExpandableHeap); AK_MAKE_NONMOVABLE(ExpandableHeap); public: typedef ExpandHeap ExpandHeapType; typedef Heap HeapType; struct SubHeap { HeapType heap; SubHeap* next { nullptr }; template SubHeap(Args&&... args) : heap(forward(args)...) { } }; ExpandableHeap(u8* memory, size_t memory_size, const ExpandHeapType& expand = ExpandHeapType()) : m_heaps(memory, memory_size) , m_expand(expand) { } ~ExpandableHeap() { // We don't own the main heap, only remove memory that we added previously SubHeap* next; for (auto* heap = m_heaps.next; heap; heap = next) { next = heap->next; heap->~SubHeap(); ExpandableHeapTraits::remove_memory(m_expand, (void*)heap); } } static size_t calculate_memory_for_bytes(size_t bytes) { return sizeof(SubHeap) + HeapType::calculate_memory_for_bytes(bytes); } bool expand_memory(size_t size) { if (m_expanding) return false; // Allocating more memory itself may trigger allocations and deallocations // on this heap. We need to prevent recursive expansion. We also disable // removing memory while trying to expand the heap. TemporaryChange change(m_expanding, true); return ExpandableHeapTraits::add_memory(m_expand, size); } void* allocate(size_t size) { int attempt = 0; do { for (auto* subheap = &m_heaps; subheap; subheap = subheap->next) { if (void* ptr = subheap->heap.allocate(size)) return ptr; } // We need to loop because we won't know how much memory was added. // Even though we make a best guess how much memory needs to be added, // it doesn't guarantee that enough will be available after adding it. // This is especially true for the kmalloc heap, where adding memory // requires several other objects to be allocated just to be able to // expand the heap. // To avoid an infinite expansion loop, limit to two attempts if (attempt++ >= 2) break; } while (expand_memory(size)); return nullptr; } void deallocate(void* ptr) { if (!ptr) return; for (auto* subheap = &m_heaps; subheap; subheap = subheap->next) { if (subheap->heap.contains(ptr)) { subheap->heap.deallocate(ptr); if (subheap->heap.allocated_chunks() == 0 && subheap != &m_heaps && !m_expanding) { // Since remove_memory may free subheap, we need to save the // next pointer before calling it auto* next_subheap = subheap->next; if (ExpandableHeapTraits::remove_memory(m_expand, subheap)) { auto* subheap2 = m_heaps.next; auto** subheap_link = &m_heaps.next; while (subheap2 != subheap) { subheap_link = &subheap2->next; subheap2 = subheap2->next; } *subheap_link = next_subheap; } } return; } } ASSERT_NOT_REACHED(); } void* reallocate(void* ptr, size_t new_size) { if (!ptr) return allocate(new_size); for (auto* subheap = &m_heaps; subheap; subheap = subheap->next) { if (subheap->heap.contains(ptr)) return subheap->heap.reallocate(ptr, new_size, *this); } ASSERT_NOT_REACHED(); } HeapType& add_subheap(void* memory, size_t memory_size) { ASSERT(memory_size > sizeof(SubHeap)); // Place the SubHeap structure at the beginning of the new memory block memory_size -= sizeof(SubHeap); SubHeap* new_heap = (SubHeap*)memory; new (new_heap) SubHeap((u8*)(new_heap + 1), memory_size); // Add the subheap to the list (but leave the main heap where it is) SubHeap* next_heap = m_heaps.next; SubHeap** next_heap_link = &m_heaps.next; while (next_heap) { if (new_heap->heap.memory() < next_heap->heap.memory()) break; next_heap_link = &next_heap->next; next_heap = next_heap->next; } new_heap->next = *next_heap_link; *next_heap_link = new_heap; return new_heap->heap; } bool contains(const void* ptr) const { for (auto* subheap = &m_heaps; subheap; subheap = subheap->next) { if (subheap->heap.contains(ptr)) return true; } return false; } size_t total_chunks() const { size_t total = 0; for (auto* subheap = &m_heaps; subheap; subheap = subheap->next) total += subheap->heap.total_chunks(); return total; } size_t total_bytes() const { return total_chunks() * CHUNK_SIZE; } size_t free_chunks() const { size_t total = 0; for (auto* subheap = &m_heaps; subheap; subheap = subheap->next) total += subheap->heap.free_chunks(); return total; } size_t free_bytes() const { return free_chunks() * CHUNK_SIZE; } size_t allocated_chunks() const { size_t total = 0; for (auto* subheap = &m_heaps; subheap; subheap = subheap->next) total += subheap->heap.allocated_chunks(); return total; } size_t allocated_bytes() const { return allocated_chunks() * CHUNK_SIZE; } private: SubHeap m_heaps; ExpandHeap m_expand; bool m_expanding { false }; }; }