/* -*- mode: C++; c-basic-offset: 2; indent-tabs-mode: nil -*- */ /* * Main authors: * Christian Schulte * * Contributing authors: * Guido Tack * * Bugfixes provided by: * Zandra Norman * * Copyright: * Christian Schulte, 2002 * Guido Tack, 2004 * * This file is part of Gecode, the generic constraint * development environment: * http://www.gecode.org * * Permission is hereby granted, free of charge, to any person obtaining * a copy of this software and associated documentation files (the * "Software"), to deal in the Software without restriction, including * without limitation the rights to use, copy, modify, merge, publish, * distribute, sublicense, and/or sell copies of the Software, and to * permit persons to whom the Software is furnished to do so, subject to * the following conditions: * * The above copyright notice and this permission notice shall be * included in all copies or substantial portions of the Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE * LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION * OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION * WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. * */ #include namespace Gecode { namespace Kernel { /// Memory chunk with size information class MemoryChunk { public: /// Next chunk MemoryChunk* next; /// Size of chunk size_t size; }; /// Memory chunk allocated from heap with proper alignment class HeapChunk : public MemoryChunk { public: /// Start of memory area inside chunk double area[1]; }; /// Shared object for several memory areas class SharedMemory { private: /// The components for shared heap memory struct { /// How many heap chunks are available for caching unsigned int n_hc; /// A list of cached heap chunks HeapChunk* hc; } heap; /// A mutex for access GECODE_KERNEL_EXPORT static Support::Mutex& m(void); public: /// Initialize SharedMemory(void); /// Destructor ~SharedMemory(void); /// \name Heap management //@ /// Return heap chunk, preferable of size \a s, but at least of size \a l HeapChunk* alloc(size_t s, size_t l); /// Free heap chunk (or cache for later) void free(HeapChunk* hc); //@} }; }} namespace Gecode { /** * \brief Base-class for freelist-managed objects * * Freelist-managed object must inherit from this class. The size * of objects of subclasses is defined by the parameters in * Gecode::MemoryConfig. * \ingroup FuncMemSpace */ class FreeList { protected: /// Pointer to next freelist object FreeList* _next; public: /// Use uninitialized FreeList(void); /// Initialize with next freelist object \a n FreeList(FreeList* n); /// Return next freelist object FreeList* next(void) const; /// Return pointer to next link in freelist object FreeList** nextRef(void); /// Set next freelist object to \a n void next(FreeList* n); }; } namespace Gecode { namespace Kernel { /// Manage memory for space class MemoryManager { public: /// Constructor initialization MemoryManager(SharedMemory& sm); /// Constructor during cloning \a mm with shared memory \a sm and for a memory area for subscriptions of size \a s_sub MemoryManager(SharedMemory& sm, MemoryManager& mm, size_t s_sub); /// Release all allocated heap chunks void release(SharedMemory& sm); private: size_t cur_hcsz; ///< Current heap chunk size HeapChunk* cur_hc; ///< Current heap chunk size_t requested; ///< Total amount of heap memory requested char* start; ///< Start of current heap area used for allocation size_t lsz; ///< Size left for allocation /// Refill current heap area (outlined) issued by request of size \a s GECODE_KERNEL_EXPORT void alloc_refill(SharedMemory& sm, size_t s); /// Do the real work for refilling void alloc_fill(SharedMemory& sm, size_t s, bool first); public: /// Allocate memory of size \a s void* alloc(SharedMemory& sm, size_t s); /// Get the memory area for subscriptions void* subscriptions(void) const; private: /// Start of free lists FreeList* fl[MemoryConfig::fl_size_max-MemoryConfig::fl_size_min+1]; /// Refill free list template void fl_refill(SharedMemory& sm); /// Translate size to index in free list static size_t sz2i(size_t); /// Translate index in free list to size static size_t i2sz(size_t); public: /// Allocate free list element of size \a s template void* fl_alloc(SharedMemory& sm); /// Release all free list elements of size \a s between f and l (inclusive) template void fl_dispose(FreeList* f, FreeList* l); private: /// Slack memory chunks MemoryChunk* slack; public: /// Store for reusal, if of sufficient size for free list void reuse(void* p, size_t s); }; }} namespace Gecode { namespace Kernel { /* * Shared memory area * */ forceinline SharedMemory::SharedMemory(void) { heap.n_hc = 0; heap.hc = nullptr; } forceinline SharedMemory::~SharedMemory(void) { while (heap.hc != nullptr) { HeapChunk* hc = heap.hc; heap.hc = static_cast(hc->next); Gecode::heap.rfree(hc); } } forceinline HeapChunk* SharedMemory::alloc(size_t s, size_t l) { // To protect from exceptions from heap.ralloc() Support::Lock guard(m()); while ((heap.hc != nullptr) && (heap.hc->size < l)) { heap.n_hc--; HeapChunk* hc = heap.hc; heap.hc = static_cast(hc->next); Gecode::heap.rfree(hc); } HeapChunk* hc; if (heap.hc == nullptr) { assert(heap.n_hc == 0); hc = static_cast(Gecode::heap.ralloc(s)); hc->size = s; } else { heap.n_hc--; hc = heap.hc; heap.hc = static_cast(hc->next); } return hc; } forceinline void SharedMemory::free(HeapChunk* hc) { Support::Lock guard(m()); if (heap.n_hc == MemoryConfig::n_hc_cache) { Gecode::heap.rfree(hc); } else { heap.n_hc++; hc->next = heap.hc; heap.hc = hc; } } }} namespace Gecode { /* * Freelists * */ forceinline FreeList::FreeList(void) {} forceinline FreeList::FreeList(FreeList* n) : _next(n) {} forceinline FreeList* FreeList::next(void) const { return _next; } forceinline FreeList** FreeList::nextRef(void) { return &_next; } forceinline void FreeList::next(FreeList* n) { _next = n; } } namespace Gecode { namespace Kernel { /* * The active memory manager * */ forceinline size_t MemoryManager::sz2i(size_t s) { assert(s >= (MemoryConfig::fl_size_min << MemoryConfig::fl_unit_size)); assert(s <= (MemoryConfig::fl_size_max << MemoryConfig::fl_unit_size)); return (s >> MemoryConfig::fl_unit_size) - MemoryConfig::fl_size_min; } forceinline size_t MemoryManager::i2sz(size_t i) { return (i + MemoryConfig::fl_size_min) << MemoryConfig::fl_unit_size; } forceinline void* MemoryManager::alloc(SharedMemory& sm, size_t sz) { assert(sz > 0); // Perform alignment MemoryConfig::align(sz); // Check whether sufficient memory left if (sz > lsz) alloc_refill(sm,sz); lsz -= sz; return start + lsz; } forceinline void* MemoryManager::subscriptions(void) const { return &cur_hc->area[0]; } forceinline void MemoryManager::alloc_fill(SharedMemory& sm, size_t sz, bool first) { // Adjust current heap chunk size if (((requested > MemoryConfig::hcsz_inc_ratio*cur_hcsz) || (sz > cur_hcsz)) && (cur_hcsz < MemoryConfig::hcsz_max) && !first) { cur_hcsz <<= 1; } // Increment the size that it caters for the initial overhead size_t overhead = sizeof(HeapChunk) - sizeof(double); sz += overhead; // Round size to next multiple of current heap chunk size size_t allocate = ((sz > cur_hcsz) ? (static_cast(sz / cur_hcsz) + 1) * cur_hcsz : cur_hcsz); // Request a chunk of preferably size allocate, but at least size sz HeapChunk* hc = sm.alloc(allocate,sz); start = ptr_cast(&hc->area[0]); lsz = hc->size - overhead; // Link heap chunk, where the first heap chunk is kept in place if (first) { requested = hc->size; hc->next = nullptr; cur_hc = hc; } else { requested += hc->size; hc->next = cur_hc->next; cur_hc->next = hc; } #ifdef GECODE_MEMORY_CHECK for (char* c = start; c < (start+lsz); c++) *c = 0; #endif } forceinline MemoryManager::MemoryManager(SharedMemory& sm) : cur_hcsz(MemoryConfig::hcsz_min), requested(0), slack(nullptr) { alloc_fill(sm,cur_hcsz,true); for (size_t i = 0; i MemoryConfig::hcsz_min) && (s_sub*2 < cur_hcsz)) cur_hcsz >>= 1; alloc_fill(sm,cur_hcsz+s_sub,true); // Skip the memory area at the beginning for subscriptions lsz -= s_sub; start += s_sub; for (size_t i = 0; i(hc->next); sm.free(t); } while (hc != nullptr); } /* * Slack memory management * */ forceinline void MemoryManager::reuse(void* p, size_t s) { #ifdef GECODE_MEMORY_CHECK { char* c = static_cast(p); char* e = c + s; while (c < e) { *c = 0; c++; } } #endif size_t aligned_s = s; void* aligned_p = std::align(std::min(static_cast(GECODE_MEMORY_ALIGNMENT), s), MemoryConfig::fl_size_min< (MemoryConfig::fl_size_max<(aligned_p); rc->next = slack; rc->size = aligned_s; slack = rc; } else { size_t i = sz2i(aligned_s); FreeList* f = static_cast(aligned_p); f->next(fl[i]); fl[i]=f; } } /* * Freelist management * */ template forceinline void* MemoryManager::fl_alloc(SharedMemory& sm) { size_t i = sz2i(s); FreeList* f = fl[i]; if (f == nullptr) { fl_refill(sm); f = fl[i]; } FreeList* n = f->next(); fl[i] = n; return f; } template forceinline void MemoryManager::fl_dispose(FreeList* f, FreeList* l) { size_t i = sz2i(s); #ifdef GECODE_AUDIT FreeList* cur = f; while (cur != l) { assert(cur->next()); cur = cur->next(); } #endif l->next(fl[i]); fl[i] = f; } template void MemoryManager::fl_refill(SharedMemory& sm) { // Try to acquire memory from slack if (slack != nullptr) { MemoryChunk* m = slack; slack = nullptr; do { char* block = ptr_cast(m); size_t s = m->size; assert(s >= sz); m = m->next; fl[sz2i(sz)] = ptr_cast(block); while (s >= 2*sz) { ptr_cast(block)->next(ptr_cast(block+sz)); block += sz; s -= sz; } ptr_cast(block)->next(nullptr); } while (m != nullptr); } else { char* block = static_cast(alloc(sm,MemoryConfig::fl_refill*sz)); fl[sz2i(sz)] = ptr_cast(block); int i = MemoryConfig::fl_refill-2; do { ptr_cast(block+static_cast(i)*sz) ->next(ptr_cast(block+ (static_cast(i)+1)*sz)); } while (--i >= 0); ptr_cast(block+ (MemoryConfig::fl_refill-1)*sz)->next (ptr_cast(nullptr)); } } }} // STATISTICS: kernel-memory