ultimatepp/uppsrc/Core/sheap.cpp
2025-04-07 10:21:39 +02:00

654 lines
14 KiB
C++

#include <Core/Core.h>
#define LTIMING(x) // RTIMING(x)
namespace Upp {
#ifdef UPP_HEAP
#include "HeapImp.h"
#define LLOG(x) // LOG((void *)this << ' ' << x)
inline void Heap::Page::Format(int k)
{
DbgFreeFill(Begin(), End() - Begin());
klass = k;
active = 0;
int sz = Ksz(k);
byte *ptr = End() - sz;
byte *b = Begin();
FreeLink *l = NULL;
while(ptr >= b) {
((FreeLink *)ptr)->next = l;
l = (FreeLink *)ptr;
ptr -= sz;
}
freelist = l;
}
Heap::Page *Heap::WorkPage(int k) // get a new workpage with empty blocks
{
LLOG("AllocK - next work page not available " << k << " empty: " << (void *)empty[k]);
Page *page = empty[k]; // hot empty page of the same klass
empty[k] = NULL;
if(!page) { // try to reacquire pages freed remotely
LLOG("AllocK - trying FreeRemote");
SmallFreeRemote();
if(work[k]->freelist) { // partially free page found
LLOG("AllocK - work available after FreeRemote " << k);
return work[k];
}
page = empty[k]; // hot empty page
empty[k] = NULL;
}
if(!page)
for(int i = 0; i < NKLASS; i++) // Try hot empty page of different klass
if(empty[i]) {
LLOG("AllocK - free page available for reformatting " << k);
page = empty[i];
empty[i] = NULL;
page->Format(k); // reformat the page for the required klass
break;
}
if(!page) { // Attempt to find page in global storage of free pages
Mutex::Lock __(mutex);
if(this != &aux) { // Try to take the page from the aux
aux.SmallFreeRemoteRaw();
if(aux.work[k]->next != aux.work[k]) { // Try page of the same klass first
page = aux.work[k]->next;
page->Unlink();
LLOG("AllocK - adopting aux page " << k << " page: " << (void *)page << ", free " << (void *)page->freelist);
}
if(!page && aux.empty[k]) { // Try free page of the same klass (no need to format it)
page = aux.empty[k];
aux.empty[k] = page->next;
free_4KB--;
ASSERT(free_4KB < max_free_spages);
LLOG("AllocK - empty aux page available of the same format " << k << " page: " << (void *)page << ", free " << (void *)page->freelist);
}
if(!page)
for(int i = 0; i < NKLASS; i++) // Finally try to find free page of different klass
if(aux.empty[i]) {
page = aux.empty[i];
aux.empty[i] = page->next;
free_4KB--;
page->Format(k);
ASSERT(free_4KB < max_free_spages);
LLOG("AllocK - empty aux page available for reformatting " << k << " page: " << (void *)page << ", free " << (void *)page->freelist);
break;
}
}
if(!page) { // No free memory was found, ask huge for the new page
page = (Page *)HugeAlloc(1);
LLOG("AllocK - allocated new system page " << (void *)page << " " << k);
page->Format(k);
}
page->heap = this;
}
page->Link(work[k]);
ASSERT(page->klass == k);
return page;
}
void *Heap::AllocK(int k)
{
LLOG("AllocK " << k);
if(!initialized)
Init();
Page *page = work[k]->next;
for(;;) {
ASSERT(page->klass == k);
FreeLink *p = page->freelist;
if(p) {
LLOG("AllocK allocating from " << (void *)page << " " << (void *)p);
page->freelist = p->next;
++page->active;
return p;
}
LLOG("AllocK - page exhausted " << k << " page: " << (void *)page);
if(page->next != page) {
LLOG("Moving " << (void *)page << " to full");
page->Unlink();
page->Link(full[k]);
page = work[k]->next;
}
if(page->next == page)
page = WorkPage(k);
}
}
force_inline
void *Heap::Allok(int k)
{ // Try to alloc from the front-cache first
LTIMING("Allok");
FreeLink *ptr = cache[k];
if(ptr) {
cachen[k]++;
cache[k] = ptr->next;
return DbgFreeCheckK(ptr, k);
}
return DbgFreeCheckK(AllocK(k), k);
}
force_inline
void *Heap::AllocSz(size_t& sz)
{
LTIMING("Alloc");
LLOG("Alloc " << asString(sz));
Stat(sz);
if(sz <= 384) {
if(sz == 0)
sz = 1;
int k = ((int)sz - 1) >> 5;
sz = (k + 1) << 5;
return Allok(k);
}
if(sz <= 992) {
if(sz <= 448) {
sz = 448;
return Allok(12);
}
if(sz <= 576) {
sz = 576;
return Allok(13);
}
if(sz <= 672) {
sz = 672;
return Allok(14);
}
if(sz <= 800) {
sz = 800;
return Allok(15);
}
sz = 992;
return Allok(16);
}
return LAlloc(sz);
}
force_inline
void Heap::FreeK(void *ptr, Page *page, int k)
{
if(page->freelist) {
LLOG("Free next in work page " << k);
((FreeLink *)ptr)->next = page->freelist;
}
else {
LLOG("Free full to work " << k << " heap: " << (void *)page->heap);
page->Unlink();
page->Link(work[k]);
((FreeLink *)ptr)->next = NULL;
}
page->freelist = (FreeLink *)ptr;
if(--page->active == 0) { // there are no more allocated blocks in this page
LLOG("Free page is empty " << (void *)page);
page->Unlink();
if(this == &aux) {
LLOG("...is aux " << asString(free_4KB));
Mutex::Lock __(mutex);
Free4KB(k, page);
}
else {
if(empty[k]) { // Keep one hot empty page per klass in thread, put rest to 'aux' global storage
Mutex::Lock __(mutex);
Free4KB(k, empty[k]); // Free current hot page to reserve/huge
}
empty[k] = page; // this empty page is now hot
}
}
}
void Heap::Free4KB(int k, Page *page)
{ // put empty 4KB to aux reserve or back to huge blocks if the reserve is full
LLOG("Global Free4KB " << k << " " << (void *)empty);
if(free_4KB < max_free_spages) { // only keep max_free_spages, release if more
page->heap = &aux;
page->next = aux.empty[k];
aux.empty[k] = page;
free_4KB++;
LLOG("Reserve 4KB " << asString(free_4KB));
}
else {
aux.HugeFree(page);
LLOG("HugeFree 4KB " << asString(free_4KB));
}
}
force_inline
void Heap::Free(void *ptr, Page *page, int k)
{
LTIMING("Small Free");
LLOG("Small free page: " << (void *)page << ", k: " << k << ", ksz: " << Ksz(k));
ASSERT((4096 - ((uintptr_t)ptr & (uintptr_t)4095)) % Ksz(k) == 0);
if(page->heap != this) { // freeing block allocated in different thread
RemoteFree(ptr, Ksz(k)); // add to originating heap's list of free pages to be properly freed later
return;
}
DbgFreeFillK(ptr, k);
if(cachen[k]) {
cachen[k]--;
FreeLink *l = (FreeLink *)ptr;
l->next = cache[k];
cache[k] = l;
return;
}
FreeK(ptr, page, k);
}
force_inline
void Heap::Free(void *ptr)
{
if(!ptr) return;
LLOG("Free " << ptr);
if(IsSmall(ptr)) {
Page *page = GetPage(ptr);
Free(ptr, page, page->klass);
}
else
LFree(ptr);
}
size_t Heap::GetBlockSize(void *ptr)
{
if(!ptr) return 0;
LLOG("GetBlockSize " << ptr);
if(IsSmall(ptr)) {
Page *page = GetPage(ptr);
int k = page->klass;
return Ksz(k);
}
return LGetBlockSize(ptr);
}
void Heap::SmallFreeDirect(void *ptr)
{ // does not need to check for target heap or small vs large
LLOG("Free Direct " << ptr);
Page *page = GetPage(ptr);
ASSERT(page->heap == this);
int k = page->klass;
LLOG("Small free page: " << (void *)page << ", k: " << k << ", ksz: " << Ksz(k));
ASSERT((4096 - ((uintptr_t)ptr & (uintptr_t)4095)) % Ksz(k) == 0);
DbgFreeFillK(ptr, k);
FreeK(ptr, page, k);
}
bool Heap::FreeSmallEmpty(int size4KB, int count)
{ // attempt to release small 4KB pages to gain count4KB space or count of releases
bool released;
do {
released = false;
for(int i = 0; i < NKLASS; i++)
if(aux.empty[i]) {
Page *q = aux.empty[i];
aux.empty[i] = q->next;
free_4KB--;
ASSERT(free_4KB < max_free_spages);
if(aux.HugeFree(q) >= size4KB || --count <= 0) // HugeFree is really static, aux needed just to compile
return true;
released = true;
}
}
while(released);
return false;
}
force_inline
void *Heap::Alloc32()
{
Stat(32);
return Allok(KLASS_32);
}
force_inline
void Heap::Free(void *ptr, int k)
{
Free(ptr, GetPage(ptr), k);
}
force_inline
void Heap::Free32(void *ptr)
{
Free(ptr, KLASS_32);
}
force_inline
void *Heap::Alloc48()
{
Stat(48);
return Allok(KLASS_48);
}
force_inline
void Heap::Free48(void *ptr)
{
Free(ptr, KLASS_48);
}
static thread_local bool heap_closed__;
static thread_local Heap *heap_tls__;
void Heap::Shutdown()
{ // Move all active blocks, "orphans", to global aux heap
LLOG("**** Shutdown heap " << asString(this));
Mutex::Lock __(mutex);
heap_closed__ = true;
heap_tls__ = NULL;
Init();
RemoteFlushRaw(); // Move remote blocks to originating heaps
FreeRemoteRaw(); // Free all remotely freed blocks
for(int i = 0; i < NKLASS; i++) { // move all small pages to aux (some heap will pick them later)
LLOG("Free cache " << asString(i));
FreeLink *l = cache[i];
while(l) {
FreeLink *h = l;
l = l->next;
SmallFreeDirect(h);
}
while(full[i]->next != full[i]) {
Page *p = full[i]->next;
p->Unlink();
p->heap = &aux;
p->Link(aux.full[i]);
LLOG("Orphan full " << (void *)p);
}
while(work[i]->next != work[i]) {
Page *p = work[i]->next;
p->Unlink();
p->heap = &aux;
p->Link(p->freelist ? aux.work[i] : aux.full[i]);
LLOG("Orphan work " << (void *)p);
}
if(empty[i]) {
ASSERT(empty[i]->freelist);
ASSERT(empty[i]->active == 0);
Free4KB(i, empty[i]);
LLOG("Orphan empty " << (void *)empty[i]);
}
}
MoveLargeTo(&aux); // move all large pages to aux, some heap will pick them later
memset(this, 0, sizeof(Heap));
LLOG("++++ Done Shutdown heap " << asString(this));
}
never_inline
void EnterHeapMutex()
{
Heap::mutex.Enter();
}
never_inline
void LeaveHeapMutex()
{
Heap::mutex.Leave();
}
struct HeapMutexLock {
HeapMutexLock() { EnterHeapMutex(); }
~HeapMutexLock() { LeaveHeapMutex(); }
};
void MemoryFreeThread();
struct HeapExitThreadGuard {
void Used() {}
~HeapExitThreadGuard() { MemoryFreeThread(); }
};
Heap *MakeHeap()
{
if(heap_closed__)
return &Heap::aux;
if(heap_tls__)
return heap_tls__;
static thread_local byte sHeap__[sizeof(Heap)]; // zero initialization is fine for us
heap_tls__ = (Heap *)sHeap__;
static thread_local HeapExitThreadGuard __;
__.Used(); // "odr-used", register allocator to be shutdown at thread exit
return heap_tls__;
}
struct CurrentHeap {
bool locked = false;
Heap *heap;
Heap *operator->() { return heap; }
CurrentHeap() {
heap = heap_tls__;
if(!heap) {
EnterHeapMutex();
locked = true;
heap = MakeHeap();
}
}
~CurrentHeap() {
if(locked)
LeaveHeapMutex();
}
};
void MemoryFreek__(int klass, void *ptr)
{
Heap *heap = heap_tls__;
if(heap)
heap->Free((void *)ptr, klass);
else {
HeapMutexLock __;
MakeHeap()->Free((void *)ptr, klass);
}
}
void *MemoryAllok__(int klass)
{
Heap *heap = heap_tls__;
if(heap)
return heap->Allok(klass);
else {
HeapMutexLock __;
return MakeHeap()->Allok(klass);
}
}
#if defined(HEAPDBG)
void *MemoryAllocSz_(size_t& sz)
{
return CurrentHeap()->AllocSz(sz);
}
void MemoryFree_(void *ptr)
{
return CurrentHeap()->Free(ptr);
}
bool MemoryTryRealloc_(void *ptr, size_t& size)
{
return CurrentHeap()->TryRealloc(ptr, size);
}
size_t GetMemoryBlockSize_(void *ptr)
{
return CurrentHeap()->GetBlockSize(ptr);
}
#else
#ifdef flagHEAPLOG // log all heap allocations and deallocations
#undef AllocSz
StaticMutex sHeapLogLock;
static FILE *sLog = fopen(GetExeDirFile("heap.log"), "w");
void LogFree(void *ptr)
{
if(sLog) {
Mutex::Lock __(sHeapLogLock);
fprintf(sLog, "-%x %p\n", (int)Thread::GetCurrentId(), ptr);
}
}
void *LogAlloc(void *ptr, size_t sz)
{
if(sLog) {
Mutex::Lock __(sHeapLogLock);
fprintf(sLog, "%x %zx %p\n", (int)Thread::GetCurrentId(), sz, ptr);
}
return ptr;
}
#else
inline void LogFree(void *ptr) {}
inline void *LogAlloc(void *ptr, size_t sz) { return ptr; }
#endif
// xxx2 routines in the following code are to make things explicit for optimiser
never_inline
void *MemoryAlloc2(size_t& sz)
{
HeapMutexLock __;
return LogAlloc(MakeHeap()->AllocSz(sz), sz);
}
void *MemoryAlloc(size_t sz)
{
LTIMING("MemoryAlloc");
Heap *heap = heap_tls__;
if(heap)
return LogAlloc(heap->AllocSz(sz), sz);
return MemoryAlloc2(sz);
}
void *MemoryAllocSz(size_t& sz)
{
LTIMING("MemoryAllocSz");
Heap *heap = heap_tls__;
if(heap)
return LogAlloc(heap->AllocSz(sz), sz);
return MemoryAlloc2(sz);
}
never_inline
void MemoryFree2(void *ptr)
{
HeapMutexLock __;
MakeHeap()->Free(ptr);
}
void MemoryFree(void *ptr)
{
LTIMING("MemoryFree");
LogFree(ptr);
Heap *heap = heap_tls__;
if(heap)
heap->Free(ptr);
else
MemoryFree2(ptr);
}
size_t GetMemoryBlockSize(void *ptr)
{
return CurrentHeap()->GetBlockSize(ptr);
}
bool MemoryTryRealloc__(void *ptr, size_t& size)
{
Heap *heap = heap_tls__;
if(heap)
return heap->TryRealloc(ptr, size);
else {
HeapMutexLock __;
return MakeHeap()->TryRealloc(ptr, size);
}
}
never_inline
void *MemoryAlloc32_2()
{
HeapMutexLock __;
return LogAlloc(MakeHeap()->Alloc32(), 32);
}
force_inline
void *MemoryAlloc32_i()
{
LTIMING("MemoryAlloc32");
Heap *heap = heap_tls__;
if(heap)
return LogAlloc(heap->Alloc32(), 32);
return MemoryAlloc32_2();
}
void *MemoryAlloc32() { return MemoryAlloc32_i(); }
never_inline
void MemoryFree32_2(void *ptr)
{
HeapMutexLock __;
MakeHeap()->Free32(ptr);
}
force_inline
void MemoryFree32_i(void *ptr)
{
LTIMING("MemoryFree32");
LogFree(ptr);
Heap *heap = heap_tls__;
if(heap)
heap->Free32(ptr);
else
MemoryFree32_2(ptr);
}
void MemoryFree32(void *ptr) { MemoryFree32_i(ptr); }
#endif
void MemoryFreeThread()
{
if(heap_closed__)
return;
CurrentHeap()->Shutdown();
}
void MemoryCheck()
{
CurrentHeap()->Check();
}
MemoryProfile::MemoryProfile()
{
CurrentHeap()->Make(*this);
}
void MemoryDumpLarge()
{
CurrentHeap()->DumpLarge();;
}
void MemoryDumpHuge()
{
CurrentHeap()->DumpHuge();
}
#ifdef HEAPDBG
force_inline void *MemoryAlloc32_i() { return MemoryAlloc32(); }
force_inline void MemoryFree32_i(void *ptr) { return MemoryFree32(ptr); }
#endif
#include "StringMem.i"
#endif
}