summaryrefslogtreecommitdiffstats
path: root/src/kernel
diff options
context:
space:
mode:
authorDoug Gilbert <dgilbert@us.ibm.com>2011-09-26 13:36:33 -0500
committerDouglas R. Gilbert <dgilbert@us.ibm.com>2011-10-25 11:16:20 -0500
commit5ab488739184f2b2649193e3f9da695ee334d04f (patch)
tree3d47e74b8dd290598527988adccff0ff57c72dc0 /src/kernel
parentd127ad9d985ffd7a42dba798bee66654242c4fe6 (diff)
downloadtalos-hostboot-5ab488739184f2b2649193e3f9da695ee334d04f.tar.gz
talos-hostboot-5ab488739184f2b2649193e3f9da695ee334d04f.zip
new HEAP manager to reduce fragmentation
Change-Id: Ibe725a43e6366d9113ec99df1cc6aafa7bbb770e Reviewed-on: http://gfw160.austin.ibm.com:8080/gerrit/431 Tested-by: Jenkins Server Reviewed-by: A. Patrick Williams III <iawillia@us.ibm.com> Reviewed-by: Douglas R. Gilbert <dgilbert@us.ibm.com>
Diffstat (limited to 'src/kernel')
-rw-r--r--src/kernel/barrier.C42
-rw-r--r--src/kernel/cpumgr.C42
-rw-r--r--src/kernel/heapmgr.C494
-rw-r--r--src/kernel/makefile2
-rw-r--r--src/kernel/pagemgr.C90
-rw-r--r--src/kernel/scheduler.C2
-rw-r--r--src/kernel/syscall.C10
7 files changed, 633 insertions, 49 deletions
diff --git a/src/kernel/barrier.C b/src/kernel/barrier.C
new file mode 100644
index 000000000..a5dadc63a
--- /dev/null
+++ b/src/kernel/barrier.C
@@ -0,0 +1,42 @@
+// IBM_PROLOG_BEGIN_TAG
+// This is an automatically generated prolog.
+//
+// $Source: src/kernel/barrier.C $
+//
+// IBM CONFIDENTIAL
+//
+// COPYRIGHT International Business Machines Corp. 2011
+//
+// p1
+//
+// Object Code Only (OCO) source materials
+// Licensed Internal Code Source Materials
+// IBM HostBoot Licensed Internal Code
+//
+// The source code for this program is not published or other-
+// wise divested of its trade secrets, irrespective of what has
+// been deposited with the U.S. Copyright Office.
+//
+// Origin: 30
+//
+// IBM_PROLOG_END
+#include <kernel/barrier.H>
+#include <arch/ppc.H>
+
+void Barrier::wait()
+{
+ iv_spinlock.lock();
+ --iv_missing;
+ if(iv_missing > 0)
+ {
+ size_t l_event = iv_event;
+ iv_spinlock.unlock();
+ while(iv_event == l_event);
+ }
+ else
+ {
+ iv_missing = iv_count;
+ ++iv_event;
+ iv_spinlock.unlock();
+ }
+}
diff --git a/src/kernel/cpumgr.C b/src/kernel/cpumgr.C
index 86e8653bf..6ec3b9f6d 100644
--- a/src/kernel/cpumgr.C
+++ b/src/kernel/cpumgr.C
@@ -34,10 +34,14 @@
#include <sys/sync.h>
#include <kernel/cpuid.H>
#include <kernel/ptmgr.H>
+#include <kernel/heapmgr.H>
cpu_t* CpuManager::cv_cpus[CpuManager::MAXCPUS] = { NULL };
bool CpuManager::cv_shutdown_requested = false;
uint64_t CpuManager::cv_shutdown_status = 0;
+Barrier CpuManager::cv_barrier;
+bool CpuManager::cv_defrag = false;
+size_t CpuManager::cv_cpuCount = 0;
CpuManager::CpuManager()
{
@@ -150,6 +154,7 @@ void CpuManager::startCPU(ssize_t i)
if (currentCPU)
{
setDEC(TimeManager::getTimeSliceCount());
+ activateCPU(cv_cpus[i]);
}
return;
}
@@ -157,10 +162,18 @@ void CpuManager::startCPU(ssize_t i)
void CpuManager::startSlaveCPU(cpu_t* cpu)
{
setDEC(TimeManager::getTimeSliceCount());
+ activateCPU(cpu);
return;
}
+void CpuManager::activateCPU(cpu_t * i_cpu)
+{
+ i_cpu->active = true;
+ ++cv_cpuCount;
+ lwsync();
+}
+
void CpuManager::executePeriodics(cpu_t * i_cpu)
{
if(i_cpu->master)
@@ -169,11 +182,11 @@ void CpuManager::executePeriodics(cpu_t * i_cpu)
if(0 == (i_cpu->periodic_count % CPU_PERIODIC_CHECK_MEMORY))
{
uint64_t pcntAvail = PageManager::queryAvail();
- if(pcntAvail < 16) // Less than 16% pages left TODO 16 ok?
+ if(pcntAvail < 16)
{
VmmManager::flushPageTable();
++(i_cpu->periodic_count); // prevent another flush below
- if(pcntAvail < 5) // TODO 5% ok
+ if(pcntAvail < 5)
{
VmmManager::castOutPages(VmmManager::CRITICAL);
}
@@ -187,6 +200,31 @@ void CpuManager::executePeriodics(cpu_t * i_cpu)
{
VmmManager::flushPageTable();
}
+ if(0 == (i_cpu->periodic_count % CPU_PERIODIC_DEFRAG))
+ {
+ // set up barrier based on # cpus cv_barrier;
+ // TODO whatif other cpus become active?
+ isync();
+ cv_barrier.init(cv_cpuCount);
+ if(!cv_shutdown_requested)
+ {
+ cv_defrag = true;
+ lwsync();
+ }
+ }
+ }
+ if(cv_defrag)
+ {
+ cv_barrier.wait();
+
+ if(i_cpu->master)
+ {
+ HeapManager::coalesce();
+ PageManager::coalesce();
+ cv_defrag = false;
+ }
+
+ cv_barrier.wait();
}
}
diff --git a/src/kernel/heapmgr.C b/src/kernel/heapmgr.C
index 7d3757b9f..e3d2ad1ea 100644
--- a/src/kernel/heapmgr.C
+++ b/src/kernel/heapmgr.C
@@ -25,87 +25,511 @@
#include <util/singleton.H>
#include <kernel/console.H>
#include <kernel/pagemgr.H>
+#include <util/align.H>
+#include <arch/ppc.H>
+
+#ifdef HOSTBOOT_DEBUG
+#define SMALL_HEAP_PAGES_TRACKED 64
+
+// track pages allocated to smallheap
+void * g_smallHeapPages[SMALL_HEAP_PAGES_TRACKED];
+
+// If these stats are to be kept then they should be modified using
+// atomic instructions
+uint16_t g_bucket_counts[HeapManager::BUCKETS];
+uint32_t g_smallheap_allocated = 0; // sum of currently allocated
+uint32_t g_smallheap_alloc_hw = 0; // allocated high water
+uint32_t g_smallheap_pages = 0; // total bytes high water (pages used)
+uint32_t g_smallheap_count = 0; // # of chunks allocated
+
+uint32_t g_big_chunks = 0;
+uint32_t g_bigheap_highwater = 0;
+#endif
+
+const size_t HeapManager::cv_chunk_size[BUCKETS] =
+{
+ HeapManager::BUCKET_SIZE0,
+ HeapManager::BUCKET_SIZE1,
+ HeapManager::BUCKET_SIZE2,
+ HeapManager::BUCKET_SIZE3,
+ HeapManager::BUCKET_SIZE4,
+ HeapManager::BUCKET_SIZE5,
+ HeapManager::BUCKET_SIZE6,
+ HeapManager::BUCKET_SIZE7,
+ HeapManager::BUCKET_SIZE8,
+ HeapManager::BUCKET_SIZE9,
+ HeapManager::BUCKET_SIZE10,
+ HeapManager::BUCKET_SIZE11
+};
+
+size_t HeapManager::cv_coalesce_count = 0;
+size_t HeapManager::cv_free_bytes;
+size_t HeapManager::cv_free_chunks;
+
void HeapManager::init()
{
Singleton<HeapManager>::instance();
}
-void* HeapManager::allocate(size_t n)
+void * HeapManager::allocate(size_t i_sz)
{
HeapManager& hmgr = Singleton<HeapManager>::instance();
- return hmgr._allocate(n);
+ if(i_sz > MAX_SMALL_ALLOC_SIZE)
+ {
+ return hmgr._allocateBig(i_sz);
+ }
+ return hmgr._allocate(i_sz);
}
-void HeapManager::free(void* p)
+void HeapManager::free(void * i_ptr)
{
HeapManager& hmgr = Singleton<HeapManager>::instance();
- return hmgr._free(p);
+ return hmgr._free(i_ptr);
}
-void* HeapManager::_allocate(size_t n)
+void* HeapManager::realloc(void* i_ptr, size_t i_sz)
{
- size_t which_bucket = 0;
- while (n > (size_t)((1 << (which_bucket + 4)) - 8)) which_bucket++;
+ return Singleton<HeapManager>::instance()._realloc(i_ptr,i_sz);
+}
- chunk_t* chunk = (chunk_t*)NULL;
+void HeapManager::coalesce( void )
+{
+ Singleton<HeapManager>::instance()._coalesce();
+}
+
+void* HeapManager::_allocate(size_t i_sz)
+{
+ size_t which_bucket = bucketIndex(i_sz+8);
+
+ chunk_t* chunk = reinterpret_cast<chunk_t*>(NULL);
chunk = pop_bucket(which_bucket);
if (NULL == chunk)
{
newPage();
- return _allocate(n);
+ return _allocate(i_sz);
}
else
{
- return &chunk->next;
+#ifdef HOSTBOOT_DEBUG
+ size_t alloc = bucketByteSize(chunk->bucket);
+ __sync_add_and_fetch(&g_smallheap_count,1);
+ __sync_add_and_fetch(&g_smallheap_allocated,alloc);
+ if (g_smallheap_allocated > g_smallheap_alloc_hw)
+ g_smallheap_alloc_hw = g_smallheap_allocated;
+ // test_pages();
+#endif
+ chunk->free = false;
+ return &chunk->next;
+ }
+}
+
+
+void* HeapManager::_realloc(void* i_ptr, size_t i_sz)
+{
+ void* new_ptr = _reallocBig(i_ptr,i_sz);
+ if(new_ptr) return new_ptr;
+
+ new_ptr = i_ptr;
+ chunk_t* chunk = reinterpret_cast<chunk_t*>(((size_t*)i_ptr)-1);
+ size_t asize = bucketByteSize(chunk->bucket)-8;
+ if(asize < i_sz)
+ {
+ new_ptr = _allocate(i_sz);
+ memcpy(new_ptr, i_ptr, asize);
+ _free(i_ptr);
}
+ return new_ptr;
}
-void HeapManager::_free(void* p)
+void* HeapManager::_reallocBig(void* i_ptr, size_t i_sz)
{
- if (NULL == p) return;
+ // Currently all large allocations fall on a page boundry,
+ // but small allocatoins never do
+ if(ALIGN_PAGE(reinterpret_cast<uint64_t>(i_ptr)) !=
+ reinterpret_cast<uint64_t>(i_ptr))
+ {
+ return NULL;
+ }
+
+ void* new_ptr = NULL;
+ big_chunk_t * bc = big_chunk_stack.first();
+ while(bc)
+ {
+ if(bc->addr == i_ptr)
+ {
+ size_t new_size = ALIGN_PAGE(i_sz)/PAGESIZE;
+ if(new_size > bc->page_count)
+ {
+#ifdef HOSTBOOT_DEBUG
+ __sync_add_and_fetch(&g_big_chunks,new_size-bc->page_count);
+ if(g_bigheap_highwater < g_big_chunks)
+ g_bigheap_highwater = g_big_chunks;
+#endif
+ new_ptr = PageManager::allocatePage(new_size);
- chunk_t* chunk = (chunk_t*)(((size_t*)p)-1);
- push_bucket(chunk, chunk->len);
+ memcpy(new_ptr,i_ptr,bc->page_count*PAGESIZE);
+
+ size_t page_count = bc->page_count;
+ bc->addr = new_ptr;
+ bc->page_count = new_size;
+ lwsync();
+
+ PageManager::freePage(i_ptr,page_count);
+ }
+ new_ptr = bc->addr;
+
+ break;
+ }
+ bc = bc->next;
+ }
+ return new_ptr;
+}
+
+void HeapManager::_free(void * i_ptr)
+{
+ if (NULL == i_ptr) return;
+
+ if(!_freeBig(i_ptr))
+ {
+ chunk_t* chunk = reinterpret_cast<chunk_t*>(((size_t*)i_ptr)-1);
+#ifdef HOSTBOOT_DEBUG
+ __sync_sub_and_fetch(&g_smallheap_count,1);
+ __sync_sub_and_fetch(&g_smallheap_allocated,bucketByteSize(chunk->bucket));
+#endif
+ push_bucket(chunk, chunk->bucket);
+ }
}
-HeapManager::chunk_t* HeapManager::pop_bucket(size_t n)
+
+HeapManager::chunk_t* HeapManager::pop_bucket(size_t i_bucket)
{
- if (n >= BUCKETS) return NULL;
+ if (i_bucket >= BUCKETS) return NULL;
- chunk_t* c = first_chunk[n].pop();
+ chunk_t* c = first_chunk[i_bucket].pop();
if (NULL == c)
{
- // Couldn't allocate from the correct size bucket, so split up an
- // item from the next sized bucket.
- c = pop_bucket(n+1);
- if (NULL != c)
- {
- chunk_t* c1 = (chunk_t*)(((uint8_t*)c) + (1 << (n + 4)));
- c->len = n;
- c1->len = n;
- push_bucket(c1, n);
- }
+ // Couldn't allocate from the correct size bucket, so split up an
+ // item from the next sized bucket.
+ c = pop_bucket(i_bucket+1);
+ if (NULL != c)
+ {
+ size_t c_size = bucketByteSize(i_bucket);
+ size_t c1_size = bucketByteSize(c->bucket) - c_size;
+ size_t c1_bucket = bucketIndex(c1_size);
+
+ chunk_t* c1 = reinterpret_cast<chunk_t*>(((uint8_t*)c) + c_size);
+ c1->bucket = c1_bucket;
+ c->bucket = i_bucket;
+
+ // c1_size should always be a valid size unless the FIB sequence is modified
+ // then we could end up with an 8 byte piece of junk.
+ if(c1_size >= MIN_BUCKET_SIZE)
+ {
+ push_bucket(c1, c1_bucket);
+ }
+ }
}
return c;
}
-void HeapManager::push_bucket(chunk_t* c, size_t n)
+
+void HeapManager::push_bucket(chunk_t* i_chunk, size_t i_bucket)
{
- if (n >= BUCKETS) return;
- first_chunk[n].push(c);
+ if (i_bucket >= BUCKETS) return;
+ i_chunk->free = true;
+ first_chunk[i_bucket].push(i_chunk);
}
+
void HeapManager::newPage()
{
void* page = PageManager::allocatePage();
- chunk_t * c = (chunk_t*)page;
- for (uint64_t i = 0; i < (PAGESIZE / (1 << (BUCKETS + 3))); i++)
+ chunk_t * c = reinterpret_cast<chunk_t*>(page);
+ size_t remaining = PAGESIZE;
+
+#ifdef HOSTBOOT_DEBUG
+ uint32_t idx = __sync_fetch_and_add(&g_smallheap_pages,1);
+ if(idx < SMALL_HEAP_PAGES_TRACKED)
+ g_smallHeapPages[idx] = page;
+#endif
+
+ while(remaining >= MIN_BUCKET_SIZE)
+ {
+ size_t bucket = bucketIndex(remaining);
+
+ // bucket might be one too big
+ if(bucket == BUCKETS || bucketByteSize(bucket) > remaining)
+ {
+ --bucket;
+ }
+ c->bucket = bucket;
+ push_bucket(c, bucket);
+
+ size_t bsize = bucketByteSize(bucket);
+ c = reinterpret_cast<chunk_t*>(((uint8_t*)c) + bsize);
+ remaining -= bsize;
+ }
+ // Note: if the fibonacci series originally used is modified, there could
+ // be a remainder. Thow it away.
+}
+
+// find smallest bucket i_sz will fit into
+size_t HeapManager::bucketIndex(size_t i_sz)
+{
+
+ // A simple linear search loop is unrolled by the compiler
+ // and generates large asm code.
+ //
+ // A manual unrole of a binary search using "if" statements is 160 bytes
+ // for this function and 160 bytes for the bucketByteSize() function
+ // but does not need the 96 byte cv_chunk_size array. Total 320 bytes
+ //
+ // This function is 120 bytes and it scales if more buckets are added
+ // bucketByteSize() using the static array uses 96 bytes. Total = 216 bytes
+
+ if(i_sz > cv_chunk_size[BUCKETS-1]) return BUCKETS;
+
+ // binary search
+ int64_t high_idx = BUCKETS - 1;
+ int64_t low_idx = 0;
+ size_t bucket = 0;
+ while(low_idx <= high_idx)
{
- c->len = BUCKETS-1;
- push_bucket(c, BUCKETS-1);
- c = (chunk_t*)(((uint8_t*)c) + (1 << (BUCKETS + 3)));
+ bucket = (low_idx + high_idx) / 2;
+ if( i_sz > bucketByteSize(bucket))
+ {
+ low_idx = bucket + 1;
+ }
+ else
+ {
+ high_idx = bucket - 1;
+ if(i_sz > bucketByteSize(high_idx)) // high_idx would be too small
+ break;
+ }
}
+ return bucket;
}
+
+
+// all other processes must be quiesced
+void HeapManager::_coalesce()
+{
+ // Clean up the smaller heap
+ bool merged_one = false;
+ chunk_t head; head.next = NULL;
+ chunk_t * c = NULL;
+
+ // make a chain out of all the free chunks
+ // OPTION: A priority queue sorted by address could be used.
+ // This would improve performance when searching for chunks to coalesce, but
+ // would require a minimum bucket size of 32 bytes.
+ // TODO: sort by address
+ for(size_t bucket = 0; bucket < BUCKETS; ++bucket)
+ {
+ while(NULL != (c = first_chunk[bucket].pop()))
+ {
+ c->next = head.next;
+ head.next = c;
+ }
+ }
+
+ do
+ {
+ merged_one = false;
+ // look for chunks to coalesce
+ for(c = head.next; c!=NULL; c = c->next)
+ {
+ size_t s = bucketByteSize(c->bucket);
+ chunk_t * c_find = reinterpret_cast<chunk_t*>(((uint8_t*)c) + s);
+
+ // c_find must be in same page
+ if(reinterpret_cast<uint64_t>(c_find) <
+ ALIGN_PAGE(reinterpret_cast<uint64_t>(c)))
+ {
+ if(c_find->free)
+ {
+ // is combinable?
+ size_t ns = s + bucketByteSize(c_find->bucket);
+ size_t n_bucket = bucketIndex(ns);
+ if(n_bucket < BUCKETS && bucketByteSize(n_bucket) == ns)
+ {
+ // find c_find in the free chain and remove it
+ chunk_t * c_prev = &head;
+ chunk_t * c_seek = head.next;
+ while(c_seek != NULL)
+ {
+ if(c_find == c_seek) // found it -> merge
+ {
+ // new bigger chunk
+ c->bucket = n_bucket;
+ merged_one = true;
+ ++cv_coalesce_count;
+
+ // remove found elment from the chain
+ c_prev->next = c_find->next;
+
+ break;
+ }
+ c_prev = c_seek;
+ c_seek = c_seek->next;
+ }
+ }
+ }
+ }
+ }
+ } while(merged_one); // repeat until nothing can be merged
+
+ // restore the free buckets
+ cv_free_chunks = 0;
+ cv_free_bytes = 0;
+ c = head.next;
+ while(c != NULL)
+ {
+ chunk_t * c_next = c->next;
+ push_bucket(c,c->bucket);
+ ++cv_free_chunks;
+ cv_free_bytes += bucketByteSize(c->bucket) - 8;
+ c = c_next;
+ }
+ printkd("HeapMgr coalesced total %ld\n",cv_coalesce_count);
+ test_pages();
+}
+
+void HeapManager::stats()
+{
+ coalesce(); // collects some of the stats
+
+ printkd("Memory Heap Stats:\n");
+ printkd(" %d Large heap pages allocated.\n",g_big_chunks);
+ printkd(" %d Large heap max allocated.\n",g_bigheap_highwater);
+ printkd(" %d Small heap pages.\n",g_smallheap_pages);
+ printkd(" %d Small heap bytes max allocated\n",g_smallheap_alloc_hw);
+ printkd(" %d Small heap bytes allocated in %d chunks\n",
+ g_smallheap_allocated,g_smallheap_count);
+ printkd(" %ld Small heap free bytes in %ld chunks\n",cv_free_bytes,cv_free_chunks);
+ printkd(" %ld Small heap total chunks coalesced\n",cv_coalesce_count);
+ printkd("Small heap bucket profile:\n");
+ for(size_t i = 0; i < BUCKETS; ++i)
+ {
+ printkd(" %d chunks of bytesize %ld\n",
+ g_bucket_counts[i],
+ cv_chunk_size[i]-8);
+ }
+
+ PageManager::coalesce();
+}
+
+
+void HeapManager::test_pages()
+{
+#ifdef HOSTBOOT_DEBUG
+ for(size_t i = 0; i < BUCKETS; ++i)
+ g_bucket_counts[i] = 0;
+
+ size_t max_idx = g_smallheap_pages;
+ if(max_idx > SMALL_HEAP_PAGES_TRACKED) max_idx = SMALL_HEAP_PAGES_TRACKED;
+ for(size_t i = 0; i < max_idx; ++i)
+ {
+ chunk_t* c = reinterpret_cast<chunk_t*>(g_smallHeapPages[i]);
+ uint8_t* c_prev = reinterpret_cast<uint8_t*>(c);
+ size_t sum = 0;
+ while(sum <= (PAGESIZE-MIN_BUCKET_SIZE))
+ {
+ size_t b = c->bucket;
+ if(b < BUCKETS)
+ {
+ size_t s = bucketByteSize(b);
+ c_prev = reinterpret_cast<uint8_t*>(c);
+ c = reinterpret_cast<chunk_t*>(((uint8_t*)c) + s);
+ sum += s;
+ ++g_bucket_counts[b];
+ }
+ else
+ {
+ printk("Heaptest: Corruption at %p on page %p."
+ " Owner of %p may have scribbled on it\n",
+ c,g_smallHeapPages[i],c_prev+8);
+ sum = PAGESIZE;
+ break;
+ }
+ }
+ if(sum > PAGESIZE)
+ {
+ printk("Heaptest: Page %p failed consistancy test\n",g_smallHeapPages[i]);
+ }
+ }
+#endif
+}
+
+void* HeapManager::_allocateBig(size_t i_sz)
+{
+ size_t pages = ALIGN_PAGE(i_sz)/PAGESIZE;
+ void* v = PageManager::allocatePage(pages);
+#ifdef HOSTBOOT_DEBUG
+ //printk("HEAPAB %p:%ld:%ld wasted %ld\n",v,pages,i_sz, pages*PAGESIZE - i_sz);
+ __sync_add_and_fetch(&g_big_chunks,pages);
+ if(g_bigheap_highwater < g_big_chunks)
+ g_bigheap_highwater = g_big_chunks;
+#endif
+
+ // If already have unused big_chunk_t object available then use it
+ // otherwise create a new one.
+ big_chunk_t * bc = big_chunk_stack.first();
+ while(bc)
+ {
+ if(bc->page_count == 0)
+ {
+ if(__sync_bool_compare_and_swap(&bc->addr,NULL,v))
+ {
+ bc->page_count = pages;
+ break;
+ }
+ }
+ bc = bc->next;
+ }
+ if(!bc)
+ {
+ bc = new big_chunk_t(v,pages);
+ big_chunk_stack.push(bc);
+ }
+
+ return v;
+}
+
+bool HeapManager::_freeBig(void* i_ptr)
+{
+ // Currently all large allocations fall on a page boundry,
+ // but small allocations never do
+ if(ALIGN_PAGE(reinterpret_cast<uint64_t>(i_ptr)) !=
+ reinterpret_cast<uint64_t>(i_ptr))
+ return false;
+
+ bool result = false;
+ big_chunk_t * bc = big_chunk_stack.first();
+ while(bc)
+ {
+ if(bc->addr == i_ptr)
+ {
+#ifdef HOSTBOOT_DEBUG
+ __sync_sub_and_fetch(&g_big_chunks,bc->page_count);
+#endif
+ size_t page_count = bc->page_count;
+ bc->page_count = 0;
+ bc->addr = NULL;
+ lwsync();
+
+ PageManager::freePage(i_ptr,page_count);
+
+ // no way to safely remove object from chain so leave it
+
+ result = true;
+ break;
+ }
+ bc = bc->next;
+ }
+ return result;
+}
+
diff --git a/src/kernel/makefile b/src/kernel/makefile
index 1ebcf98d5..2ec3db0d9 100644
--- a/src/kernel/makefile
+++ b/src/kernel/makefile
@@ -26,7 +26,7 @@ OBJS = start.o kernel.o console.o pagemgr.o heapmgr.o taskmgr.o cpumgr.o
OBJS += syscall.o scheduler.o spinlock.o exception.o vmmmgr.o timemgr.o
OBJS += futexmgr.o ptmgr.o segmentmgr.o devicesegment.o basesegment.o
OBJS += block.o cpuid.o misc.o msghandler.o blockmsghdlr.o stacksegment.o
-OBJS += softpatch_p7.o
+OBJS += softpatch_p7.o barrier.o
include ${ROOTPATH}/config.mk
diff --git a/src/kernel/pagemgr.C b/src/kernel/pagemgr.C
index 826e1bc4d..82e743ef1 100644
--- a/src/kernel/pagemgr.C
+++ b/src/kernel/pagemgr.C
@@ -24,8 +24,10 @@
#include <kernel/pagemgr.H>
#include <util/singleton.H>
#include <kernel/console.H>
-#include <sys/vfs.h>
#include <arch/ppc.H>
+#include <util/locked/pqueue.H>
+
+size_t PageManager::cv_coalesce_count = 0;
void PageManager::init()
{
@@ -51,14 +53,11 @@ uint64_t PageManager::queryAvail()
PageManager::PageManager() : iv_pagesAvail(0), iv_pagesTotal(0)
{
- // Determine first page of un-allocated memory.
- uint64_t addr = (uint64_t) VFS_LAST_ADDRESS;
- if (0 != (addr % PAGESIZE))
- addr = (addr - (addr % PAGESIZE)) + PAGESIZE;
-
- // Determine number of pages available.
- page_t* page = (page_t*)((void*) addr);
+ // Determine first page of un-allocated memory
+ // and number of pages available.
+ uint64_t addr = firstPageAddr();
size_t length = (MEMLEN - addr) / PAGESIZE;
+ page_t* page = reinterpret_cast<page_t *>(addr);
iv_pagesTotal = length;
// Update statistics.
@@ -70,7 +69,7 @@ PageManager::PageManager() : iv_pagesAvail(0), iv_pagesTotal(0)
(uint64_t)page);
// Populate L3 cache lines.
- uint64_t* cache_line = (uint64_t*) addr;
+ uint64_t* cache_line = reinterpret_cast<uint64_t*>(addr);
uint64_t* end_cache_line = (uint64_t*) VmmManager::FULL_MEM_SIZE;
while (cache_line != end_cache_line)
{
@@ -163,3 +162,76 @@ void PageManager::push_bucket(page_t* p, size_t n)
first_page[n].push(p);
}
+void PageManager::coalesce( void )
+{
+ Singleton<PageManager>::instance()._coalesce();
+}
+
+
+// Coalsesce adjacent free memory blocks
+void PageManager::_coalesce( void )
+{
+ // Look at all the "free buckets" and find blocks to merge
+ // Since this is binary, all merges will be from the same free bucket
+ // Each bucket is a stack of non-allocated memory blocks of the same size
+ // Once two blocks are merged they become a single block twice the size.
+ // The source blocks must be removed from the current bucket (stack) and
+ // the new block needs to be pushed onto the next biggest stack.
+ for(size_t bucket = 0; bucket < (BUCKETS-1); ++bucket)
+ {
+ // Move the this stack bucket into a priority queue
+ // sorted by address, highest to lowest
+ Util::Locked::PQueue<page_t,page_t*> pq;
+ page_t * p = NULL;
+ while(NULL != (p = first_page[bucket].pop()))
+ {
+ p->key = p;
+ pq.insert(p);
+ }
+
+ while(NULL != (p = pq.remove()))
+ {
+ // p needs to be the even buddy to prevent merging of wrong block.
+ // To determine this, get the index of the block as if the whole
+ // page memory space were blocks of this size.
+ uint64_t p_idx = (reinterpret_cast<uint64_t>(p) - firstPageAddr())/
+ ((1 << bucket)*PAGESIZE);
+ if(0 != (p_idx % 2)) // odd index
+ {
+ first_page[bucket].push(p); // can't merge
+ }
+ else // it's even
+ {
+ // If p can be merged then the next block in pq will be the
+ // match. The address of p also can't be greater than what's
+ // in pq or something is really messed up, therefore if
+ // pq.remove_if() returns something then it's a match.
+ page_t * p_seek = (page_t*)((uint64_t)p +
+ (1 << bucket)*PAGESIZE);
+ page_t * p_next = pq.remove_if(p_seek);
+ if(p_next == p_seek)
+ {
+ // new block is twice the size and goes into the next
+ // bucket size
+ push_bucket(p,bucket+1);
+ ++cv_coalesce_count;
+ }
+ else
+ {
+ // Can't merge p
+ first_page[bucket].push(p);
+
+ if(p_next) // This should be null - if then overlaping mem
+ {
+ first_page[bucket].push(p_next);
+ printk("pagemgr::coalesce Expected %p, got %p\n",
+ p_seek, p_next);
+ }
+ }
+ }
+ }
+ }
+ printkd("PAGEMGR coalesced total %ld\n", cv_coalesce_count);
+}
+
+
diff --git a/src/kernel/scheduler.C b/src/kernel/scheduler.C
index 8a8da89a4..10dda6708 100644
--- a/src/kernel/scheduler.C
+++ b/src/kernel/scheduler.C
@@ -81,8 +81,6 @@ void Scheduler::setNextRunnable()
task_t* t = NULL;
cpu_t* cpu = CpuManager::getCurrentCPU();
- CpuManager::executePeriodics(cpu);
-
// Check for ready task in local run-queue, if it exists.
if (NULL != cpu->scheduler_extra)
{
diff --git a/src/kernel/syscall.C b/src/kernel/syscall.C
index 53fdb287e..87cfba6bd 100644
--- a/src/kernel/syscall.C
+++ b/src/kernel/syscall.C
@@ -38,6 +38,7 @@
#include <kernel/msghandler.H>
#include <kernel/vmmmgr.H>
#include <kernel/stacksegment.H>
+#include <kernel/heapmgr.H>
extern "C"
void kernel_execute_decrementer()
@@ -47,8 +48,17 @@ void kernel_execute_decrementer()
TimeManager::checkReleaseTasks(s);
s->returnRunnable();
+ CpuManager::executePeriodics(c);//TODO is there still a potential deadlock?
+
if (CpuManager::isShutdownRequested())
{
+ // The code below could cause a hang during shutdown
+ // The stats can be retrieved from global variables as needed.
+ // This can be uncommented for debug if desired
+#ifdef __MEMSTATS__
+ if(c->master)
+ HeapManager::stats();
+#endif
KernelMisc::shutdown();
}
OpenPOWER on IntegriCloud