vm: make heap data-type object-oriented
parent
d017a53227
commit
c046ff4b23
2
Makefile
2
Makefile
|
@ -38,7 +38,6 @@ DLL_OBJS = $(PLAF_DLL_OBJS) \
|
|||
vm/byte_arrays.o \
|
||||
vm/callstack.o \
|
||||
vm/code_block.o \
|
||||
vm/code_gc.o \
|
||||
vm/code_heap.o \
|
||||
vm/contexts.o \
|
||||
vm/data_gc.o \
|
||||
|
@ -47,6 +46,7 @@ DLL_OBJS = $(PLAF_DLL_OBJS) \
|
|||
vm/dispatch.o \
|
||||
vm/errors.o \
|
||||
vm/factor.o \
|
||||
vm/heap.o \
|
||||
vm/image.o \
|
||||
vm/inline_cache.o \
|
||||
vm/io.o \
|
||||
|
|
|
@ -371,7 +371,7 @@ void factor_vm::update_word_references(code_block *compiled)
|
|||
the code heap with dead PICs that will be freed on the next
|
||||
GC, we add them to the free list immediately. */
|
||||
else if(compiled->type == PIC_TYPE)
|
||||
heap_free(&code,compiled);
|
||||
code->heap_free(compiled);
|
||||
else
|
||||
{
|
||||
iterate_relocations(compiled,factor::update_word_references_step);
|
||||
|
@ -411,7 +411,7 @@ void factor_vm::mark_code_block(code_block *compiled)
|
|||
{
|
||||
check_code_address((cell)compiled);
|
||||
|
||||
mark_block(compiled);
|
||||
code->mark_block(compiled);
|
||||
|
||||
copy_handle(&compiled->literals);
|
||||
copy_handle(&compiled->relocation);
|
||||
|
@ -503,19 +503,19 @@ void factor_vm::fixup_labels(array *labels, code_block *compiled)
|
|||
/* Might GC */
|
||||
code_block *factor_vm::allot_code_block(cell size)
|
||||
{
|
||||
heap_block *block = heap_allot(&code,size + sizeof(code_block));
|
||||
heap_block *block = code->heap_allot(size + sizeof(code_block));
|
||||
|
||||
/* If allocation failed, do a code GC */
|
||||
if(block == NULL)
|
||||
{
|
||||
gc();
|
||||
block = heap_allot(&code,size + sizeof(code_block));
|
||||
block = code->heap_allot(size + sizeof(code_block));
|
||||
|
||||
/* Insufficient room even after code GC, give up */
|
||||
if(block == NULL)
|
||||
{
|
||||
cell used, total_free, max_free;
|
||||
heap_usage(&code,&used,&total_free,&max_free);
|
||||
code->heap_usage(&used,&total_free,&max_free);
|
||||
|
||||
print_string("Code heap stats:\n");
|
||||
print_string("Used: "); print_cell(used); nl();
|
||||
|
|
|
@ -1,38 +0,0 @@
|
|||
namespace factor
|
||||
{
|
||||
|
||||
static const cell free_list_count = 16;
|
||||
static const cell block_size_increment = 32;
|
||||
|
||||
struct heap_free_list {
|
||||
free_heap_block *small_blocks[free_list_count];
|
||||
free_heap_block *large_blocks;
|
||||
};
|
||||
|
||||
struct heap {
|
||||
segment *seg;
|
||||
heap_free_list free;
|
||||
};
|
||||
|
||||
typedef void (*heap_iterator)(heap_block *compiled,factor_vm *vm);
|
||||
|
||||
inline static heap_block *next_block(heap *h, heap_block *block)
|
||||
{
|
||||
cell next = ((cell)block + block->size);
|
||||
if(next == h->seg->end)
|
||||
return NULL;
|
||||
else
|
||||
return (heap_block *)next;
|
||||
}
|
||||
|
||||
inline static heap_block *first_block(heap *h)
|
||||
{
|
||||
return (heap_block *)h->seg->start;
|
||||
}
|
||||
|
||||
inline static heap_block *last_block(heap *h)
|
||||
{
|
||||
return (heap_block *)h->seg->end;
|
||||
}
|
||||
|
||||
}
|
|
@ -6,12 +6,12 @@ namespace factor
|
|||
/* Allocate a code heap during startup */
|
||||
void factor_vm::init_code_heap(cell size)
|
||||
{
|
||||
new_heap(&code,size);
|
||||
code = new heap(this,size);
|
||||
}
|
||||
|
||||
bool factor_vm::in_code_heap_p(cell ptr)
|
||||
{
|
||||
return (ptr >= code.seg->start && ptr <= code.seg->end);
|
||||
return (ptr >= code->seg->start && ptr <= code->seg->end);
|
||||
}
|
||||
|
||||
/* Compile a word definition with the non-optimizing compiler. Allocates memory */
|
||||
|
@ -31,13 +31,13 @@ void factor_vm::jit_compile_word(cell word_, cell def_, bool relocate)
|
|||
/* Apply a function to every code block */
|
||||
void factor_vm::iterate_code_heap(code_heap_iterator iter)
|
||||
{
|
||||
heap_block *scan = first_block(&code);
|
||||
heap_block *scan = code->first_block();
|
||||
|
||||
while(scan)
|
||||
{
|
||||
if(scan->status != B_FREE)
|
||||
iter((code_block *)scan,this);
|
||||
scan = next_block(&code,scan);
|
||||
scan = code->next_block(scan);
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -112,8 +112,8 @@ PRIMITIVE_FORWARD(modify_code_heap)
|
|||
inline void factor_vm::primitive_code_room()
|
||||
{
|
||||
cell used, total_free, max_free;
|
||||
heap_usage(&code,&used,&total_free,&max_free);
|
||||
dpush(tag_fixnum(code.seg->size / 1024));
|
||||
code->heap_usage(&used,&total_free,&max_free);
|
||||
dpush(tag_fixnum(code->seg->size / 1024));
|
||||
dpush(tag_fixnum(used / 1024));
|
||||
dpush(tag_fixnum(total_free / 1024));
|
||||
dpush(tag_fixnum(max_free / 1024));
|
||||
|
@ -220,20 +220,20 @@ void factor_vm::compact_code_heap()
|
|||
gc();
|
||||
|
||||
/* Figure out where the code heap blocks are going to end up */
|
||||
cell size = compute_heap_forwarding(&code, forwarding);
|
||||
cell size = code->compute_heap_forwarding(forwarding);
|
||||
|
||||
/* Update word and quotation code pointers */
|
||||
forward_object_xts();
|
||||
|
||||
/* Actually perform the compaction */
|
||||
compact_heap(&code,forwarding);
|
||||
code->compact_heap(forwarding);
|
||||
|
||||
/* Update word and quotation XTs */
|
||||
fixup_object_xts();
|
||||
|
||||
/* Now update the free list; there will be a single free block at
|
||||
the end */
|
||||
build_free_list(&code,size);
|
||||
code->build_free_list(size);
|
||||
}
|
||||
|
||||
}
|
||||
|
|
|
@ -1,5 +1,6 @@
|
|||
namespace factor
|
||||
{
|
||||
|
||||
struct factor_vm;
|
||||
typedef void (*code_heap_iterator)(code_block *compiled, factor_vm *myvm);
|
||||
|
||||
|
|
|
@ -509,7 +509,7 @@ void factor_vm::garbage_collection(cell gen,bool growing_data_heap_,cell request
|
|||
growing_data_heap = true;
|
||||
|
||||
/* see the comment in unmark_marked() */
|
||||
unmark_marked(&code);
|
||||
code->unmark_marked();
|
||||
}
|
||||
/* we try collecting aging space twice before going on to
|
||||
collect tenured */
|
||||
|
@ -546,7 +546,7 @@ void factor_vm::garbage_collection(cell gen,bool growing_data_heap_,cell request
|
|||
code_heap_scans++;
|
||||
|
||||
if(collecting_gen == data->tenured())
|
||||
free_unmarked(&code,(heap_iterator)factor::update_literal_and_word_references);
|
||||
code->free_unmarked((heap_iterator)factor::update_literal_and_word_references);
|
||||
else
|
||||
copy_code_heap_roots();
|
||||
|
||||
|
|
|
@ -297,7 +297,7 @@ void factor_vm::dump_code_heap()
|
|||
{
|
||||
cell reloc_size = 0, literal_size = 0;
|
||||
|
||||
heap_block *scan = first_block(&code);
|
||||
heap_block *scan = code->first_block();
|
||||
|
||||
while(scan)
|
||||
{
|
||||
|
@ -326,7 +326,7 @@ void factor_vm::dump_code_heap()
|
|||
print_cell_hex(scan->size); print_string(" ");
|
||||
print_string(status); print_string("\n");
|
||||
|
||||
scan = next_block(&code,scan);
|
||||
scan = code->next_block(scan);
|
||||
}
|
||||
|
||||
print_cell(reloc_size); print_string(" bytes of relocation data\n");
|
||||
|
|
|
@ -1,37 +1,36 @@
|
|||
#include "master.hpp"
|
||||
|
||||
/* This malloc-style heap code is reasonably generic. Maybe in the future, it
|
||||
will be used for the data heap too, if we ever get mark/sweep/compact GC. */
|
||||
|
||||
namespace factor
|
||||
{
|
||||
|
||||
void factor_vm::clear_free_list(heap *heap)
|
||||
void heap::clear_free_list()
|
||||
{
|
||||
memset(&heap->free,0,sizeof(heap_free_list));
|
||||
memset(&free,0,sizeof(heap_free_list));
|
||||
}
|
||||
|
||||
/* This malloc-style heap code is reasonably generic. Maybe in the future, it
|
||||
will be used for the data heap too, if we ever get incremental
|
||||
mark/sweep/compact GC. */
|
||||
void factor_vm::new_heap(heap *heap, cell size)
|
||||
heap::heap(factor_vm *myvm_, cell size)
|
||||
{
|
||||
heap->seg = alloc_segment(align_page(size));
|
||||
if(!heap->seg)
|
||||
fatal_error("Out of memory in new_heap",size);
|
||||
|
||||
clear_free_list(heap);
|
||||
myvm = myvm_;
|
||||
seg = myvm->alloc_segment(myvm->align_page(size));
|
||||
if(!seg) fatal_error("Out of memory in new_heap",size);
|
||||
clear_free_list();
|
||||
}
|
||||
|
||||
void factor_vm::add_to_free_list(heap *heap, free_heap_block *block)
|
||||
void heap::add_to_free_list(free_heap_block *block)
|
||||
{
|
||||
if(block->size < free_list_count * block_size_increment)
|
||||
{
|
||||
int index = block->size / block_size_increment;
|
||||
block->next_free = heap->free.small_blocks[index];
|
||||
heap->free.small_blocks[index] = block;
|
||||
block->next_free = free.small_blocks[index];
|
||||
free.small_blocks[index] = block;
|
||||
}
|
||||
else
|
||||
{
|
||||
block->next_free = heap->free.large_blocks;
|
||||
heap->free.large_blocks = block;
|
||||
block->next_free = free.large_blocks;
|
||||
free.large_blocks = block;
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -39,16 +38,16 @@ void factor_vm::add_to_free_list(heap *heap, free_heap_block *block)
|
|||
|
||||
In the former case, we must add a large free block from compiling.base + size to
|
||||
compiling.limit. */
|
||||
void factor_vm::build_free_list(heap *heap, cell size)
|
||||
void heap::build_free_list(cell size)
|
||||
{
|
||||
heap_block *prev = NULL;
|
||||
|
||||
clear_free_list(heap);
|
||||
clear_free_list();
|
||||
|
||||
size = (size + block_size_increment - 1) & ~(block_size_increment - 1);
|
||||
|
||||
heap_block *scan = first_block(heap);
|
||||
free_heap_block *end = (free_heap_block *)(heap->seg->start + size);
|
||||
heap_block *scan = first_block();
|
||||
free_heap_block *end = (free_heap_block *)(seg->start + size);
|
||||
|
||||
/* Add all free blocks to the free list */
|
||||
while(scan && scan < (heap_block *)end)
|
||||
|
@ -56,28 +55,28 @@ void factor_vm::build_free_list(heap *heap, cell size)
|
|||
switch(scan->status)
|
||||
{
|
||||
case B_FREE:
|
||||
add_to_free_list(heap,(free_heap_block *)scan);
|
||||
add_to_free_list((free_heap_block *)scan);
|
||||
break;
|
||||
case B_ALLOCATED:
|
||||
break;
|
||||
default:
|
||||
critical_error("Invalid scan->status",(cell)scan);
|
||||
myvm->critical_error("Invalid scan->status",(cell)scan);
|
||||
break;
|
||||
}
|
||||
|
||||
prev = scan;
|
||||
scan = next_block(heap,scan);
|
||||
scan = next_block(scan);
|
||||
}
|
||||
|
||||
/* If there is room at the end of the heap, add a free block. This
|
||||
branch is only taken after loading a new image, not after code GC */
|
||||
if((cell)(end + 1) <= heap->seg->end)
|
||||
if((cell)(end + 1) <= seg->end)
|
||||
{
|
||||
end->status = B_FREE;
|
||||
end->size = heap->seg->end - (cell)end;
|
||||
end->size = seg->end - (cell)end;
|
||||
|
||||
/* add final free block */
|
||||
add_to_free_list(heap,end);
|
||||
add_to_free_list(end);
|
||||
}
|
||||
/* This branch is taken if the newly loaded image fits exactly, or
|
||||
after code GC */
|
||||
|
@ -86,30 +85,30 @@ void factor_vm::build_free_list(heap *heap, cell size)
|
|||
/* even if there's no room at the end of the heap for a new
|
||||
free block, we might have to jigger it up by a few bytes in
|
||||
case prev + prev->size */
|
||||
if(prev) prev->size = heap->seg->end - (cell)prev;
|
||||
if(prev) prev->size = seg->end - (cell)prev;
|
||||
}
|
||||
|
||||
}
|
||||
|
||||
void factor_vm::assert_free_block(free_heap_block *block)
|
||||
void heap::assert_free_block(free_heap_block *block)
|
||||
{
|
||||
if(block->status != B_FREE)
|
||||
critical_error("Invalid block in free list",(cell)block);
|
||||
myvm->critical_error("Invalid block in free list",(cell)block);
|
||||
}
|
||||
|
||||
|
||||
free_heap_block *factor_vm::find_free_block(heap *heap, cell size)
|
||||
free_heap_block *heap::find_free_block(cell size)
|
||||
{
|
||||
cell attempt = size;
|
||||
|
||||
while(attempt < free_list_count * block_size_increment)
|
||||
{
|
||||
int index = attempt / block_size_increment;
|
||||
free_heap_block *block = heap->free.small_blocks[index];
|
||||
free_heap_block *block = free.small_blocks[index];
|
||||
if(block)
|
||||
{
|
||||
assert_free_block(block);
|
||||
heap->free.small_blocks[index] = block->next_free;
|
||||
free.small_blocks[index] = block->next_free;
|
||||
return block;
|
||||
}
|
||||
|
||||
|
@ -117,7 +116,7 @@ free_heap_block *factor_vm::find_free_block(heap *heap, cell size)
|
|||
}
|
||||
|
||||
free_heap_block *prev = NULL;
|
||||
free_heap_block *block = heap->free.large_blocks;
|
||||
free_heap_block *block = free.large_blocks;
|
||||
|
||||
while(block)
|
||||
{
|
||||
|
@ -127,7 +126,7 @@ free_heap_block *factor_vm::find_free_block(heap *heap, cell size)
|
|||
if(prev)
|
||||
prev->next_free = block->next_free;
|
||||
else
|
||||
heap->free.large_blocks = block->next_free;
|
||||
free.large_blocks = block->next_free;
|
||||
return block;
|
||||
}
|
||||
|
||||
|
@ -138,7 +137,7 @@ free_heap_block *factor_vm::find_free_block(heap *heap, cell size)
|
|||
return NULL;
|
||||
}
|
||||
|
||||
free_heap_block *factor_vm::split_free_block(heap *heap, free_heap_block *block, cell size)
|
||||
free_heap_block *heap::split_free_block(free_heap_block *block, cell size)
|
||||
{
|
||||
if(block->size != size )
|
||||
{
|
||||
|
@ -148,21 +147,21 @@ free_heap_block *factor_vm::split_free_block(heap *heap, free_heap_block *block,
|
|||
split->size = block->size - size;
|
||||
split->next_free = block->next_free;
|
||||
block->size = size;
|
||||
add_to_free_list(heap,split);
|
||||
add_to_free_list(split);
|
||||
}
|
||||
|
||||
return block;
|
||||
}
|
||||
|
||||
/* Allocate a block of memory from the mark and sweep GC heap */
|
||||
heap_block *factor_vm::heap_allot(heap *heap, cell size)
|
||||
heap_block *heap::heap_allot(cell size)
|
||||
{
|
||||
size = (size + block_size_increment - 1) & ~(block_size_increment - 1);
|
||||
|
||||
free_heap_block *block = find_free_block(heap,size);
|
||||
free_heap_block *block = find_free_block(size);
|
||||
if(block)
|
||||
{
|
||||
block = split_free_block(heap,block,size);
|
||||
block = split_free_block(block,size);
|
||||
|
||||
block->status = B_ALLOCATED;
|
||||
return block;
|
||||
|
@ -172,13 +171,13 @@ heap_block *factor_vm::heap_allot(heap *heap, cell size)
|
|||
}
|
||||
|
||||
/* Deallocates a block manually */
|
||||
void factor_vm::heap_free(heap *heap, heap_block *block)
|
||||
void heap::heap_free(heap_block *block)
|
||||
{
|
||||
block->status = B_FREE;
|
||||
add_to_free_list(heap,(free_heap_block *)block);
|
||||
add_to_free_list((free_heap_block *)block);
|
||||
}
|
||||
|
||||
void factor_vm::mark_block(heap_block *block)
|
||||
void heap::mark_block(heap_block *block)
|
||||
{
|
||||
/* If already marked, do nothing */
|
||||
switch(block->status)
|
||||
|
@ -189,41 +188,41 @@ void factor_vm::mark_block(heap_block *block)
|
|||
block->status = B_MARKED;
|
||||
break;
|
||||
default:
|
||||
critical_error("Marking the wrong block",(cell)block);
|
||||
myvm->critical_error("Marking the wrong block",(cell)block);
|
||||
break;
|
||||
}
|
||||
}
|
||||
|
||||
/* If in the middle of code GC, we have to grow the heap, data GC restarts from
|
||||
scratch, so we have to unmark any marked blocks. */
|
||||
void factor_vm::unmark_marked(heap *heap)
|
||||
void heap::unmark_marked()
|
||||
{
|
||||
heap_block *scan = first_block(heap);
|
||||
heap_block *scan = first_block();
|
||||
|
||||
while(scan)
|
||||
{
|
||||
if(scan->status == B_MARKED)
|
||||
scan->status = B_ALLOCATED;
|
||||
|
||||
scan = next_block(heap,scan);
|
||||
scan = next_block(scan);
|
||||
}
|
||||
}
|
||||
|
||||
/* After code GC, all referenced code blocks have status set to B_MARKED, so any
|
||||
which are allocated and not marked can be reclaimed. */
|
||||
void factor_vm::free_unmarked(heap *heap, heap_iterator iter)
|
||||
void heap::free_unmarked(heap_iterator iter)
|
||||
{
|
||||
clear_free_list(heap);
|
||||
clear_free_list();
|
||||
|
||||
heap_block *prev = NULL;
|
||||
heap_block *scan = first_block(heap);
|
||||
heap_block *scan = first_block();
|
||||
|
||||
while(scan)
|
||||
{
|
||||
switch(scan->status)
|
||||
{
|
||||
case B_ALLOCATED:
|
||||
if(secure_gc)
|
||||
if(myvm->secure_gc)
|
||||
memset(scan + 1,0,scan->size - sizeof(heap_block));
|
||||
|
||||
if(prev && prev->status == B_FREE)
|
||||
|
@ -242,30 +241,30 @@ void factor_vm::free_unmarked(heap *heap, heap_iterator iter)
|
|||
break;
|
||||
case B_MARKED:
|
||||
if(prev && prev->status == B_FREE)
|
||||
add_to_free_list(heap,(free_heap_block *)prev);
|
||||
add_to_free_list((free_heap_block *)prev);
|
||||
scan->status = B_ALLOCATED;
|
||||
prev = scan;
|
||||
iter(scan,this);
|
||||
iter(scan,myvm);
|
||||
break;
|
||||
default:
|
||||
critical_error("Invalid scan->status",(cell)scan);
|
||||
myvm->critical_error("Invalid scan->status",(cell)scan);
|
||||
}
|
||||
|
||||
scan = next_block(heap,scan);
|
||||
scan = next_block(scan);
|
||||
}
|
||||
|
||||
if(prev && prev->status == B_FREE)
|
||||
add_to_free_list(heap,(free_heap_block *)prev);
|
||||
add_to_free_list((free_heap_block *)prev);
|
||||
}
|
||||
|
||||
/* Compute total sum of sizes of free blocks, and size of largest free block */
|
||||
void factor_vm::heap_usage(heap *heap, cell *used, cell *total_free, cell *max_free)
|
||||
void heap::heap_usage(cell *used, cell *total_free, cell *max_free)
|
||||
{
|
||||
*used = 0;
|
||||
*total_free = 0;
|
||||
*max_free = 0;
|
||||
|
||||
heap_block *scan = first_block(heap);
|
||||
heap_block *scan = first_block();
|
||||
|
||||
while(scan)
|
||||
{
|
||||
|
@ -280,34 +279,34 @@ void factor_vm::heap_usage(heap *heap, cell *used, cell *total_free, cell *max_f
|
|||
*max_free = scan->size;
|
||||
break;
|
||||
default:
|
||||
critical_error("Invalid scan->status",(cell)scan);
|
||||
myvm->critical_error("Invalid scan->status",(cell)scan);
|
||||
}
|
||||
|
||||
scan = next_block(heap,scan);
|
||||
scan = next_block(scan);
|
||||
}
|
||||
}
|
||||
|
||||
/* The size of the heap, not including the last block if it's free */
|
||||
cell factor_vm::heap_size(heap *heap)
|
||||
cell heap::heap_size()
|
||||
{
|
||||
heap_block *scan = first_block(heap);
|
||||
heap_block *scan = first_block();
|
||||
|
||||
while(next_block(heap,scan) != NULL)
|
||||
scan = next_block(heap,scan);
|
||||
while(next_block(scan) != NULL)
|
||||
scan = next_block(scan);
|
||||
|
||||
/* this is the last block in the heap, and it is free */
|
||||
if(scan->status == B_FREE)
|
||||
return (cell)scan - heap->seg->start;
|
||||
return (cell)scan - seg->start;
|
||||
/* otherwise the last block is allocated */
|
||||
else
|
||||
return heap->seg->size;
|
||||
return seg->size;
|
||||
}
|
||||
|
||||
/* Compute where each block is going to go, after compaction */
|
||||
cell factor_vm::compute_heap_forwarding(heap *heap, unordered_map<heap_block *,char *> &forwarding)
|
||||
cell heap::compute_heap_forwarding(unordered_map<heap_block *,char *> &forwarding)
|
||||
{
|
||||
heap_block *scan = first_block(heap);
|
||||
char *address = (char *)first_block(heap);
|
||||
heap_block *scan = first_block();
|
||||
char *address = (char *)first_block();
|
||||
|
||||
while(scan)
|
||||
{
|
||||
|
@ -317,21 +316,21 @@ cell factor_vm::compute_heap_forwarding(heap *heap, unordered_map<heap_block *,c
|
|||
address += scan->size;
|
||||
}
|
||||
else if(scan->status == B_MARKED)
|
||||
critical_error("Why is the block marked?",0);
|
||||
myvm->critical_error("Why is the block marked?",0);
|
||||
|
||||
scan = next_block(heap,scan);
|
||||
scan = next_block(scan);
|
||||
}
|
||||
|
||||
return (cell)address - heap->seg->start;
|
||||
return (cell)address - seg->start;
|
||||
}
|
||||
|
||||
void factor_vm::compact_heap(heap *heap, unordered_map<heap_block *,char *> &forwarding)
|
||||
void heap::compact_heap(unordered_map<heap_block *,char *> &forwarding)
|
||||
{
|
||||
heap_block *scan = first_block(heap);
|
||||
heap_block *scan = first_block();
|
||||
|
||||
while(scan)
|
||||
{
|
||||
heap_block *next = next_block(heap,scan);
|
||||
heap_block *next = next_block(scan);
|
||||
|
||||
if(scan->status == B_ALLOCATED)
|
||||
memmove(forwarding[scan],scan,scan->size);
|
|
@ -0,0 +1,59 @@
|
|||
namespace factor
|
||||
{
|
||||
|
||||
static const cell free_list_count = 16;
|
||||
static const cell block_size_increment = 32;
|
||||
|
||||
struct heap_free_list {
|
||||
free_heap_block *small_blocks[free_list_count];
|
||||
free_heap_block *large_blocks;
|
||||
};
|
||||
|
||||
typedef void (*heap_iterator)(heap_block *compiled, factor_vm *vm);
|
||||
|
||||
struct heap {
|
||||
factor_vm *myvm;
|
||||
segment *seg;
|
||||
heap_free_list free;
|
||||
|
||||
heap(factor_vm *myvm, cell size);
|
||||
|
||||
inline heap_block *next_block(heap_block *block)
|
||||
{
|
||||
cell next = ((cell)block + block->size);
|
||||
if(next == seg->end)
|
||||
return NULL;
|
||||
else
|
||||
return (heap_block *)next;
|
||||
}
|
||||
|
||||
inline heap_block *first_block()
|
||||
{
|
||||
return (heap_block *)seg->start;
|
||||
}
|
||||
|
||||
inline heap_block *last_block()
|
||||
{
|
||||
return (heap_block *)seg->end;
|
||||
}
|
||||
|
||||
void clear_free_list();
|
||||
void new_heap(cell size);
|
||||
void add_to_free_list(free_heap_block *block);
|
||||
void build_free_list(cell size);
|
||||
void assert_free_block(free_heap_block *block);
|
||||
free_heap_block *find_free_block(cell size);
|
||||
free_heap_block *split_free_block(free_heap_block *block, cell size);
|
||||
heap_block *heap_allot(cell size);
|
||||
void heap_free(heap_block *block);
|
||||
void mark_block(heap_block *block);
|
||||
void unmark_marked();
|
||||
void free_unmarked(heap_iterator iter);
|
||||
void heap_usage(cell *used, cell *total_free, cell *max_free);
|
||||
cell heap_size();
|
||||
cell compute_heap_forwarding(unordered_map<heap_block *,char *> &forwarding);
|
||||
void compact_heap(unordered_map<heap_block *,char *> &forwarding);
|
||||
|
||||
};
|
||||
|
||||
}
|
12
vm/image.cpp
12
vm/image.cpp
|
@ -56,7 +56,7 @@ void factor_vm::load_code_heap(FILE *file, image_header *h, vm_parameters *p)
|
|||
|
||||
if(h->code_size != 0)
|
||||
{
|
||||
size_t bytes_read = fread(first_block(&code),1,h->code_size,file);
|
||||
size_t bytes_read = fread(code->first_block(),1,h->code_size,file);
|
||||
if(bytes_read != h->code_size)
|
||||
{
|
||||
print_string("truncated image: ");
|
||||
|
@ -69,7 +69,7 @@ void factor_vm::load_code_heap(FILE *file, image_header *h, vm_parameters *p)
|
|||
}
|
||||
|
||||
code_relocation_base = h->code_relocation_base;
|
||||
build_free_list(&code,h->code_size);
|
||||
code->build_free_list(h->code_size);
|
||||
}
|
||||
|
||||
/* Save the current image to disk */
|
||||
|
@ -92,8 +92,8 @@ bool factor_vm::save_image(const vm_char *filename)
|
|||
h.version = image_version;
|
||||
h.data_relocation_base = tenured->start;
|
||||
h.data_size = tenured->here - tenured->start;
|
||||
h.code_relocation_base = code.seg->start;
|
||||
h.code_size = heap_size(&code);
|
||||
h.code_relocation_base = code->seg->start;
|
||||
h.code_size = code->heap_size();
|
||||
|
||||
h.t = T;
|
||||
h.bignum_zero = bignum_zero;
|
||||
|
@ -107,7 +107,7 @@ bool factor_vm::save_image(const vm_char *filename)
|
|||
|
||||
if(fwrite(&h,sizeof(image_header),1,file) != 1) ok = false;
|
||||
if(fwrite((void*)tenured->start,h.data_size,1,file) != 1) ok = false;
|
||||
if(fwrite(first_block(&code),h.code_size,1,file) != 1) ok = false;
|
||||
if(fwrite(code->first_block(),h.code_size,1,file) != 1) ok = false;
|
||||
if(fclose(file)) ok = false;
|
||||
|
||||
if(!ok)
|
||||
|
@ -175,7 +175,7 @@ void data_fixup(cell *cell, factor_vm *myvm)
|
|||
template <typename TYPE> void factor_vm::code_fixup(TYPE **handle)
|
||||
{
|
||||
TYPE *ptr = *handle;
|
||||
TYPE *new_ptr = (TYPE *)(((cell)ptr) + (code.seg->start - code_relocation_base));
|
||||
TYPE *new_ptr = (TYPE *)(((cell)ptr) + (code->seg->start - code_relocation_base));
|
||||
*handle = new_ptr;
|
||||
}
|
||||
|
||||
|
|
|
@ -24,7 +24,7 @@ void factor_vm::deallocate_inline_cache(cell return_address)
|
|||
#endif
|
||||
|
||||
if(old_type == PIC_TYPE)
|
||||
heap_free(&code,old_block);
|
||||
code->heap_free(old_block);
|
||||
}
|
||||
|
||||
/* Figure out what kind of type check the PIC needs based on the methods
|
||||
|
|
|
@ -64,7 +64,7 @@
|
|||
#include "math.hpp"
|
||||
#include "float_bits.hpp"
|
||||
#include "io.hpp"
|
||||
#include "code_gc.hpp"
|
||||
#include "heap.hpp"
|
||||
#include "code_heap.hpp"
|
||||
#include "image.hpp"
|
||||
#include "callstack.hpp"
|
||||
|
|
|
@ -83,7 +83,7 @@ struct factor_vm_data {
|
|||
cell bignum_neg_one;
|
||||
|
||||
//code_heap
|
||||
heap code;
|
||||
heap *code;
|
||||
unordered_map<heap_block *, char *> forwarding;
|
||||
|
||||
//image
|
||||
|
|
18
vm/vm.hpp
18
vm/vm.hpp
|
@ -381,24 +381,6 @@ struct factor_vm : factor_vm_data {
|
|||
inline void primitive_fflush();
|
||||
inline void primitive_fclose();
|
||||
|
||||
//code_gc
|
||||
void clear_free_list(heap *heap);
|
||||
void new_heap(heap *heap, cell size);
|
||||
void add_to_free_list(heap *heap, free_heap_block *block);
|
||||
void build_free_list(heap *heap, cell size);
|
||||
void assert_free_block(free_heap_block *block);
|
||||
free_heap_block *find_free_block(heap *heap, cell size);
|
||||
free_heap_block *split_free_block(heap *heap, free_heap_block *block, cell size);
|
||||
heap_block *heap_allot(heap *heap, cell size);
|
||||
void heap_free(heap *heap, heap_block *block);
|
||||
void mark_block(heap_block *block);
|
||||
void unmark_marked(heap *heap);
|
||||
void free_unmarked(heap *heap, heap_iterator iter);
|
||||
void heap_usage(heap *heap, cell *used, cell *total_free, cell *max_free);
|
||||
cell heap_size(heap *heap);
|
||||
cell compute_heap_forwarding(heap *heap, unordered_map<heap_block *,char *> &forwarding);
|
||||
void compact_heap(heap *heap, unordered_map<heap_block *,char *> &forwarding);
|
||||
|
||||
//code_block
|
||||
relocation_type relocation_type_of(relocation_entry r);
|
||||
relocation_class relocation_class_of(relocation_entry r);
|
||||
|
|
Loading…
Reference in New Issue