summaryrefslogtreecommitdiff
path: root/js/src/nanojit/CodeAlloc.cpp
diff options
context:
space:
mode:
Diffstat (limited to 'js/src/nanojit/CodeAlloc.cpp')
-rw-r--r--js/src/nanojit/CodeAlloc.cpp570
1 files changed, 570 insertions, 0 deletions
diff --git a/js/src/nanojit/CodeAlloc.cpp b/js/src/nanojit/CodeAlloc.cpp
new file mode 100644
index 0000000..61b711a
--- /dev/null
+++ b/js/src/nanojit/CodeAlloc.cpp
@@ -0,0 +1,570 @@
+/* -*- Mode: C++; c-basic-offset: 4; indent-tabs-mode: nil; tab-width: 4 -*- */
+/* vi: set ts=4 sw=4 expandtab: (add to ~/.vimrc: set modeline modelines=5) */
+/* ***** BEGIN LICENSE BLOCK *****
+ * Version: MPL 1.1/GPL 2.0/LGPL 2.1
+ *
+ * The contents of this file are subject to the Mozilla Public License Version
+ * 1.1 (the "License"); you may not use this file except in compliance with
+ * the License. You may obtain a copy of the License at
+ * http://www.mozilla.org/MPL/
+ *
+ * Software distributed under the License is distributed on an "AS IS" basis,
+ * WITHOUT WARRANTY OF ANY KIND, either express or implied. See the License
+ * for the specific language governing rights and limitations under the
+ * License.
+ *
+ * The Original Code is [Open Source Virtual Machine].
+ *
+ * The Initial Developer of the Original Code is
+ * Adobe System Incorporated.
+ * Portions created by the Initial Developer are Copyright (C) 2004-2007
+ * the Initial Developer. All Rights Reserved.
+ *
+ * Contributor(s):
+ * Adobe AS3 Team
+ *
+ * Alternatively, the contents of this file may be used under the terms of
+ * either the GNU General Public License Version 2 or later (the "GPL"), or
+ * the GNU Lesser General Public License Version 2.1 or later (the "LGPL"),
+ * in which case the provisions of the GPL or the LGPL are applicable instead
+ * of those above. If you wish to allow use of your version of this file only
+ * under the terms of either the GPL or the LGPL, and not to allow others to
+ * use your version of this file under the terms of the MPL, indicate your
+ * decision by deleting the provisions above and replace them with the notice
+ * and other provisions required by the GPL or the LGPL. If you do not delete
+ * the provisions above, a recipient may use your version of this file under
+ * the terms of any one of the MPL, the GPL or the LGPL.
+ *
+ * ***** END LICENSE BLOCK ***** */
+
+#include "nanojit.h"
+
+//#define DOPROF
+#include "../vprof/vprof.h"
+
+#ifdef FEATURE_NANOJIT
+
+namespace nanojit
+{
+ static const bool verbose = false;
+#ifdef VMCFG_VTUNE
+ // vtune jit profiling api can't handle non-contiguous methods,
+ // so make the allocation size huge to avoid non-contiguous methods
+ static const int pagesPerAlloc = 128; // 1MB
+#elif defined(NANOJIT_ARM)
+ // ARM requires single-page allocations, due to the constant pool that
+ // lives on each page that must be reachable by a 4kb pcrel load.
+ static const int pagesPerAlloc = 1;
+#else
+ static const int pagesPerAlloc = 16;
+#endif
+
+ CodeAlloc::CodeAlloc()
+ : heapblocks(0)
+ , availblocks(0)
+ , totalAllocated(0)
+ , bytesPerPage(VMPI_getVMPageSize())
+ , bytesPerAlloc(pagesPerAlloc * bytesPerPage)
+ {
+ }
+
+ CodeAlloc::~CodeAlloc() {
+ reset();
+ }
+
+ void CodeAlloc::reset() {
+ // give all memory back to gcheap. Assumption is that all
+ // code is done being used by now.
+ for (CodeList* hb = heapblocks; hb != 0; ) {
+ _nvprof("free page",1);
+ CodeList* next = hb->next;
+ CodeList* fb = firstBlock(hb);
+ markBlockWrite(fb);
+ freeCodeChunk(fb, bytesPerAlloc);
+ totalAllocated -= bytesPerAlloc;
+ hb = next;
+ }
+ NanoAssert(!totalAllocated);
+ heapblocks = availblocks = 0;
+ }
+
+ CodeList* CodeAlloc::firstBlock(CodeList* term) {
+ // use uintptr_t, rather than char*, to avoid "increases required alignment" warning
+ uintptr_t end = (uintptr_t)alignUp(term, bytesPerPage);
+ return (CodeList*) (end - (uintptr_t)bytesPerAlloc);
+ }
+
+ static int round(size_t x) {
+ return (int)((x + 512) >> 10);
+ }
+
+ void CodeAlloc::logStats() {
+ size_t total = 0;
+ size_t frag_size = 0;
+ size_t free_size = 0;
+ int free_count = 0;
+ for (CodeList* hb = heapblocks; hb != 0; hb = hb->next) {
+ total += bytesPerAlloc;
+ for (CodeList* b = hb->lower; b != 0; b = b->lower) {
+ if (b->isFree) {
+ free_count++;
+ free_size += b->blockSize();
+ if (b->size() < minAllocSize)
+ frag_size += b->blockSize();
+ }
+ }
+ }
+ avmplus::AvmLog("code-heap: %dk free %dk fragmented %d\n",
+ round(total), round(free_size), frag_size);
+ }
+
+ inline void CodeAlloc::markBlockWrite(CodeList* b) {
+ NanoAssert(b->terminator != NULL);
+ CodeList* term = b->terminator;
+ if (term->isExec) {
+ markCodeChunkWrite(firstBlock(term), bytesPerAlloc);
+ term->isExec = false;
+ }
+ }
+
+ void CodeAlloc::alloc(NIns* &start, NIns* &end, size_t byteLimit) {
+ if (!availblocks) {
+ // no free mem, get more
+ addMem();
+ }
+
+ // grab a block
+ NanoAssert(!byteLimit || byteLimit > blkSpaceFor(2)); // if a limit is imposed it must be bigger than 2x minimum block size (see below)
+ markBlockWrite(availblocks);
+ CodeList* b = removeBlock(availblocks);
+
+ // limit imposed (byteLimit > 0) and the block is too big? then break it apart
+ if (byteLimit > 0 && b->size() > byteLimit) {
+
+ size_t consume; // # bytes to extract from the free block
+
+ // enough space to carve out a perfectly sized blk? (leaving at least a full free blk)
+ if (b->size() >= byteLimit + headerSpaceFor(1) + blkSpaceFor(1)) {
+ // yes, then take exactly what we need
+ consume = byteLimit + headerSpaceFor(1);
+ } else {
+ // no, then we should only take the min amount
+ consume = blkSpaceFor(1);
+
+ // ... and since b->size() > byteLimit && byteLimit > blkSpaceFor(2)
+ NanoAssert( b->size() > blkSpaceFor(2) );
+ NanoAssert( b->size() - consume > blkSpaceFor(1) ); // thus, we know that at least 1 blk left.
+ }
+
+ // break block into 2 pieces, returning the lower portion to the free list
+ CodeList* higher = b->higher;
+ b->end = (NIns*) ( (uintptr_t)b->end - consume );
+ CodeList* b1 = b->higher;
+ higher->lower = b1;
+ b1->higher = higher;
+ b1->lower = b;
+ b1->terminator = b->terminator;
+ NanoAssert(b->size() > minAllocSize);
+ addBlock(availblocks, b); // put back the rest of the block
+ b = b1;
+ }
+ NanoAssert(b->size() >= minAllocSize);
+ b->next = 0; // not technically needed (except for debug builds), but good hygiene.
+ b->isFree = false;
+ start = b->start();
+ end = b->end;
+ if (verbose)
+ avmplus::AvmLog("CodeAlloc(%p).alloc %p-%p %d\n", this, start, end, int(end-start));
+ debug_only(sanity_check();)
+ }
+
+ void CodeAlloc::free(NIns* start, NIns *end) {
+ NanoAssert(heapblocks);
+ CodeList *blk = getBlock(start, end);
+ if (verbose)
+ avmplus::AvmLog("free %p-%p %d\n", start, end, (int)blk->size());
+
+ NanoAssert(!blk->isFree);
+
+ // coalesce adjacent blocks.
+ bool already_on_avail_list;
+
+ if (blk->lower && blk->lower->isFree) {
+ // combine blk into blk->lower (destroy blk)
+ CodeList* lower = blk->lower;
+ CodeList* higher = blk->higher;
+ already_on_avail_list = lower->size() >= minAllocSize;
+ lower->higher = higher;
+ higher->lower = lower;
+ blk = lower;
+ }
+ else
+ already_on_avail_list = false;
+
+ // the last block in each heapblock is a terminator block,
+ // which is never free, therefore blk->higher != null
+ if (blk->higher->isFree) {
+ CodeList *higher = blk->higher->higher;
+ CodeList *coalescedBlock = blk->higher;
+
+ if ( coalescedBlock->size() >= minAllocSize ) {
+ // Unlink coalescedBlock from the available block chain.
+ if ( availblocks == coalescedBlock ) {
+ removeBlock(availblocks);
+ }
+ else {
+ CodeList* free_block = availblocks;
+ while ( free_block && free_block->next != coalescedBlock) {
+ NanoAssert(free_block->size() >= minAllocSize);
+ NanoAssert(free_block->isFree);
+ NanoAssert(free_block->next);
+ free_block = free_block->next;
+ }
+ NanoAssert(free_block && free_block->next == coalescedBlock);
+ free_block->next = coalescedBlock->next;
+ }
+ }
+
+ // combine blk->higher into blk (destroy coalescedBlock)
+ blk->higher = higher;
+ higher->lower = blk;
+ }
+ blk->isFree = true;
+ NanoAssert(!blk->lower || !blk->lower->isFree);
+ NanoAssert(blk->higher && !blk->higher->isFree);
+ //memset(blk->start(), 0xCC, blk->size()); // INT 3 instruction
+ if ( !already_on_avail_list && blk->size() >= minAllocSize )
+ addBlock(availblocks, blk);
+
+ NanoAssert(heapblocks);
+ debug_only(sanity_check();)
+ }
+
+ void CodeAlloc::freeAll(CodeList* &code) {
+ while (code) {
+ CodeList *b = removeBlock(code);
+ free(b->start(), b->end);
+ }
+ }
+
+#if defined NANOJIT_ARM && defined UNDER_CE
+ // Use a single flush for the whole CodeList, when we have no
+ // finer-granularity flush support, as on WinCE.
+ void CodeAlloc::flushICache(CodeList* &/*blocks*/) {
+ FlushInstructionCache(GetCurrentProcess(), NULL, NULL);
+ }
+#else
+ void CodeAlloc::flushICache(CodeList* &blocks) {
+ for (CodeList *b = blocks; b != 0; b = b->next)
+ flushICache(b->start(), b->size());
+ }
+#endif
+
+#if defined(AVMPLUS_UNIX) && defined(NANOJIT_ARM)
+#include <asm/unistd.h>
+extern "C" void __clear_cache(char *BEG, char *END);
+#endif
+
+#if defined(AVMPLUS_UNIX) && defined(NANOJIT_MIPS)
+#include <asm/cachectl.h>
+extern "C" int cacheflush(char *addr, int nbytes, int cache);
+#endif
+
+#ifdef AVMPLUS_SPARC
+// Note: the linux #define provided by the compiler.
+#ifdef linux // bugzilla 502369
+void sync_instruction_memory(caddr_t v, u_int len)
+{
+ caddr_t end = v + len;
+ caddr_t p = v;
+ while (p < end) {
+ asm("flush %0" : : "r" (p));
+ p += 32;
+ }
+}
+#else
+extern "C" void sync_instruction_memory(caddr_t v, u_int len);
+#endif
+#endif
+
+#if defined NANOJIT_IA32 || defined NANOJIT_X64
+ // intel chips have dcache/icache interlock
+ void CodeAlloc::flushICache(void *start, size_t len) {
+ // Tell Valgrind that new code has been generated, and it must flush
+ // any translations it has for the memory range generated into.
+ (void)start;
+ (void)len;
+ VALGRIND_DISCARD_TRANSLATIONS(start, len);
+ }
+
+#elif defined NANOJIT_ARM && defined UNDER_CE
+ // On arm/winmo, just flush the whole icache. The
+ // WinCE docs indicate that this function actually ignores its
+ // 2nd and 3rd arguments, and wants them to be NULL.
+ void CodeAlloc::flushICache(void *, size_t) {
+ FlushInstructionCache(GetCurrentProcess(), NULL, NULL);
+ }
+
+#elif defined NANOJIT_ARM && defined DARWIN
+ void CodeAlloc::flushICache(void *, size_t) {
+ VMPI_debugBreak();
+ }
+
+#elif defined AVMPLUS_MAC && defined NANOJIT_PPC
+
+# ifdef NANOJIT_64BIT
+ extern "C" void sys_icache_invalidate(const void*, size_t len);
+ extern "C" void sys_dcache_flush(const void*, size_t len);
+
+ // mac 64bit requires 10.5 so use that api
+ void CodeAlloc::flushICache(void *start, size_t len) {
+ sys_dcache_flush(start, len);
+ sys_icache_invalidate(start, len);
+ }
+# else
+ // mac ppc 32 could be 10.0 or later
+ // uses MakeDataExecutable() from Carbon api, OSUtils.h
+ // see http://developer.apple.com/documentation/Carbon/Reference/Memory_Manag_nt_Utilities/Reference/reference.html#//apple_ref/c/func/MakeDataExecutable
+ void CodeAlloc::flushICache(void *start, size_t len) {
+ MakeDataExecutable(start, len);
+ }
+# endif
+
+#elif defined NANOJIT_ARM && defined VMCFG_SYMBIAN
+ void CodeAlloc::flushICache(void *ptr, size_t len) {
+ uint32_t start = (uint32_t)ptr;
+ uint32_t rangeEnd = start + len;
+ User::IMB_Range((TAny*)start, (TAny*)rangeEnd);
+ }
+
+#elif defined AVMPLUS_SPARC
+ // fixme: sync_instruction_memory is a solaris api, test for solaris not sparc
+ void CodeAlloc::flushICache(void *start, size_t len) {
+ sync_instruction_memory((char*)start, len);
+ }
+
+#elif defined NANOJIT_SH4
+#include <asm/cachectl.h> /* CACHEFLUSH_*, */
+#include <sys/syscall.h> /* __NR_cacheflush, */
+ void CodeAlloc::flushICache(void *start, size_t len) {
+ syscall(__NR_cacheflush, start, len, CACHEFLUSH_D_WB | CACHEFLUSH_I);
+ }
+
+#elif defined(AVMPLUS_UNIX) && defined(NANOJIT_MIPS)
+ void CodeAlloc::flushICache(void *start, size_t len) {
+ // FIXME Use synci on MIPS32R2
+ cacheflush((char *)start, len, BCACHE);
+ }
+
+#elif defined AVMPLUS_UNIX
+ #ifdef ANDROID
+ void CodeAlloc::flushICache(void *start, size_t len) {
+ cacheflush((int)start, (int)start + len, 0);
+ }
+ #else
+ // fixme: __clear_cache is a libgcc feature, test for libgcc or gcc
+ void CodeAlloc::flushICache(void *start, size_t len) {
+ __clear_cache((char*)start, (char*)start + len);
+ }
+ #endif
+#endif // AVMPLUS_MAC && NANOJIT_PPC
+
+ void CodeAlloc::addBlock(CodeList* &blocks, CodeList* b) {
+ NanoAssert(b->terminator != NULL); // should not be mucking with terminator blocks
+ b->next = blocks;
+ blocks = b;
+ }
+
+ void CodeAlloc::addMem() {
+ void *mem = allocCodeChunk(bytesPerAlloc); // allocations never fail
+ totalAllocated += bytesPerAlloc;
+ NanoAssert(mem != NULL); // see allocCodeChunk contract in CodeAlloc.h
+ _nvprof("alloc page", uintptr_t(mem)>>12);
+
+ CodeList* b = (CodeList*)mem;
+ b->lower = 0;
+ b->next = 0;
+ b->end = (NIns*) (uintptr_t(mem) + bytesPerAlloc - sizeofMinBlock);
+ b->isFree = true;
+
+ // create a tiny terminator block, add to fragmented list, this way
+ // all other blocks have a valid block at b->higher
+ CodeList* terminator = b->higher;
+ b->terminator = terminator;
+ terminator->lower = b;
+ terminator->end = 0; // this is how we identify the terminator
+ terminator->isFree = false;
+ terminator->isExec = false;
+ terminator->terminator = 0;
+ debug_only(sanity_check();)
+
+ // add terminator to heapblocks list so we can track whole blocks
+ terminator->next = heapblocks;
+ heapblocks = terminator;
+
+ addBlock(availblocks, b); // add to free list
+ }
+
+ CodeList* CodeAlloc::getBlock(NIns* start, NIns* end) {
+ CodeList* b = (CodeList*) (uintptr_t(start) - offsetof(CodeList, code));
+ NanoAssert(b->end == end && b->next == 0); (void) end;
+ return b;
+ }
+
+ CodeList* CodeAlloc::removeBlock(CodeList* &blocks) {
+ CodeList* b = blocks;
+ NanoAssert(b != NULL);
+ NanoAssert(b->terminator != NULL); // should not be mucking with terminator blocks
+ blocks = b->next;
+ b->next = 0;
+ return b;
+ }
+
+ void CodeAlloc::add(CodeList* &blocks, NIns* start, NIns* end) {
+ addBlock(blocks, getBlock(start, end));
+ }
+
+ /**
+ * split a block by freeing the hole in the middle defined by [holeStart,holeEnd),
+ * and adding the used prefix and suffix parts to the blocks CodeList.
+ */
+ void CodeAlloc::addRemainder(CodeList* &blocks, NIns* start, NIns* end, NIns* holeStart, NIns* holeEnd) {
+ NanoAssert(start < end && start <= holeStart && holeStart <= holeEnd && holeEnd <= end);
+ // shrink the hole by aligning holeStart forward and holeEnd backward
+ holeStart = (NIns*) ((uintptr_t(holeStart) + sizeof(NIns*)-1) & ~(sizeof(NIns*)-1));
+ holeEnd = (NIns*) (uintptr_t(holeEnd) & ~(sizeof(NIns*)-1));
+ // hole needs to be big enough for 2 headers + 1 block of free space (subtraction not used in check to avoid wraparound)
+ size_t minHole = headerSpaceFor(2) + blkSpaceFor(1);
+ if (uintptr_t(holeEnd) < minHole + uintptr_t(holeStart) ) {
+ // the hole is too small to make a new free block and a new used block. just keep
+ // the whole original block and don't free anything.
+ add(blocks, start, end);
+ } else if (holeStart == start && holeEnd == end) {
+ // totally empty block. free whole start-end range
+ this->free(start, end);
+ } else if (holeStart == start) {
+ // hole is lower-aligned with start, so just need one new block
+ // b1 b2
+ CodeList* b1 = getBlock(start, end);
+ CodeList* b2 = (CodeList*) (uintptr_t(holeEnd) - offsetof(CodeList, code));
+ b2->terminator = b1->terminator;
+ b2->isFree = false;
+ b2->next = 0;
+ b2->higher = b1->higher;
+ b2->lower = b1;
+ b2->higher->lower = b2;
+ b1->higher = b2;
+ debug_only(sanity_check();)
+ this->free(b1->start(), b1->end);
+ addBlock(blocks, b2);
+ } else if (holeEnd == end) {
+ // hole is right-aligned with end, just need one new block
+ // todo
+ NanoAssert(false);
+ } else {
+ // there's enough space left to split into three blocks (two new ones)
+ CodeList* b1 = getBlock(start, end);
+ CodeList* b2 = (CodeList*) (void*) holeStart;
+ CodeList* b3 = (CodeList*) (uintptr_t(holeEnd) - offsetof(CodeList, code));
+ b1->higher = b2;
+ b2->lower = b1;
+ b2->higher = b3;
+ b2->isFree = false; // redundant, since we're about to free, but good hygiene
+ b2->terminator = b1->terminator;
+ b3->lower = b2;
+ b3->end = end;
+ b3->isFree = false;
+ b3->higher->lower = b3;
+ b3->terminator = b1->terminator;
+ b2->next = 0;
+ b3->next = 0;
+ debug_only(sanity_check();)
+ this->free(b2->start(), b2->end);
+ addBlock(blocks, b3);
+ addBlock(blocks, b1);
+ }
+ }
+
+#ifdef PERFM
+ // This method is used only for profiling purposes.
+ // See CodegenLIR::emitMD() in Tamarin for an example.
+
+ size_t CodeAlloc::size(const CodeList* blocks) {
+ size_t size = 0;
+ for (const CodeList* b = blocks; b != 0; b = b->next)
+ size += int((uintptr_t)b->end - (uintptr_t)b);
+ return size;
+ }
+#endif
+
+ size_t CodeAlloc::size() {
+ return totalAllocated;
+ }
+
+ // check that all block neighbors are correct
+ #ifdef _DEBUG
+ void CodeAlloc::sanity_check() {
+ for (CodeList* hb = heapblocks; hb != 0; hb = hb->next) {
+ NanoAssert(hb->higher == 0);
+ for (CodeList* b = hb->lower; b != 0; b = b->lower) {
+ NanoAssert(b->higher->lower == b);
+ }
+ }
+ for (CodeList* avail = this->availblocks; avail; avail = avail->next) {
+ NanoAssert(avail->isFree && avail->size() >= minAllocSize);
+ }
+
+ #if CROSS_CHECK_FREE_LIST
+ for(CodeList* term = heapblocks; term; term = term->next) {
+ for(CodeList* hb = term->lower; hb; hb = hb->lower) {
+ if (hb->isFree && hb->size() >= minAllocSize) {
+ bool found_on_avail = false;
+ for (CodeList* avail = this->availblocks; !found_on_avail && avail; avail = avail->next) {
+ found_on_avail = avail == hb;
+ }
+
+ NanoAssert(found_on_avail);
+ }
+ }
+ }
+ for (CodeList* avail = this->availblocks; avail; avail = avail->next) {
+ bool found_in_heapblocks = false;
+ for(CodeList* term = heapblocks; !found_in_heapblocks && term; term = term->next) {
+ for(CodeList* hb = term->lower; !found_in_heapblocks && hb; hb = hb->lower) {
+ found_in_heapblocks = hb == avail;
+ }
+ }
+ NanoAssert(found_in_heapblocks);
+ }
+ #endif /* CROSS_CHECK_FREE_LIST */
+ }
+ #endif
+
+ // Loop through a list of blocks marking the chunks executable. If we encounter
+ // multiple blocks in the same chunk, only the first block will cause the
+ // chunk to become executable, the other calls will no-op (isExec flag checked)
+ void CodeAlloc::markExec(CodeList* &blocks) {
+ for (CodeList *b = blocks; b != 0; b = b->next) {
+ markChunkExec(b->terminator);
+ }
+ }
+
+ // Variant of markExec(CodeList*) that walks all heapblocks (i.e. chunks) marking
+ // each one executable. On systems where bytesPerAlloc is low (i.e. have lots
+ // of elements in the list) this can be expensive.
+ void CodeAlloc::markAllExec() {
+ for (CodeList* hb = heapblocks; hb != NULL; hb = hb->next) {
+ markChunkExec(hb);
+ }
+ }
+
+ // make an entire chunk executable
+ void CodeAlloc::markChunkExec(CodeList* term) {
+ NanoAssert(term->terminator == NULL);
+ if (!term->isExec) {
+ term->isExec = true;
+ markCodeChunkExec(firstBlock(term), bytesPerAlloc);
+ }
+ }
+}
+#endif // FEATURE_NANOJIT