diff options
author | Austin Seipp <austin@well-typed.com> | 2014-07-21 20:21:54 -0500 |
---|---|---|
committer | Austin Seipp <austin@well-typed.com> | 2014-07-28 09:29:15 -0500 |
commit | 4f5966b3edd62997552d2ebe371542861ed81405 (patch) | |
tree | b8b2c40d0d9ba7a7319439532c9e6bae0da23441 /rts/Arena.c | |
parent | 9e8d2580f0a7498664ded1ff7cdae6bca07b8dda (diff) | |
download | haskell-4f5966b3edd62997552d2ebe371542861ed81405.tar.gz |
rts: delint/detab/dewhitespace Arena.c
Signed-off-by: Austin Seipp <austin@well-typed.com>
Diffstat (limited to 'rts/Arena.c')
-rw-r--r-- | rts/Arena.c | 51 |
1 files changed, 25 insertions, 26 deletions
diff --git a/rts/Arena.c b/rts/Arena.c index 361c6c41be..813e60ebf6 100644 --- a/rts/Arena.c +++ b/rts/Arena.c @@ -8,7 +8,7 @@ Do not assume that sequentially allocated objects will be adjacent in memory. - + Quirks: this allocator makes use of the RTS block allocator. If the current block doesn't have enough room for the requested object, then a new block is allocated. This means that allocating @@ -27,11 +27,11 @@ // Each arena struct is allocated using malloc(). struct _Arena { bdescr *current; - StgWord *free; // ptr to next free byte in current block - StgWord *lim; // limit (== last free byte + 1) + StgWord *free; // ptr to next free byte in current block + StgWord *lim; // limit (== last free byte + 1) }; -// We like to keep track of how many blocks we've allocated for +// We like to keep track of how many blocks we've allocated for // Storage.c:memInventory(). static long arena_blocks = 0; @@ -74,26 +74,26 @@ arenaAlloc( Arena *arena, size_t size ) size_w = B_TO_W(size); if ( arena->free + size_w < arena->lim ) { - // enough room in the current block... - p = arena->free; - arena->free += size_w; - return p; + // enough room in the current block... + p = arena->free; + arena->free += size_w; + return p; } else { - // allocate a fresh block... - req_blocks = (W_)BLOCK_ROUND_UP(size) / BLOCK_SIZE; - bd = allocGroup_lock(req_blocks); - arena_blocks += req_blocks; + // allocate a fresh block... + req_blocks = (W_)BLOCK_ROUND_UP(size) / BLOCK_SIZE; + bd = allocGroup_lock(req_blocks); + arena_blocks += req_blocks; - bd->gen_no = 0; - bd->gen = NULL; + bd->gen_no = 0; + bd->gen = NULL; bd->dest_no = 0; - bd->flags = 0; - bd->free = bd->start; - bd->link = arena->current; - arena->current = bd; - arena->free = bd->free + size_w; - arena->lim = bd->free + bd->blocks * BLOCK_SIZE_W; - return bd->start; + bd->flags = 0; + bd->free = bd->start; + bd->link = arena->current; + arena->current = bd; + arena->free = bd->free + size_w; + arena->lim = bd->free + bd->blocks * BLOCK_SIZE_W; + return bd->start; } } @@ -104,10 +104,10 @@ arenaFree( Arena *arena ) bdescr *bd, *next; for (bd = arena->current; bd != NULL; bd = next) { - next = bd->link; - arena_blocks -= bd->blocks; - ASSERT(arena_blocks >= 0); - freeGroup_lock(bd); + next = bd->link; + arena_blocks -= bd->blocks; + ASSERT(arena_blocks >= 0); + freeGroup_lock(bd); } stgFree(arena); } @@ -117,4 +117,3 @@ arenaBlocks( void ) { return arena_blocks; } - |