diff options
author | Simon Marlow <marlowsd@gmail.com> | 2013-09-04 10:37:10 +0100 |
---|---|---|
committer | Simon Marlow <marlowsd@gmail.com> | 2013-09-04 11:00:32 +0100 |
commit | aa779e092c4f4d6a6691f3a4fc4074e6359337f8 (patch) | |
tree | f4c4e22da3aa71eff569b01af603836d7b5fd6a5 /rts/sm | |
parent | 5a3918febb7354e0900c4f04151599d833716032 (diff) | |
download | haskell-aa779e092c4f4d6a6691f3a4fc4074e6359337f8.tar.gz |
Don't move Capabilities in setNumCapabilities (#8209)
We have various problems with reallocating the array of Capabilities,
due to threads in waitForReturnCapability that are already holding a
pointer to a Capability.
Rather than add more locking to make this safer, I decided it would be
easier to ensure that we never move the Capabilities at all. The
capabilities array is now an array of pointers to Capabaility. There
are extra indirections, but it rarely matters - we don't often access
Capabilities via the array, normally we already have a pointer to
one. I ran the parallel benchmarks and didn't see any difference.
Diffstat (limited to 'rts/sm')
-rw-r--r-- | rts/sm/Compact.c | 2 | ||||
-rw-r--r-- | rts/sm/GC.c | 50 | ||||
-rw-r--r-- | rts/sm/Sanity.c | 14 | ||||
-rw-r--r-- | rts/sm/Storage.c | 10 |
4 files changed, 35 insertions, 41 deletions
diff --git a/rts/sm/Compact.c b/rts/sm/Compact.c index 247f1a01c6..375aeea2f2 100644 --- a/rts/sm/Compact.c +++ b/rts/sm/Compact.c @@ -933,7 +933,7 @@ compact(StgClosure *static_objects) bdescr *bd; StgPtr p; for (n = 0; n < n_capabilities; n++) { - for (bd = capabilities[n].mut_lists[g]; + for (bd = capabilities[n]->mut_lists[g]; bd != NULL; bd = bd->link) { for (p = bd->start; p < bd->free; p++) { thread((StgClosure **)p); diff --git a/rts/sm/GC.c b/rts/sm/GC.c index 8bbdda2266..6fd3b1b1c9 100644 --- a/rts/sm/GC.c +++ b/rts/sm/GC.c @@ -236,8 +236,8 @@ GarbageCollect (nat collect_gen, // attribute any costs to CCS_GC #ifdef PROFILING for (n = 0; n < n_capabilities; n++) { - save_CCS[n] = capabilities[n].r.rCCCS; - capabilities[n].r.rCCCS = CCS_GC; + save_CCS[n] = capabilities[n]->r.rCCCS; + capabilities[n]->r.rCCCS = CCS_GC; } #endif @@ -339,18 +339,18 @@ GarbageCollect (nat collect_gen, if (n_gc_threads == 1) { for (n = 0; n < n_capabilities; n++) { #if defined(THREADED_RTS) - scavenge_capability_mut_Lists1(&capabilities[n]); + scavenge_capability_mut_Lists1(capabilities[n]); #else - scavenge_capability_mut_lists(&capabilities[n]); + scavenge_capability_mut_lists(capabilities[n]); #endif } } else { scavenge_capability_mut_lists(gct->cap); for (n = 0; n < n_capabilities; n++) { if (gc_threads[n]->idle) { - markCapability(mark_root, gct, &capabilities[n], + markCapability(mark_root, gct, capabilities[n], rtsTrue/*don't mark sparks*/); - scavenge_capability_mut_lists(&capabilities[n]); + scavenge_capability_mut_lists(capabilities[n]); } } } @@ -363,7 +363,7 @@ GarbageCollect (nat collect_gen, gct->evac_gen_no = 0; if (n_gc_threads == 1) { for (n = 0; n < n_capabilities; n++) { - markCapability(mark_root, gct, &capabilities[n], + markCapability(mark_root, gct, capabilities[n], rtsTrue/*don't mark sparks*/); } } else { @@ -417,12 +417,12 @@ GarbageCollect (nat collect_gen, #ifdef THREADED_RTS if (n_gc_threads == 1) { for (n = 0; n < n_capabilities; n++) { - pruneSparkQueue(&capabilities[n]); + pruneSparkQueue(capabilities[n]); } } else { for (n = 0; n < n_capabilities; n++) { if (n == cap->no || gc_threads[n]->idle) { - pruneSparkQueue(&capabilities[n]); + pruneSparkQueue(capabilities[n]); } } } @@ -495,7 +495,7 @@ GarbageCollect (nat collect_gen, if (g > 0) { W_ mut_list_size = 0; for (n = 0; n < n_capabilities; n++) { - mut_list_size += countOccupied(capabilities[n].mut_lists[g]); + mut_list_size += countOccupied(capabilities[n]->mut_lists[g]); } copied += mut_list_size; @@ -646,14 +646,14 @@ GarbageCollect (nat collect_gen, // Reset the nursery: make the blocks empty if (DEBUG_IS_ON || n_gc_threads == 1) { for (n = 0; n < n_capabilities; n++) { - clearNursery(&capabilities[n]); + clearNursery(capabilities[n]); } } else { // When doing parallel GC, clearNursery() is called by the // worker threads for (n = 0; n < n_capabilities; n++) { if (gc_threads[n]->idle) { - clearNursery(&capabilities[n]); + clearNursery(capabilities[n]); } } } @@ -753,7 +753,7 @@ GarbageCollect (nat collect_gen, // restore enclosing cost centre #ifdef PROFILING for (n = 0; n < n_capabilities; n++) { - capabilities[n].r.rCCCS = save_CCS[n]; + capabilities[n]->r.rCCCS = save_CCS[n]; } #endif @@ -794,7 +794,7 @@ new_gc_thread (nat n, gc_thread *t) nat g; gen_workspace *ws; - t->cap = &capabilities[n]; + t->cap = capabilities[n]; #ifdef THREADED_RTS t->id = 0; @@ -866,12 +866,6 @@ initGcThreads (nat from USED_IF_THREADS, nat to USED_IF_THREADS) "initGcThreads"); } - // We have to update the gct->cap pointers to point to the new - // Capability array now. - for (i = 0; i < from; i++) { - gc_threads[i]->cap = &capabilities[gc_threads[i]->cap->no]; - } - for (i = from; i < to; i++) { gc_threads[i] = stgMallocBytes(sizeof(gc_thread) + @@ -1124,7 +1118,7 @@ waitForGcThreads (Capability *cap USED_IF_THREADS) for (i=0; i < n_threads; i++) { if (i == me || gc_threads[i]->idle) continue; if (gc_threads[i]->wakeup != GC_THREAD_STANDING_BY) { - prodCapability(&capabilities[i], cap->running_task); + prodCapability(capabilities[i], cap->running_task); } } for (j=0; j < 10; j++) { @@ -1132,7 +1126,7 @@ waitForGcThreads (Capability *cap USED_IF_THREADS) for (i=0; i < n_threads; i++) { if (i == me || gc_threads[i]->idle) continue; write_barrier(); - interruptCapability(&capabilities[i]); + interruptCapability(capabilities[i]); if (gc_threads[i]->wakeup != GC_THREAD_STANDING_BY) { retry = rtsTrue; } @@ -1228,8 +1222,8 @@ prepare_collected_gen (generation *gen) g = gen->no; if (g != 0) { for (i = 0; i < n_capabilities; i++) { - freeChain(capabilities[i].mut_lists[g]); - capabilities[i].mut_lists[g] = allocBlock(); + freeChain(capabilities[i]->mut_lists[g]); + capabilities[i]->mut_lists[g] = allocBlock(); } } @@ -1360,7 +1354,7 @@ prepare_uncollected_gen (generation *gen) // allocate a fresh block for each one. We'll traverse these // mutable lists as roots early on in the GC. for (i = 0; i < n_capabilities; i++) { - stash_mut_list(&capabilities[i], gen->no); + stash_mut_list(capabilities[i], gen->no); } ASSERT(gen->scavenged_large_objects == NULL); @@ -1429,7 +1423,7 @@ collect_pinned_object_blocks (void) for (n = 0; n < n_capabilities; n++) { prev = NULL; - for (bd = capabilities[n].pinned_object_blocks; bd != NULL; bd = bd->link) { + for (bd = capabilities[n]->pinned_object_blocks; bd != NULL; bd = bd->link) { prev = bd; } if (prev != NULL) { @@ -1437,8 +1431,8 @@ collect_pinned_object_blocks (void) if (g0->large_objects != NULL) { g0->large_objects->u.back = prev; } - g0->large_objects = capabilities[n].pinned_object_blocks; - capabilities[n].pinned_object_blocks = 0; + g0->large_objects = capabilities[n]->pinned_object_blocks; + capabilities[n]->pinned_object_blocks = 0; } } } diff --git a/rts/sm/Sanity.c b/rts/sm/Sanity.c index 9b579abbbc..0d48ba2724 100644 --- a/rts/sm/Sanity.c +++ b/rts/sm/Sanity.c @@ -615,7 +615,7 @@ checkLocalMutableLists (nat cap_no) { nat g; for (g = 1; g < RtsFlags.GcFlags.generations; g++) { - checkMutableList(capabilities[cap_no].mut_lists[g], g); + checkMutableList(capabilities[cap_no]->mut_lists[g], g); } } @@ -756,7 +756,7 @@ findMemoryLeak (void) nat g, i; for (g = 0; g < RtsFlags.GcFlags.generations; g++) { for (i = 0; i < n_capabilities; i++) { - markBlocks(capabilities[i].mut_lists[g]); + markBlocks(capabilities[i]->mut_lists[g]); markBlocks(gc_threads[i]->gens[g].part_list); markBlocks(gc_threads[i]->gens[g].scavd_list); markBlocks(gc_threads[i]->gens[g].todo_bd); @@ -767,7 +767,7 @@ findMemoryLeak (void) for (i = 0; i < n_capabilities; i++) { markBlocks(nurseries[i].blocks); - markBlocks(capabilities[i].pinned_object_block); + markBlocks(capabilities[i]->pinned_object_block); } #ifdef PROFILING @@ -847,7 +847,7 @@ memInventory (rtsBool show) for (g = 0; g < RtsFlags.GcFlags.generations; g++) { gen_blocks[g] = 0; for (i = 0; i < n_capabilities; i++) { - gen_blocks[g] += countBlocks(capabilities[i].mut_lists[g]); + gen_blocks[g] += countBlocks(capabilities[i]->mut_lists[g]); gen_blocks[g] += countBlocks(gc_threads[i]->gens[g].part_list); gen_blocks[g] += countBlocks(gc_threads[i]->gens[g].scavd_list); gen_blocks[g] += countBlocks(gc_threads[i]->gens[g].todo_bd); @@ -859,10 +859,10 @@ memInventory (rtsBool show) for (i = 0; i < n_capabilities; i++) { ASSERT(countBlocks(nurseries[i].blocks) == nurseries[i].n_blocks); nursery_blocks += nurseries[i].n_blocks; - if (capabilities[i].pinned_object_block != NULL) { - nursery_blocks += capabilities[i].pinned_object_block->blocks; + if (capabilities[i]->pinned_object_block != NULL) { + nursery_blocks += capabilities[i]->pinned_object_block->blocks; } - nursery_blocks += countBlocks(capabilities[i].pinned_object_blocks); + nursery_blocks += countBlocks(capabilities[i]->pinned_object_blocks); } retainer_blocks = 0; diff --git a/rts/sm/Storage.c b/rts/sm/Storage.c index a5337bc5b2..b575fc3e52 100644 --- a/rts/sm/Storage.c +++ b/rts/sm/Storage.c @@ -215,7 +215,7 @@ void storageAddCapabilities (nat from, nat to) // we've moved the nurseries, so we have to update the rNursery // pointers from the Capabilities. for (i = 0; i < to; i++) { - capabilities[i].r.rNursery = &nurseries[i]; + capabilities[i]->r.rNursery = &nurseries[i]; } /* The allocation area. Policy: keep the allocation area @@ -229,7 +229,7 @@ void storageAddCapabilities (nat from, nat to) // allocate a block for each mut list for (n = from; n < to; n++) { for (g = 1; g < RtsFlags.GcFlags.generations; g++) { - capabilities[n].mut_lists[g] = allocBlock(); + capabilities[n]->mut_lists[g] = allocBlock(); } } @@ -493,8 +493,8 @@ assignNurseriesToCapabilities (nat from, nat to) nat i; for (i = from; i < to; i++) { - capabilities[i].r.rCurrentNursery = nurseries[i].blocks; - capabilities[i].r.rCurrentAlloc = NULL; + capabilities[i]->r.rCurrentNursery = nurseries[i].blocks; + capabilities[i]->r.rCurrentAlloc = NULL; } } @@ -939,7 +939,7 @@ void updateNurseriesStats (void) nat i; for (i = 0; i < n_capabilities; i++) { - capabilities[i].total_allocated += countOccupied(nurseries[i].blocks); + capabilities[i]->total_allocated += countOccupied(nurseries[i].blocks); } } |