diff options
author | Josh Meredith <joshmeredith2008@gmail.com> | 2019-12-04 23:39:28 +1100 |
---|---|---|
committer | Josh Meredith <joshmeredith2008@gmail.com> | 2019-12-04 23:39:28 +1100 |
commit | a8435165b84c32fd2ebdd1281dd6ee077e07ad5a (patch) | |
tree | 791936d014aeaa26174c2dcbef34c14f3329dd04 /rts | |
parent | 7805441b4d5e22eb63a501e1e40383d10380dc92 (diff) | |
parent | f03a41d4bf9418ee028ecb51654c928b2da74edd (diff) | |
download | haskell-wip/binary-readerT.tar.gz |
Merge branch 'master' into wip/binary-readerTwip/binary-readerT
Diffstat (limited to 'rts')
-rw-r--r-- | rts/Apply.cmm | 2 | ||||
-rw-r--r-- | rts/Hpc.c | 2 | ||||
-rw-r--r-- | rts/LinkerInternals.h | 2 | ||||
-rw-r--r-- | rts/RetainerProfile.c | 2 | ||||
-rw-r--r-- | rts/StablePtr.c | 2 | ||||
-rw-r--r-- | rts/StgCRun.c | 2 | ||||
-rw-r--r-- | rts/linker/M32Alloc.c | 2 | ||||
-rw-r--r-- | rts/linker/MachO.c | 2 | ||||
-rw-r--r-- | rts/linker/PEi386.c | 4 | ||||
-rw-r--r-- | rts/posix/OSMem.c | 2 | ||||
-rw-r--r-- | rts/sm/GC.c | 2 | ||||
-rw-r--r-- | rts/sm/NonMoving.c | 4 | ||||
-rw-r--r-- | rts/sm/NonMovingMark.c | 2 | ||||
-rw-r--r-- | rts/sm/Storage.c | 2 |
14 files changed, 16 insertions, 16 deletions
diff --git a/rts/Apply.cmm b/rts/Apply.cmm index eeb760c5ed..dcfaa446f2 100644 --- a/rts/Apply.cmm +++ b/rts/Apply.cmm @@ -571,7 +571,7 @@ being carried out by TSO 2 and attempt to suspend it. The suspension process proceeds by invoking raiseAsync, which walks the stack from the top looking for update frames. For each update frame we take any stack -frames preceeding it and construct an AP_STACK heap object from them. We then +frames preceding it and construct an AP_STACK heap object from them. We then replace the updatee of the frame with an indirection pointing to the AP_STACK. So, after suspending the first update frame we have, @@ -373,7 +373,7 @@ freeHpcModuleInfo (HpcModuleInfo *mod) } /* Called at the end of execution, to write out the Hpc *.tix file - * for this exection. Safe to call, even if coverage is not used. + * for this execution. Safe to call, even if coverage is not used. */ void exitHpc(void) { diff --git a/rts/LinkerInternals.h b/rts/LinkerInternals.h index 5cd35f2e5b..0f47b82761 100644 --- a/rts/LinkerInternals.h +++ b/rts/LinkerInternals.h @@ -361,7 +361,7 @@ void freeSegments(ObjectCode *oc); #define MAP_ANONYMOUS MAP_ANON #endif -/* Which object file format are we targetting? */ +/* Which object file format are we targeting? */ #if defined(linux_HOST_OS) || defined(solaris2_HOST_OS) \ || defined(linux_android_HOST_OS) \ || defined(freebsd_HOST_OS) || defined(kfreebsdgnu_HOST_OS) \ diff --git a/rts/RetainerProfile.c b/rts/RetainerProfile.c index 97c4b35a71..a836c5bf2a 100644 --- a/rts/RetainerProfile.c +++ b/rts/RetainerProfile.c @@ -374,7 +374,7 @@ computeRetainerSet( traverseState *ts ) // all are guaranteed to be valid, or reachable. // // The following code assumes that WEAK objects are considered to be roots - // for retainer profilng. + // for retainer profiling. for (n = 0; n < n_capabilities; n++) { // NB: after a GC, all nursery weak_ptr_lists have been migrated // to the global lists living in the generations diff --git a/rts/StablePtr.c b/rts/StablePtr.c index 2181b83d90..edcd863183 100644 --- a/rts/StablePtr.c +++ b/rts/StablePtr.c @@ -32,7 +32,7 @@ for garbage collection because the act of passing them makes a copy from the heap, stack or wherever they are onto the C-world stack. However, if we were to pass a heap object such as a (Haskell) @String@ - and a garbage collection occured before we finished using it, we'd run + and a garbage collection occurred before we finished using it, we'd run into problems since the heap object might have been moved or even deleted. diff --git a/rts/StgCRun.c b/rts/StgCRun.c index 934926e0f3..a6ef7054a4 100644 --- a/rts/StgCRun.c +++ b/rts/StgCRun.c @@ -131,7 +131,7 @@ STG_NO_OPTIMIZE StgWord8 *win32AllocStack(void) * ABI requires this (x64, Mac OSX 32bit/64bit) as well as interfacing with * other libraries through the FFI. * - * As part of this arrangment we must maintain the stack at a 16-byte boundary + * As part of this arrangement we must maintain the stack at a 16-byte boundary * - word_size-bytes (so 16n - 4 for i386 and 16n - 8 for x64) on entry to a * procedure since both GCC and LLVM expect this. This is because the stack * should have been 16-byte boundary aligned and then a call made which pushes diff --git a/rts/linker/M32Alloc.c b/rts/linker/M32Alloc.c index dbba711df4..148a5043cf 100644 --- a/rts/linker/M32Alloc.c +++ b/rts/linker/M32Alloc.c @@ -96,7 +96,7 @@ life-cycle it is in: allocation) Allocation (in the case of a small request) consists of walking the nursery to -find a page that will accomodate the request. If none exists then we allocate a +find a page that will accommodate the request. If none exists then we allocate a new nursery page (flushing an existing one to the filled list if the nursery is full). diff --git a/rts/linker/MachO.c b/rts/linker/MachO.c index 5ae7620fc7..12e84578fc 100644 --- a/rts/linker/MachO.c +++ b/rts/linker/MachO.c @@ -427,7 +427,7 @@ isGotLoad(struct relocation_info * ri) { /* This is very similar to makeSymbolExtra * However, as we load sections into different - * pages, that may be further appart than + * pages, that may be further apart than * branching allows, we'll use some extra * space at the end of each section allocated * for stubs. diff --git a/rts/linker/PEi386.c b/rts/linker/PEi386.c index 2f65f6a208..81308779a1 100644 --- a/rts/linker/PEi386.c +++ b/rts/linker/PEi386.c @@ -105,7 +105,7 @@ contain the name of the actual dll to load. This will be the only content of the section. In the symbol table, the last symbol will be the name used to refer to the dll in the relocation tables. This name will always - be in the format "symbol_name_iname", however when refered to, the format + be in the format "symbol_name_iname", however when referred to, the format "_head_symbol_name" is used. We record this symbol early on during GetNames and load the dll and use @@ -1535,7 +1535,7 @@ ocGetNames_PEi386 ( ObjectCode* oc ) } setImportSymbol (oc, sname); - /* Don't process this oc any futher. Just exit. */ + /* Don't process this oc any further. Just exit. */ oc->n_symbols = 0; oc->symbols = NULL; stgFree (oc->image); diff --git a/rts/posix/OSMem.c b/rts/posix/OSMem.c index 15f1b62dc9..8e797681c9 100644 --- a/rts/posix/OSMem.c +++ b/rts/posix/OSMem.c @@ -183,7 +183,7 @@ my_mmap (void *addr, W_ size, int operation) #if defined(darwin_HOST_OS) // Without MAP_FIXED, Apple's mmap ignores addr. - // With MAP_FIXED, it overwrites already mapped regions, whic + // With MAP_FIXED, it overwrites already mapped regions, which // mmap(0, ... MAP_FIXED ...) is worst of all: It unmaps the program text // and replaces it with zeroes, causing instant death. // This behaviour seems to be conformant with IEEE Std 1003.1-2001. diff --git a/rts/sm/GC.c b/rts/sm/GC.c index 83e9c97bd9..a560c2c8fd 100644 --- a/rts/sm/GC.c +++ b/rts/sm/GC.c @@ -743,7 +743,7 @@ GarbageCollect (uint32_t collect_gen, // oldest_gen->scavenged_large_objects back to oldest_gen->large_objects. ASSERT(oldest_gen->scavenged_large_objects == NULL); if (RtsFlags.GcFlags.useNonmoving && major_gc) { - // All threads in non-moving heap should be found to be alive, becuase + // All threads in non-moving heap should be found to be alive, because // threads in the non-moving generation's list should live in the // non-moving heap, and we consider non-moving objects alive during // preparation. diff --git a/rts/sm/NonMoving.c b/rts/sm/NonMoving.c index 20ec5a45ba..5cb7c8b7e5 100644 --- a/rts/sm/NonMoving.c +++ b/rts/sm/NonMoving.c @@ -200,7 +200,7 @@ Mutex concurrent_coll_finished_lock; * generation. * * - Note [Aging under the non-moving collector] (NonMoving.c) describes how - * we accomodate aging + * we accommodate aging * * - Note [Large objects in the non-moving collector] (NonMovingMark.c) * describes how we track large objects. @@ -890,7 +890,7 @@ void nonmovingCollect(StgWeak **dead_weaks, StgTSO **resurrected_threads) // If we're interrupting or shutting down, do not let this capability go and // run a STW collection. Reason: we won't be able to acquire this capability // again for the sync if we let it go, because it'll immediately start doing - // a major GC, becuase that's what we do when exiting scheduler (see + // a major GC, because that's what we do when exiting scheduler (see // exitScheduler()). if (sched_state == SCHED_RUNNING) { concurrent_coll_running = true; diff --git a/rts/sm/NonMovingMark.c b/rts/sm/NonMovingMark.c index 085d7827df..3ebd039c43 100644 --- a/rts/sm/NonMovingMark.c +++ b/rts/sm/NonMovingMark.c @@ -88,7 +88,7 @@ memcount n_nonmoving_marked_compact_blocks = 0; * move the same large object to nonmoving_marked_large_objects more than once. */ static Mutex nonmoving_large_objects_mutex; -// Note that we don't need a similar lock for compact objects becuase we never +// Note that we don't need a similar lock for compact objects because we never // mark a compact object eagerly in a write barrier; all compact objects are // marked by the mark thread, so there can't be any races here. #endif diff --git a/rts/sm/Storage.c b/rts/sm/Storage.c index f04b3c5929..cdae368150 100644 --- a/rts/sm/Storage.c +++ b/rts/sm/Storage.c @@ -406,7 +406,7 @@ freeStorage (bool free_heap) but can assume only two bits are available in STATIC_LINK (due to 32-bit systems). - To accomodate this we move handling of static objects entirely to the + To accommodate this we move handling of static objects entirely to the oldest generation when the nonmoving collector is in use. To do this safely and efficiently we allocate the blackhole created by lockCAF() directly in the non-moving heap. This means that the moving collector can completely |