diff options
Diffstat (limited to 'rts')
-rw-r--r-- | rts/LdvProfile.c | 2 | ||||
-rw-r--r-- | rts/PrimOps.cmm | 2 | ||||
-rw-r--r-- | rts/STM.c | 4 | ||||
-rw-r--r-- | rts/Schedule.c | 4 | ||||
-rw-r--r-- | rts/TraverseHeap.c | 2 | ||||
-rw-r--r-- | rts/linker/Elf.c | 2 | ||||
-rw-r--r-- | rts/linker/MachO.c | 6 | ||||
-rw-r--r-- | rts/sm/GC.c | 4 | ||||
-rw-r--r-- | rts/sm/NonMovingMark.c | 4 |
9 files changed, 15 insertions, 15 deletions
diff --git a/rts/LdvProfile.c b/rts/LdvProfile.c index f4ab02886d..cf57f28eae 100644 --- a/rts/LdvProfile.c +++ b/rts/LdvProfile.c @@ -111,7 +111,7 @@ processHeapClosureForDead( const StgClosure *c ) case BLACKHOLE: case BLOCKING_QUEUE: /* - 'Ingore' cases + 'Ignore' cases */ // Why can we ignore IND closures? We assume that // any census is preceded by a major garbage collection, which diff --git a/rts/PrimOps.cmm b/rts/PrimOps.cmm index b5930363a1..0486399b46 100644 --- a/rts/PrimOps.cmm +++ b/rts/PrimOps.cmm @@ -1313,7 +1313,7 @@ stg_atomicallyzh (P_ stm) code = stm; frame_result = NO_TREC; - /* Start the memory transcation */ + /* Start the memory transaction */ ("ptr" new_trec) = ccall stmStartTransaction(MyCapability() "ptr", old_trec "ptr"); StgTSO_trec(CurrentTSO) = new_trec; @@ -828,7 +828,7 @@ static StgBool validate_and_acquire_ownership (Capability *cap, // check_read_only : check that we've seen an atomic snapshot of the // non-updated TVars accessed by a trec. This checks that the last TRec to // commit an update to the TVar is unchanged since the value was stashed in -// validate_and_acquire_ownership. If no udpate is seen to any TVar than +// validate_and_acquire_ownership. If no update is seen to any TVar than // all of them contained their expected values at the start of the call to // check_read_only. // @@ -1252,7 +1252,7 @@ StgBool stmReWait(Capability *cap, StgTSO *tso) { park_tso(tso); revert_ownership(cap, trec, true); } else { - // The transcation has become invalid. We can now remove it from the wait + // The transaction has become invalid. We can now remove it from the wait // queues. if (trec -> state != TREC_CONDEMNED) { remove_watch_queue_entries_for_trec (cap, trec); diff --git a/rts/Schedule.c b/rts/Schedule.c index 9323915dfe..ce1a1fc060 100644 --- a/rts/Schedule.c +++ b/rts/Schedule.c @@ -226,7 +226,7 @@ schedule (Capability *initialCapability, Task *task) // In order to cleanly shut down the runtime, we want to: // * make sure that all main threads return to their callers // with the state 'Interrupted'. - // * clean up all OS threads assocated with the runtime + // * clean up all OS threads associated with the runtime // * free all memory etc. // // So the sequence goes like this: @@ -2906,7 +2906,7 @@ raiseExceptionHelper (StgRegTable *reg, StgTSO *tso, StgClosure *exception) // OLD COMMENT (we don't have MIN_UPD_SIZE now): // LDV profiling: stg_raise_info has THUNK as its closure // type. Since a THUNK takes at least MIN_UPD_SIZE words in its - // payload, MIN_UPD_SIZE is more approprate than 1. It seems that + // payload, MIN_UPD_SIZE is more appropriate than 1. It seems that // 1 does not cause any problem unless profiling is performed. // However, when LDV profiling goes on, we need to linearly scan // small object pool, where raise_closure is stored, so we should diff --git a/rts/TraverseHeap.c b/rts/TraverseHeap.c index 62d44c1636..8bf58c11ee 100644 --- a/rts/TraverseHeap.c +++ b/rts/TraverseHeap.c @@ -324,7 +324,7 @@ pushStackElement(traverseState *ts, const stackElement se) newStackBlock(ts, nbd); } - // adjust stackTop (acutal push) + // adjust stackTop (actual push) ts->stackTop--; // If the size of stackElement was huge, we would better replace the // following statement by either a memcpy() call or a switch statement diff --git a/rts/linker/Elf.c b/rts/linker/Elf.c index 3e19d3a2db..c240a97d5e 100644 --- a/rts/linker/Elf.c +++ b/rts/linker/Elf.c @@ -775,7 +775,7 @@ ocGetNames_ELF ( ObjectCode* oc ) alloc = SECTION_NOMEM; } // use the m32 allocator if either the image is not mapped - // (i.e. we cannot map the secions separately), or if the section + // (i.e. we cannot map the sections separately), or if the section // size is small. else if (!oc->imageMapped || size < getPageSize() / 3) { bool executable = kind == SECTIONKIND_CODE_OR_RODATA; diff --git a/rts/linker/MachO.c b/rts/linker/MachO.c index 12e84578fc..6e379ffb53 100644 --- a/rts/linker/MachO.c +++ b/rts/linker/MachO.c @@ -52,7 +52,7 @@ int64_t signExtend(uint64_t val, uint8_t bits); bool isVectorPp(uint32_t *p); bool isLoadStore(uint32_t *p); -/* aarch64 relocations may contain an addend alreay in the position +/* aarch64 relocations may contain an addend already in the position * where we want to write the address offset to. Thus decoding as well * as encoding is needed. */ @@ -305,7 +305,7 @@ decodeAddend(ObjectCode * oc, Section * section, MachORelocationInfo * ri) { } case ARM64_RELOC_BRANCH26: /* take the lower 26 bits and shift them by 2. The last two are - * implicilty 0 (as the instructions must be aligned!) and sign + * implicitly 0 (as the instructions must be aligned!) and sign * extend to 64 bits. */ return signExtend( (*p & 0x03FFFFFF) << 2, 28 ); @@ -543,7 +543,7 @@ relocateSectionAarch64(ObjectCode * oc, Section * section) * * - loaded the sections (potentially into non-contiguous memory), * (in ocGetNames_MachO) - * - registered exported sybmols + * - registered exported symbols * (in ocGetNames_MachO) * - and fixed the nlist[i].n_value for common storage symbols (N_UNDF, * N_EXT and n_value != 0) so that they point into the common storage. diff --git a/rts/sm/GC.c b/rts/sm/GC.c index a560c2c8fd..aac8d974ae 100644 --- a/rts/sm/GC.c +++ b/rts/sm/GC.c @@ -596,7 +596,7 @@ GarbageCollect (uint32_t collect_gen, /* free old memory and shift to-space into from-space for all * the collected generations (except the allocation area). These - * freed blocks will probaby be quickly recycled. + * freed blocks will probably be quickly recycled. */ if (gen->mark) { @@ -646,7 +646,7 @@ GarbageCollect (uint32_t collect_gen, ASSERT(countBlocks(gen->blocks) == gen->n_blocks); ASSERT(countOccupied(gen->blocks) == gen->n_words); } - else // not copacted + else // not compacted { freeChain(gen->old_blocks); } diff --git a/rts/sm/NonMovingMark.c b/rts/sm/NonMovingMark.c index 3ebd039c43..81000566cb 100644 --- a/rts/sm/NonMovingMark.c +++ b/rts/sm/NonMovingMark.c @@ -1265,7 +1265,7 @@ mark_closure (MarkQueue *queue, const StgClosure *p0, StgClosure **origin) ASSERT(!IS_FORWARDING_PTR(p->header.info)); // N.B. only the first block of a compact region is guaranteed to carry - // BF_NONMOVING; conseqently we must separately check for BF_COMPACT. + // BF_NONMOVING; consequently we must separately check for BF_COMPACT. if (bd->flags & (BF_COMPACT | BF_NONMOVING)) { if (bd->flags & BF_COMPACT) { @@ -1315,7 +1315,7 @@ mark_closure (MarkQueue *queue, const StgClosure *p0, StgClosure **origin) } // A pinned object that is still attached to a capability (because it's not - // filled yet). No need to trace it pinned objects can't contain poiners. + // filled yet). No need to trace it pinned objects can't contain pointers. else if (bd->flags & BF_PINNED) { #if defined(DEBUG) bool found_it = false; |