/* -*- Mode: C++; tab-width: 8; indent-tabs-mode: nil; c-basic-offset: 4 -*- * vim: set ts=8 sts=4 et sw=4 tw=99: * This Source Code Form is subject to the terms of the Mozilla Public * License, v. 2.0. If a copy of the MPL was not distributed with this * file, You can obtain one at http://mozilla.org/MPL/2.0/. */ #ifndef jsgcinlines_h #define jsgcinlines_h #include "jsgc.h" #include "jscntxt.h" #include "jscompartment.h" #include "jslock.h" #include "js/RootingAPI.h" #include "js/TemplateLib.h" #include "vm/Shape.h" #include "vm/ForkJoin.h" namespace js { class Shape; /* * This auto class should be used around any code that might cause a mark bit to * be set on an object in a dead zone. See AutoMaybeTouchDeadZones * for more details. */ struct AutoMarkInDeadZone { AutoMarkInDeadZone(JS::Zone *zone) : zone(zone), scheduled(zone->scheduledForDestruction) { if (zone->rt->gcManipulatingDeadZones && zone->scheduledForDestruction) { zone->rt->gcObjectsMarkedInDeadZones++; zone->scheduledForDestruction = false; } } ~AutoMarkInDeadZone() { zone->scheduledForDestruction = scheduled; } private: JS::Zone *zone; bool scheduled; }; inline Allocator *const ThreadSafeContext::allocator() { JS_ASSERT_IF(isJSContext(), &asJSContext()->zone()->allocator == allocator_); return allocator_; } namespace gc { /* Capacity for slotsToThingKind */ const size_t SLOTS_TO_THING_KIND_LIMIT = 17; extern const AllocKind slotsToThingKind[]; /* Get the best kind to use when making an object with the given slot count. */ static inline AllocKind GetGCObjectKind(size_t numSlots) { if (numSlots >= SLOTS_TO_THING_KIND_LIMIT) return FINALIZE_OBJECT16; return slotsToThingKind[numSlots]; } static inline AllocKind GetGCObjectKind(Class *clasp) { if (clasp == FunctionClassPtr) return JSFunction::FinalizeKind; uint32_t nslots = JSCLASS_RESERVED_SLOTS(clasp); if (clasp->flags & JSCLASS_HAS_PRIVATE) nslots++; return GetGCObjectKind(nslots); } /* As for GetGCObjectKind, but for dense array allocation. */ static inline AllocKind GetGCArrayKind(size_t numSlots) { /* * Dense arrays can use their fixed slots to hold their elements array * (less two Values worth of ObjectElements header), but if more than the * maximum number of fixed slots is needed then the fixed slots will be * unused. */ JS_STATIC_ASSERT(ObjectElements::VALUES_PER_HEADER == 2); if (numSlots > JSObject::NELEMENTS_LIMIT || numSlots + 2 >= SLOTS_TO_THING_KIND_LIMIT) return FINALIZE_OBJECT2; return slotsToThingKind[numSlots + 2]; } static inline AllocKind GetGCObjectFixedSlotsKind(size_t numFixedSlots) { JS_ASSERT(numFixedSlots < SLOTS_TO_THING_KIND_LIMIT); return slotsToThingKind[numFixedSlots]; } static inline AllocKind GetBackgroundAllocKind(AllocKind kind) { JS_ASSERT(!IsBackgroundFinalized(kind)); JS_ASSERT(kind <= FINALIZE_OBJECT_LAST); return (AllocKind) (kind + 1); } /* * Try to get the next larger size for an object, keeping BACKGROUND * consistent. */ static inline bool TryIncrementAllocKind(AllocKind *kindp) { size_t next = size_t(*kindp) + 2; if (next >= size_t(FINALIZE_OBJECT_LIMIT)) return false; *kindp = AllocKind(next); return true; } /* Get the number of fixed slots and initial capacity associated with a kind. */ static inline size_t GetGCKindSlots(AllocKind thingKind) { /* Using a switch in hopes that thingKind will usually be a compile-time constant. */ switch (thingKind) { case FINALIZE_OBJECT0: case FINALIZE_OBJECT0_BACKGROUND: return 0; case FINALIZE_OBJECT2: case FINALIZE_OBJECT2_BACKGROUND: return 2; case FINALIZE_OBJECT4: case FINALIZE_OBJECT4_BACKGROUND: return 4; case FINALIZE_OBJECT8: case FINALIZE_OBJECT8_BACKGROUND: return 8; case FINALIZE_OBJECT12: case FINALIZE_OBJECT12_BACKGROUND: return 12; case FINALIZE_OBJECT16: case FINALIZE_OBJECT16_BACKGROUND: return 16; default: JS_NOT_REACHED("Bad object finalize kind"); return 0; } } static inline size_t GetGCKindSlots(AllocKind thingKind, Class *clasp) { size_t nslots = GetGCKindSlots(thingKind); /* An object's private data uses the space taken by its last fixed slot. */ if (clasp->flags & JSCLASS_HAS_PRIVATE) { JS_ASSERT(nslots > 0); nslots--; } /* * Functions have a larger finalize kind than FINALIZE_OBJECT to reserve * space for the extra fields in JSFunction, but have no fixed slots. */ if (clasp == FunctionClassPtr) nslots = 0; return nslots; } #ifdef JSGC_GENERATIONAL inline bool ShouldNurseryAllocate(const Nursery &nursery, AllocKind kind, InitialHeap heap) { return nursery.isEnabled() && IsNurseryAllocable(kind) && heap != TenuredHeap; } #endif inline bool IsInsideNursery(JSRuntime *rt, const void *thing) { #ifdef JSGC_GENERATIONAL return rt->gcNursery.isInside(thing); #endif return false; } inline JSGCTraceKind GetGCThingTraceKind(const void *thing) { JS_ASSERT(thing); const Cell *cell = static_cast(thing); #ifdef JSGC_GENERATIONAL if (IsInsideNursery(cell->runtime(), cell)) return JSTRACE_OBJECT; #endif return MapAllocToTraceKind(cell->tenuredGetAllocKind()); } static inline void GCPoke(JSRuntime *rt) { rt->gcPoke = true; #ifdef JS_GC_ZEAL /* Schedule a GC to happen "soon" after a GC poke. */ if (rt->gcZeal() == js::gc::ZealPokeValue) rt->gcNextScheduled = 1; #endif } class ArenaIter { ArenaHeader *aheader; ArenaHeader *remainingHeader; public: ArenaIter() { init(); } ArenaIter(JS::Zone *zone, AllocKind kind) { init(zone, kind); } void init() { aheader = NULL; remainingHeader = NULL; } void init(ArenaHeader *aheaderArg) { aheader = aheaderArg; remainingHeader = NULL; } void init(JS::Zone *zone, AllocKind kind) { aheader = zone->allocator.arenas.getFirstArena(kind); remainingHeader = zone->allocator.arenas.getFirstArenaToSweep(kind); if (!aheader) { aheader = remainingHeader; remainingHeader = NULL; } } bool done() { return !aheader; } ArenaHeader *get() { return aheader; } void next() { aheader = aheader->next; if (!aheader) { aheader = remainingHeader; remainingHeader = NULL; } } }; class CellIterImpl { size_t firstThingOffset; size_t thingSize; ArenaIter aiter; FreeSpan firstSpan; const FreeSpan *span; uintptr_t thing; Cell *cell; protected: CellIterImpl() { } void initSpan(JS::Zone *zone, AllocKind kind) { JS_ASSERT(zone->allocator.arenas.isSynchronizedFreeList(kind)); firstThingOffset = Arena::firstThingOffset(kind); thingSize = Arena::thingSize(kind); firstSpan.initAsEmpty(); span = &firstSpan; thing = span->first; } void init(ArenaHeader *singleAheader) { initSpan(singleAheader->zone, singleAheader->getAllocKind()); aiter.init(singleAheader); next(); aiter.init(); } void init(JS::Zone *zone, AllocKind kind) { initSpan(zone, kind); aiter.init(zone, kind); next(); } public: bool done() const { return !cell; } template T *get() const { JS_ASSERT(!done()); return static_cast(cell); } Cell *getCell() const { JS_ASSERT(!done()); return cell; } void next() { for (;;) { if (thing != span->first) break; if (JS_LIKELY(span->hasNext())) { thing = span->last + thingSize; span = span->nextSpan(); break; } if (aiter.done()) { cell = NULL; return; } ArenaHeader *aheader = aiter.get(); firstSpan = aheader->getFirstFreeSpan(); span = &firstSpan; thing = aheader->arenaAddress() | firstThingOffset; aiter.next(); } cell = reinterpret_cast(thing); thing += thingSize; } }; class CellIterUnderGC : public CellIterImpl { public: CellIterUnderGC(JS::Zone *zone, AllocKind kind) { JS_ASSERT(zone->rt->isHeapBusy()); init(zone, kind); } CellIterUnderGC(ArenaHeader *aheader) { JS_ASSERT(aheader->zone->rt->isHeapBusy()); init(aheader); } }; class CellIter : public CellIterImpl { ArenaLists *lists; AllocKind kind; #ifdef DEBUG size_t *counter; #endif public: CellIter(JS::Zone *zone, AllocKind kind) : lists(&zone->allocator.arenas), kind(kind) { /* * We have a single-threaded runtime, so there's no need to protect * against other threads iterating or allocating. However, we do have * background finalization; we have to wait for this to finish if it's * currently active. */ if (IsBackgroundFinalized(kind) && zone->allocator.arenas.needBackgroundFinalizeWait(kind)) { gc::FinishBackgroundFinalize(zone->rt); } if (lists->isSynchronizedFreeList(kind)) { lists = NULL; } else { JS_ASSERT(!zone->rt->isHeapBusy()); lists->copyFreeListToArena(kind); } #ifdef DEBUG counter = &zone->rt->noGCOrAllocationCheck; ++*counter; #endif init(zone, kind); } ~CellIter() { #ifdef DEBUG JS_ASSERT(*counter > 0); --*counter; #endif if (lists) lists->clearFreeListInArena(kind); } }; class GCZonesIter { private: ZonesIter zone; public: GCZonesIter(JSRuntime *rt) : zone(rt) { if (!zone->isCollecting()) next(); } bool done() const { return zone.done(); } void next() { JS_ASSERT(!done()); do { zone.next(); } while (!zone.done() && !zone->isCollecting()); } JS::Zone *get() const { JS_ASSERT(!done()); return zone; } operator JS::Zone *() const { return get(); } JS::Zone *operator->() const { return get(); } }; typedef CompartmentsIterT GCCompartmentsIter; /* Iterates over all zones in the current zone group. */ class GCZoneGroupIter { private: JS::Zone *current; public: GCZoneGroupIter(JSRuntime *rt) { JS_ASSERT(rt->isHeapBusy()); current = rt->gcCurrentZoneGroup; } bool done() const { return !current; } void next() { JS_ASSERT(!done()); current = current->nextNodeInGroup(); } JS::Zone *get() const { JS_ASSERT(!done()); return current; } operator JS::Zone *() const { return get(); } JS::Zone *operator->() const { return get(); } }; typedef CompartmentsIterT GCCompartmentGroupIter; #ifdef JSGC_GENERATIONAL /* * Attempt to allocate a new GC thing out of the nursery. If there is not enough * room in the nursery or there is an OOM, this method will return NULL. */ template inline T * TryNewNurseryGCThing(ThreadSafeContext *tcx, size_t thingSize) { /* TODO: Integrate PJS with generational GC. */ JSContext *cx = tcx->asJSContext(); JS_ASSERT(!IsAtomsCompartment(cx->compartment())); JSRuntime *rt = cx->runtime(); Nursery &nursery = rt->gcNursery; T *t = static_cast(nursery.allocate(thingSize)); if (t) return t; if (allowGC && !rt->mainThread.suppressGC) { MinorGC(rt, JS::gcreason::OUT_OF_NURSERY); /* Exceeding gcMaxBytes while tenuring can disable the Nursery. */ if (nursery.isEnabled()) { t = static_cast(nursery.allocate(thingSize)); JS_ASSERT(t); return t; } } return NULL; } #endif /* JSGC_GENERATIONAL */ /* * Allocates a new GC thing. After a successful allocation the caller must * fully initialize the thing before calling any function that can potentially * trigger GC. This will ensure that GC tracing never sees junk values stored * in the partially initialized thing. */ template inline T * NewGCThing(js::ThreadSafeContext *tcx, AllocKind kind, size_t thingSize, InitialHeap heap) { JS_ASSERT(thingSize == js::gc::Arena::thingSize(kind)); if (tcx->isJSContext()) { JSContext *cx = tcx->asJSContext(); JS_ASSERT_IF(cx->compartment() == cx->runtime()->atomsCompartment, kind == FINALIZE_STRING || kind == FINALIZE_SHORT_STRING || kind == FINALIZE_IONCODE); JS_ASSERT(!cx->runtime()->isHeapBusy()); JS_ASSERT(!cx->runtime()->noGCOrAllocationCheck); /* For testing out of memory conditions */ JS_OOM_POSSIBLY_FAIL_REPORT(cx); #ifdef JS_GC_ZEAL if (cx->runtime()->needZealousGC() && allowGC) js::gc::RunDebugGC(cx); #endif if (allowGC) MaybeCheckStackRoots(cx); } #if defined(JSGC_GENERATIONAL) if (tcx->hasNursery() && ShouldNurseryAllocate(tcx->nursery(), kind, heap)) { T *t = TryNewNurseryGCThing(tcx, thingSize); if (t) return t; } #endif T *t = static_cast(tcx->allocator()->arenas.allocateFromFreeList(kind, thingSize)); if (!t) t = static_cast(js::gc::ArenaLists::refillFreeList(tcx, kind)); #ifdef DEBUG if (tcx->isJSContext()) { Zone *zone = tcx->asJSContext()->zone(); JS_ASSERT_IF(t && zone->wasGCStarted() && (zone->isGCMarking() || zone->isGCSweeping()), t->arenaHeader()->allocatedDuringIncremental); } #endif return t; } } /* namespace gc */ } /* namespace js */ template inline JSObject * js_NewGCObject(js::ThreadSafeContext *tcx, js::gc::AllocKind kind, js::gc::InitialHeap heap) { JS_ASSERT(kind >= js::gc::FINALIZE_OBJECT0 && kind <= js::gc::FINALIZE_OBJECT_LAST); return js::gc::NewGCThing(tcx, kind, js::gc::Arena::thingSize(kind), heap); } template inline JSString * js_NewGCString(js::ThreadSafeContext *tcx) { return js::gc::NewGCThing(tcx, js::gc::FINALIZE_STRING, sizeof(JSString), js::gc::TenuredHeap); } template inline JSShortString * js_NewGCShortString(js::ThreadSafeContext *tcx) { return js::gc::NewGCThing(tcx, js::gc::FINALIZE_SHORT_STRING, sizeof(JSShortString), js::gc::TenuredHeap); } inline JSExternalString * js_NewGCExternalString(js::ThreadSafeContext *tcx) { return js::gc::NewGCThing(tcx, js::gc::FINALIZE_EXTERNAL_STRING, sizeof(JSExternalString), js::gc::TenuredHeap); } inline JSScript * js_NewGCScript(js::ThreadSafeContext *tcx) { return js::gc::NewGCThing(tcx, js::gc::FINALIZE_SCRIPT, sizeof(JSScript), js::gc::TenuredHeap); } inline js::LazyScript * js_NewGCLazyScript(js::ThreadSafeContext *tcx) { return js::gc::NewGCThing(tcx, js::gc::FINALIZE_LAZY_SCRIPT, sizeof(js::LazyScript), js::gc::TenuredHeap); } inline js::Shape * js_NewGCShape(js::ThreadSafeContext *tcx) { return js::gc::NewGCThing(tcx, js::gc::FINALIZE_SHAPE, sizeof(js::Shape), js::gc::TenuredHeap); } template inline js::BaseShape * js_NewGCBaseShape(js::ThreadSafeContext *tcx) { return js::gc::NewGCThing(tcx, js::gc::FINALIZE_BASE_SHAPE, sizeof(js::BaseShape), js::gc::TenuredHeap); } #endif /* jsgcinlines_h */