/* -*- Mode: C++; tab-width: 8; indent-tabs-mode: nil; c-basic-offset: 4 -*- * vim: set ts=8 sts=4 et sw=4 tw=99: * This Source Code Form is subject to the terms of the Mozilla Public * License, v. 2.0. If a copy of the MPL was not distributed with this * file, You can obtain one at http://mozilla.org/MPL/2.0/. */ #ifndef gc_Marking_h #define gc_Marking_h #include "mozilla/HashFunctions.h" #include "mozilla/Move.h" #include "jsfriendapi.h" #include "ds/OrderedHashTable.h" #include "gc/Heap.h" #include "gc/Tracer.h" #include "js/GCAPI.h" #include "js/HeapAPI.h" #include "js/SliceBudget.h" #include "js/TracingAPI.h" #include "vm/TaggedProto.h" class JSLinearString; class JSRope; namespace js { class BaseShape; class GCMarker; class LazyScript; class NativeObject; class ObjectGroup; class WeakMapBase; namespace gc { class Arena; } // namespace gc namespace jit { class JitCode; } // namespace jit static const size_t NON_INCREMENTAL_MARK_STACK_BASE_CAPACITY = 4096; static const size_t INCREMENTAL_MARK_STACK_BASE_CAPACITY = 32768; /* * When the native stack is low, the GC does not call js::TraceChildren to mark * the reachable "children" of the thing. Rather the thing is put aside and * js::TraceChildren is called later with more space on the C stack. * * To implement such delayed marking of the children with minimal overhead for * the normal case of sufficient native stack, the code adds a field per arena. * The field markingDelay->link links all arenas with delayed things into a * stack list with the pointer to stack top in GCMarker::unmarkedArenaStackTop. * GCMarker::delayMarkingChildren adds arenas to the stack as necessary while * markDelayedChildren pops the arenas from the stack until it empties. */ class MarkStack { friend class GCMarker; uintptr_t* stack_; uintptr_t* tos_; uintptr_t* end_; // The capacity we start with and reset() to. size_t baseCapacity_; size_t maxCapacity_; public: explicit MarkStack(size_t maxCapacity) : stack_(nullptr), tos_(nullptr), end_(nullptr), baseCapacity_(0), maxCapacity_(maxCapacity) {} ~MarkStack() { js_free(stack_); } size_t capacity() { return end_ - stack_; } ptrdiff_t position() const { return tos_ - stack_; } void setStack(uintptr_t* stack, size_t tosIndex, size_t capacity) { stack_ = stack; tos_ = stack + tosIndex; end_ = stack + capacity; } MOZ_MUST_USE bool init(JSGCMode gcMode); void setBaseCapacity(JSGCMode mode); size_t maxCapacity() const { return maxCapacity_; } void setMaxCapacity(size_t maxCapacity); MOZ_MUST_USE bool push(uintptr_t item) { if (tos_ == end_) { if (!enlarge(1)) return false; } MOZ_ASSERT(tos_ < end_); *tos_++ = item; return true; } MOZ_MUST_USE bool push(uintptr_t item1, uintptr_t item2, uintptr_t item3) { uintptr_t* nextTos = tos_ + 3; if (nextTos > end_) { if (!enlarge(3)) return false; nextTos = tos_ + 3; } MOZ_ASSERT(nextTos <= end_); tos_[0] = item1; tos_[1] = item2; tos_[2] = item3; tos_ = nextTos; return true; } bool isEmpty() const { return tos_ == stack_; } uintptr_t pop() { MOZ_ASSERT(!isEmpty()); return *--tos_; } void reset(); /* Grow the stack, ensuring there is space for at least count elements. */ MOZ_MUST_USE bool enlarge(unsigned count); void setGCMode(JSGCMode gcMode); size_t sizeOfExcludingThis(mozilla::MallocSizeOf mallocSizeOf) const; }; namespace gc { struct WeakKeyTableHashPolicy { typedef JS::GCCellPtr Lookup; static HashNumber hash(const Lookup& v, const mozilla::HashCodeScrambler&) { return mozilla::HashGeneric(v.asCell()); } static bool match(const JS::GCCellPtr& k, const Lookup& l) { return k == l; } static bool isEmpty(const JS::GCCellPtr& v) { return !v; } static void makeEmpty(JS::GCCellPtr* vp) { *vp = nullptr; } }; struct WeakMarkable { WeakMapBase* weakmap; JS::GCCellPtr key; WeakMarkable(WeakMapBase* weakmapArg, JS::GCCellPtr keyArg) : weakmap(weakmapArg), key(keyArg) {} }; using WeakEntryVector = Vector; using WeakKeyTable = OrderedHashMap; } /* namespace gc */ class GCMarker : public JSTracer { public: explicit GCMarker(JSRuntime* rt); MOZ_MUST_USE bool init(JSGCMode gcMode); void setMaxCapacity(size_t maxCap) { stack.setMaxCapacity(maxCap); } size_t maxCapacity() const { return stack.maxCapacity(); } void start(); void stop(); void reset(); // Mark the given GC thing and traverse its children at some point. template void traverse(T thing); // Calls traverse on target after making additional assertions. template void traverseEdge(S source, T* target); template void traverseEdge(S source, const T& target); // Notes a weak graph edge for later sweeping. template void noteWeakEdge(T* edge); /* * Care must be taken changing the mark color from gray to black. The cycle * collector depends on the invariant that there are no black to gray edges * in the GC heap. This invariant lets the CC not trace through black * objects. If this invariant is violated, the cycle collector may free * objects that are still reachable. */ void setMarkColorGray() { MOZ_ASSERT(isDrained()); MOZ_ASSERT(color == gc::BLACK); color = gc::GRAY; } void setMarkColorBlack() { MOZ_ASSERT(isDrained()); MOZ_ASSERT(color == gc::GRAY); color = gc::BLACK; } uint32_t markColor() const { return color; } void enterWeakMarkingMode(); void leaveWeakMarkingMode(); void abortLinearWeakMarking() { leaveWeakMarkingMode(); linearWeakMarkingDisabled_ = true; } void delayMarkingArena(gc::Arena* arena); void delayMarkingChildren(const void* thing); void markDelayedChildren(gc::Arena* arena); MOZ_MUST_USE bool markDelayedChildren(SliceBudget& budget); bool hasDelayedChildren() const { return !!unmarkedArenaStackTop; } bool isDrained() { return isMarkStackEmpty() && !unmarkedArenaStackTop; } MOZ_MUST_USE bool drainMarkStack(SliceBudget& budget); void setGCMode(JSGCMode mode) { stack.setGCMode(mode); } size_t sizeOfExcludingThis(mozilla::MallocSizeOf mallocSizeOf) const; #ifdef DEBUG bool shouldCheckCompartments() { return strictCompartmentChecking; } #endif void markEphemeronValues(gc::Cell* markedCell, gc::WeakEntryVector& entry); private: #ifdef DEBUG void checkZone(void* p); #else void checkZone(void* p) {} #endif /* * We use a common mark stack to mark GC things of different types and use * the explicit tags to distinguish them when it cannot be deduced from * the context of push or pop operation. */ enum StackTag { ValueArrayTag, ObjectTag, GroupTag, SavedValueArrayTag, JitCodeTag, ScriptTag, LastTag = JitCodeTag }; static const uintptr_t StackTagMask = 7; static_assert(StackTagMask >= uintptr_t(LastTag), "The tag mask must subsume the tags."); static_assert(StackTagMask <= gc::CellMask, "The tag mask must be embeddable in a Cell*."); // Push an object onto the stack for later tracing and assert that it has // already been marked. void repush(JSObject* obj) { MOZ_ASSERT(gc::TenuredCell::fromPointer(obj)->isMarked(markColor())); pushTaggedPtr(ObjectTag, obj); } template void markAndTraceChildren(T* thing); template void markAndPush(StackTag tag, T* thing); template void markAndScan(T* thing); template void markImplicitEdgesHelper(T oldThing); template void markImplicitEdges(T* oldThing); void eagerlyMarkChildren(JSLinearString* str); void eagerlyMarkChildren(JSRope* rope); void eagerlyMarkChildren(JSString* str); void eagerlyMarkChildren(LazyScript *thing); void eagerlyMarkChildren(Shape* shape); void eagerlyMarkChildren(Scope* scope); void lazilyMarkChildren(ObjectGroup* group); // We may not have concrete types yet, so this has to be outside the header. template void dispatchToTraceChildren(T* thing); // Mark the given GC thing, but do not trace its children. Return true // if the thing became marked. template MOZ_MUST_USE bool mark(T* thing); void pushTaggedPtr(StackTag tag, void* ptr) { checkZone(ptr); uintptr_t addr = reinterpret_cast(ptr); MOZ_ASSERT(!(addr & StackTagMask)); if (!stack.push(addr | uintptr_t(tag))) delayMarkingChildren(ptr); } void pushValueArray(JSObject* obj, HeapSlot* start, HeapSlot* end) { checkZone(obj); MOZ_ASSERT(start <= end); uintptr_t tagged = reinterpret_cast(obj) | GCMarker::ValueArrayTag; uintptr_t startAddr = reinterpret_cast(start); uintptr_t endAddr = reinterpret_cast(end); /* * Push in the reverse order so obj will be on top. If we cannot push * the array, we trigger delay marking for the whole object. */ if (!stack.push(endAddr, startAddr, tagged)) delayMarkingChildren(obj); } bool isMarkStackEmpty() { return stack.isEmpty(); } MOZ_MUST_USE bool restoreValueArray(JSObject* obj, void** vpp, void** endp); void saveValueRanges(); inline void processMarkStackTop(SliceBudget& budget); /* The mark stack. Pointers in this stack are "gray" in the GC sense. */ MarkStack stack; /* The color is only applied to objects and functions. */ uint32_t color; /* Pointer to the top of the stack of arenas we are delaying marking on. */ js::gc::Arena* unmarkedArenaStackTop; /* * If the weakKeys table OOMs, disable the linear algorithm and fall back * to iterating until the next GC. */ bool linearWeakMarkingDisabled_; #ifdef DEBUG /* Count of arenas that are currently in the stack. */ size_t markLaterArenas; /* Assert that start and stop are called with correct ordering. */ bool started; /* * If this is true, all marked objects must belong to a compartment being * GCed. This is used to look for compartment bugs. */ bool strictCompartmentChecking; #endif // DEBUG }; #ifdef DEBUG // Return true if this trace is happening on behalf of gray buffering during // the marking phase of incremental GC. bool IsBufferGrayRootsTracer(JSTracer* trc); #endif namespace gc { /*** Special Cases ***/ void PushArena(GCMarker* gcmarker, Arena* arena); /*** Liveness ***/ // Report whether a thing has been marked. Things which are in zones that are // not currently being collected or are owned by another runtime are always // reported as being marked. template bool IsMarkedUnbarriered(JSRuntime* rt, T* thingp); // Report whether a thing has been marked. Things which are in zones that are // not currently being collected or are owned by another runtime are always // reported as being marked. template bool IsMarked(JSRuntime* rt, WriteBarrieredBase* thingp); template bool IsAboutToBeFinalizedUnbarriered(T* thingp); template bool IsAboutToBeFinalized(WriteBarrieredBase* thingp); template bool IsAboutToBeFinalized(ReadBarrieredBase* thingp); bool IsAboutToBeFinalizedDuringSweep(TenuredCell& tenured); inline Cell* ToMarkable(const Value& v) { if (v.isGCThing()) return (Cell*)v.toGCThing(); return nullptr; } inline Cell* ToMarkable(Cell* cell) { return cell; } // Return true if the pointer is nullptr, or if it is a tagged pointer to // nullptr. MOZ_ALWAYS_INLINE bool IsNullTaggedPointer(void* p) { return uintptr_t(p) <= LargestTaggedNullCellPointer; } // Wrap a GC thing pointer into a new Value or jsid. The type system enforces // that the thing pointer is a wrappable type. template struct RewrapTaggedPointer{}; #define DECLARE_REWRAP(S, T, method, prefix) \ template <> struct RewrapTaggedPointer { \ static S wrap(T* thing) { return method ( prefix thing ); } \ } DECLARE_REWRAP(JS::Value, JSObject, JS::ObjectOrNullValue, ); DECLARE_REWRAP(JS::Value, JSString, JS::StringValue, ); DECLARE_REWRAP(JS::Value, JS::Symbol, JS::SymbolValue, ); DECLARE_REWRAP(jsid, JSString, NON_INTEGER_ATOM_TO_JSID, (JSAtom*)); DECLARE_REWRAP(jsid, JS::Symbol, SYMBOL_TO_JSID, ); DECLARE_REWRAP(js::TaggedProto, JSObject, js::TaggedProto, ); #undef DECLARE_REWRAP template struct IsPrivateGCThingInValue : public mozilla::EnableIf::value && !mozilla::IsBaseOf::value && !mozilla::IsBaseOf::value && !mozilla::IsBaseOf::value, T> { static_assert(!mozilla::IsSame::value && !mozilla::IsSame::value, "T must not be Cell or TenuredCell"); }; template struct RewrapTaggedPointer { static Value wrap(typename IsPrivateGCThingInValue::Type* thing) { return JS::PrivateGCThingValue(thing); } }; } /* namespace gc */ // The return value indicates if anything was unmarked. bool UnmarkGrayShapeRecursively(Shape* shape); template void CheckTracedThing(JSTracer* trc, T* thing); template void CheckTracedThing(JSTracer* trc, T thing); } /* namespace js */ #endif /* gc_Marking_h */