summaryrefslogtreecommitdiffstats
path: root/js/src/gc/GCRuntime.h
diff options
context:
space:
mode:
authorMatt A. Tobin <mattatobin@localhost.localdomain>2018-02-02 04:16:08 -0500
committerMatt A. Tobin <mattatobin@localhost.localdomain>2018-02-02 04:16:08 -0500
commit5f8de423f190bbb79a62f804151bc24824fa32d8 (patch)
tree10027f336435511475e392454359edea8e25895d /js/src/gc/GCRuntime.h
parent49ee0794b5d912db1f95dce6eb52d781dc210db5 (diff)
downloadUXP-5f8de423f190bbb79a62f804151bc24824fa32d8.tar
UXP-5f8de423f190bbb79a62f804151bc24824fa32d8.tar.gz
UXP-5f8de423f190bbb79a62f804151bc24824fa32d8.tar.lz
UXP-5f8de423f190bbb79a62f804151bc24824fa32d8.tar.xz
UXP-5f8de423f190bbb79a62f804151bc24824fa32d8.zip
Add m-esr52 at 52.6.0
Diffstat (limited to 'js/src/gc/GCRuntime.h')
-rw-r--r--js/src/gc/GCRuntime.h1467
1 files changed, 1467 insertions, 0 deletions
diff --git a/js/src/gc/GCRuntime.h b/js/src/gc/GCRuntime.h
new file mode 100644
index 000000000..8c9322849
--- /dev/null
+++ b/js/src/gc/GCRuntime.h
@@ -0,0 +1,1467 @@
+/* -*- Mode: C++; tab-width: 8; indent-tabs-mode: nil; c-basic-offset: 4 -*-
+ * vim: set ts=8 sts=4 et sw=4 tw=99:
+ * This Source Code Form is subject to the terms of the Mozilla Public
+ * License, v. 2.0. If a copy of the MPL was not distributed with this
+ * file, You can obtain one at http://mozilla.org/MPL/2.0/. */
+
+#ifndef gc_GCRuntime_h
+#define gc_GCRuntime_h
+
+#include "mozilla/Atomics.h"
+#include "mozilla/EnumSet.h"
+
+#include "jsfriendapi.h"
+#include "jsgc.h"
+
+#include "gc/Heap.h"
+#include "gc/Nursery.h"
+#include "gc/Statistics.h"
+#include "gc/StoreBuffer.h"
+#include "gc/Tracer.h"
+#include "js/GCAnnotations.h"
+
+namespace js {
+
+class AutoLockGC;
+class AutoLockHelperThreadState;
+class VerifyPreTracer;
+
+namespace gc {
+
+typedef Vector<JS::Zone*, 4, SystemAllocPolicy> ZoneVector;
+using BlackGrayEdgeVector = Vector<TenuredCell*, 0, SystemAllocPolicy>;
+
+class AutoMaybeStartBackgroundAllocation;
+class MarkingValidator;
+class AutoTraceSession;
+struct MovingTracer;
+
+class ChunkPool
+{
+ Chunk* head_;
+ size_t count_;
+
+ public:
+ ChunkPool() : head_(nullptr), count_(0) {}
+
+ size_t count() const { return count_; }
+
+ Chunk* head() { MOZ_ASSERT(head_); return head_; }
+ Chunk* pop();
+ void push(Chunk* chunk);
+ Chunk* remove(Chunk* chunk);
+
+#ifdef DEBUG
+ bool contains(Chunk* chunk) const;
+ bool verify() const;
+#endif
+
+ // Pool mutation does not invalidate an Iter unless the mutation
+ // is of the Chunk currently being visited by the Iter.
+ class Iter {
+ public:
+ explicit Iter(ChunkPool& pool) : current_(pool.head_) {}
+ bool done() const { return !current_; }
+ void next();
+ Chunk* get() const { return current_; }
+ operator Chunk*() const { return get(); }
+ Chunk* operator->() const { return get(); }
+ private:
+ Chunk* current_;
+ };
+};
+
+// Performs extra allocation off the main thread so that when memory is
+// required on the main thread it will already be available and waiting.
+class BackgroundAllocTask : public GCParallelTask
+{
+ // Guarded by the GC lock.
+ JSRuntime* runtime;
+ ChunkPool& chunkPool_;
+
+ const bool enabled_;
+
+ public:
+ BackgroundAllocTask(JSRuntime* rt, ChunkPool& pool);
+ bool enabled() const { return enabled_; }
+
+ protected:
+ void run() override;
+};
+
+// Search the provided Chunks for free arenas and decommit them.
+class BackgroundDecommitTask : public GCParallelTask
+{
+ public:
+ using ChunkVector = mozilla::Vector<Chunk*>;
+
+ explicit BackgroundDecommitTask(JSRuntime *rt) : runtime(rt) {}
+ void setChunksToScan(ChunkVector &chunks);
+
+ protected:
+ void run() override;
+
+ private:
+ JSRuntime* runtime;
+ ChunkVector toDecommit;
+};
+
+/*
+ * Encapsulates all of the GC tunables. These are effectively constant and
+ * should only be modified by setParameter.
+ */
+class GCSchedulingTunables
+{
+ /*
+ * Soft limit on the number of bytes we are allowed to allocate in the GC
+ * heap. Attempts to allocate gcthings over this limit will return null and
+ * subsequently invoke the standard OOM machinery, independent of available
+ * physical memory.
+ */
+ size_t gcMaxBytes_;
+
+ /*
+ * The base value used to compute zone->trigger.gcBytes(). When
+ * usage.gcBytes() surpasses threshold.gcBytes() for a zone, the zone may
+ * be scheduled for a GC, depending on the exact circumstances.
+ */
+ size_t gcZoneAllocThresholdBase_;
+
+ /* Fraction of threshold.gcBytes() which triggers an incremental GC. */
+ double zoneAllocThresholdFactor_;
+
+ /*
+ * Number of bytes to allocate between incremental slices in GCs triggered
+ * by the zone allocation threshold.
+ */
+ size_t zoneAllocDelayBytes_;
+
+ /*
+ * Totally disables |highFrequencyGC|, the HeapGrowthFactor, and other
+ * tunables that make GC non-deterministic.
+ */
+ bool dynamicHeapGrowthEnabled_;
+
+ /*
+ * We enter high-frequency mode if we GC a twice within this many
+ * microseconds. This value is stored directly in microseconds.
+ */
+ uint64_t highFrequencyThresholdUsec_;
+
+ /*
+ * When in the |highFrequencyGC| mode, these parameterize the per-zone
+ * "HeapGrowthFactor" computation.
+ */
+ uint64_t highFrequencyLowLimitBytes_;
+ uint64_t highFrequencyHighLimitBytes_;
+ double highFrequencyHeapGrowthMax_;
+ double highFrequencyHeapGrowthMin_;
+
+ /*
+ * When not in |highFrequencyGC| mode, this is the global (stored per-zone)
+ * "HeapGrowthFactor".
+ */
+ double lowFrequencyHeapGrowth_;
+
+ /*
+ * Doubles the length of IGC slices when in the |highFrequencyGC| mode.
+ */
+ bool dynamicMarkSliceEnabled_;
+
+ /*
+ * Controls whether painting can trigger IGC slices.
+ */
+ bool refreshFrameSlicesEnabled_;
+
+ /*
+ * Controls the number of empty chunks reserved for future allocation.
+ */
+ uint32_t minEmptyChunkCount_;
+ uint32_t maxEmptyChunkCount_;
+
+ public:
+ GCSchedulingTunables()
+ : gcMaxBytes_(0),
+ gcZoneAllocThresholdBase_(30 * 1024 * 1024),
+ zoneAllocThresholdFactor_(0.9),
+ zoneAllocDelayBytes_(1024 * 1024),
+ dynamicHeapGrowthEnabled_(false),
+ highFrequencyThresholdUsec_(1000 * 1000),
+ highFrequencyLowLimitBytes_(100 * 1024 * 1024),
+ highFrequencyHighLimitBytes_(500 * 1024 * 1024),
+ highFrequencyHeapGrowthMax_(3.0),
+ highFrequencyHeapGrowthMin_(1.5),
+ lowFrequencyHeapGrowth_(1.5),
+ dynamicMarkSliceEnabled_(false),
+ refreshFrameSlicesEnabled_(true),
+ minEmptyChunkCount_(1),
+ maxEmptyChunkCount_(30)
+ {}
+
+ size_t gcMaxBytes() const { return gcMaxBytes_; }
+ size_t gcZoneAllocThresholdBase() const { return gcZoneAllocThresholdBase_; }
+ double zoneAllocThresholdFactor() const { return zoneAllocThresholdFactor_; }
+ size_t zoneAllocDelayBytes() const { return zoneAllocDelayBytes_; }
+ bool isDynamicHeapGrowthEnabled() const { return dynamicHeapGrowthEnabled_; }
+ uint64_t highFrequencyThresholdUsec() const { return highFrequencyThresholdUsec_; }
+ uint64_t highFrequencyLowLimitBytes() const { return highFrequencyLowLimitBytes_; }
+ uint64_t highFrequencyHighLimitBytes() const { return highFrequencyHighLimitBytes_; }
+ double highFrequencyHeapGrowthMax() const { return highFrequencyHeapGrowthMax_; }
+ double highFrequencyHeapGrowthMin() const { return highFrequencyHeapGrowthMin_; }
+ double lowFrequencyHeapGrowth() const { return lowFrequencyHeapGrowth_; }
+ bool isDynamicMarkSliceEnabled() const { return dynamicMarkSliceEnabled_; }
+ bool areRefreshFrameSlicesEnabled() const { return refreshFrameSlicesEnabled_; }
+ unsigned minEmptyChunkCount(const AutoLockGC&) const { return minEmptyChunkCount_; }
+ unsigned maxEmptyChunkCount() const { return maxEmptyChunkCount_; }
+
+ MOZ_MUST_USE bool setParameter(JSGCParamKey key, uint32_t value, const AutoLockGC& lock);
+};
+
+/*
+ * GC Scheduling Overview
+ * ======================
+ *
+ * Scheduling GC's in SpiderMonkey/Firefox is tremendously complicated because
+ * of the large number of subtle, cross-cutting, and widely dispersed factors
+ * that must be taken into account. A summary of some of the more important
+ * factors follows.
+ *
+ * Cost factors:
+ *
+ * * GC too soon and we'll revisit an object graph almost identical to the
+ * one we just visited; since we are unlikely to find new garbage, the
+ * traversal will be largely overhead. We rely heavily on external factors
+ * to signal us that we are likely to find lots of garbage: e.g. "a tab
+ * just got closed".
+ *
+ * * GC too late and we'll run out of memory to allocate (e.g. Out-Of-Memory,
+ * hereafter simply abbreviated to OOM). If this happens inside
+ * SpiderMonkey we may be able to recover, but most embedder allocations
+ * will simply crash on OOM, even if the GC has plenty of free memory it
+ * could surrender.
+ *
+ * * Memory fragmentation: if we fill the process with GC allocations, a
+ * request for a large block of contiguous memory may fail because no
+ * contiguous block is free, despite having enough memory available to
+ * service the request.
+ *
+ * * Management overhead: if our GC heap becomes large, we create extra
+ * overhead when managing the GC's structures, even if the allocations are
+ * mostly unused.
+ *
+ * Heap Management Factors:
+ *
+ * * GC memory: The GC has its own allocator that it uses to make fixed size
+ * allocations for GC managed things. In cases where the GC thing requires
+ * larger or variable sized memory to implement itself, it is responsible
+ * for using the system heap.
+ *
+ * * C Heap Memory: Rather than allowing for large or variable allocations,
+ * the SpiderMonkey GC allows GC things to hold pointers to C heap memory.
+ * It is the responsibility of the thing to free this memory with a custom
+ * finalizer (with the sole exception of NativeObject, which knows about
+ * slots and elements for performance reasons). C heap memory has different
+ * performance and overhead tradeoffs than GC internal memory, which need
+ * to be considered with scheduling a GC.
+ *
+ * Application Factors:
+ *
+ * * Most applications allocate heavily at startup, then enter a processing
+ * stage where memory utilization remains roughly fixed with a slower
+ * allocation rate. This is not always the case, however, so while we may
+ * optimize for this pattern, we must be able to handle arbitrary
+ * allocation patterns.
+ *
+ * Other factors:
+ *
+ * * Other memory: This is memory allocated outside the purview of the GC.
+ * Data mapped by the system for code libraries, data allocated by those
+ * libraries, data in the JSRuntime that is used to manage the engine,
+ * memory used by the embedding that is not attached to a GC thing, memory
+ * used by unrelated processes running on the hardware that use space we
+ * could otherwise use for allocation, etc. While we don't have to manage
+ * it, we do have to take it into account when scheduling since it affects
+ * when we will OOM.
+ *
+ * * Physical Reality: All real machines have limits on the number of bits
+ * that they are physically able to store. While modern operating systems
+ * can generally make additional space available with swapping, at some
+ * point there are simply no more bits to allocate. There is also the
+ * factor of address space limitations, particularly on 32bit machines.
+ *
+ * * Platform Factors: Each OS makes use of wildly different memory
+ * management techniques. These differences result in different performance
+ * tradeoffs, different fragmentation patterns, and different hard limits
+ * on the amount of physical and/or virtual memory that we can use before
+ * OOMing.
+ *
+ *
+ * Reasons for scheduling GC
+ * -------------------------
+ *
+ * While code generally takes the above factors into account in only an ad-hoc
+ * fashion, the API forces the user to pick a "reason" for the GC. We have a
+ * bunch of JS::gcreason reasons in GCAPI.h. These fall into a few categories
+ * that generally coincide with one or more of the above factors.
+ *
+ * Embedding reasons:
+ *
+ * 1) Do a GC now because the embedding knows something useful about the
+ * zone's memory retention state. These are gcreasons like LOAD_END,
+ * PAGE_HIDE, SET_NEW_DOCUMENT, DOM_UTILS. Mostly, Gecko uses these to
+ * indicate that a significant fraction of the scheduled zone's memory is
+ * probably reclaimable.
+ *
+ * 2) Do some known amount of GC work now because the embedding knows now is
+ * a good time to do a long, unblockable operation of a known duration.
+ * These are INTER_SLICE_GC and REFRESH_FRAME.
+ *
+ * Correctness reasons:
+ *
+ * 3) Do a GC now because correctness depends on some GC property. For
+ * example, CC_WAITING is where the embedding requires the mark bits
+ * to be set correct. Also, EVICT_NURSERY where we need to work on the tenured
+ * heap.
+ *
+ * 4) Do a GC because we are shutting down: e.g. SHUTDOWN_CC or DESTROY_*.
+ *
+ * 5) Do a GC because a compartment was accessed between GC slices when we
+ * would have otherwise discarded it. We have to do a second GC to clean
+ * it up: e.g. COMPARTMENT_REVIVED.
+ *
+ * Emergency Reasons:
+ *
+ * 6) Do an all-zones, non-incremental GC now because the embedding knows it
+ * cannot wait: e.g. MEM_PRESSURE.
+ *
+ * 7) OOM when fetching a new Chunk results in a LAST_DITCH GC.
+ *
+ * Heap Size Limitation Reasons:
+ *
+ * 8) Do an incremental, zonal GC with reason MAYBEGC when we discover that
+ * the gc's allocated size is approaching the current trigger. This is
+ * called MAYBEGC because we make this check in the MaybeGC function.
+ * MaybeGC gets called at the top of the main event loop. Normally, it is
+ * expected that this callback will keep the heap size limited. It is
+ * relatively inexpensive, because it is invoked with no JS running and
+ * thus few stack roots to scan. For this reason, the GC's "trigger" bytes
+ * is less than the GC's "max" bytes as used by the trigger below.
+ *
+ * 9) Do an incremental, zonal GC with reason MAYBEGC when we go to allocate
+ * a new GC thing and find that the GC heap size has grown beyond the
+ * configured maximum (JSGC_MAX_BYTES). We trigger this GC by returning
+ * nullptr and then calling maybeGC at the top level of the allocator.
+ * This is then guaranteed to fail the "size greater than trigger" check
+ * above, since trigger is always less than max. After performing the GC,
+ * the allocator unconditionally returns nullptr to force an OOM exception
+ * is raised by the script.
+ *
+ * Note that this differs from a LAST_DITCH GC where we actually run out
+ * of memory (i.e., a call to a system allocator fails) when trying to
+ * allocate. Unlike above, LAST_DITCH GC only happens when we are really
+ * out of memory, not just when we cross an arbitrary trigger; despite
+ * this, it may still return an allocation at the end and allow the script
+ * to continue, if the LAST_DITCH GC was able to free up enough memory.
+ *
+ * 10) Do a GC under reason ALLOC_TRIGGER when we are over the GC heap trigger
+ * limit, but in the allocator rather than in a random call to maybeGC.
+ * This occurs if we allocate too much before returning to the event loop
+ * and calling maybeGC; this is extremely common in benchmarks and
+ * long-running Worker computations. Note that this uses a wildly
+ * different mechanism from the above in that it sets the interrupt flag
+ * and does the GC at the next loop head, before the next alloc, or
+ * maybeGC. The reason for this is that this check is made after the
+ * allocation and we cannot GC with an uninitialized thing in the heap.
+ *
+ * 11) Do an incremental, zonal GC with reason TOO_MUCH_MALLOC when we have
+ * malloced more than JSGC_MAX_MALLOC_BYTES in a zone since the last GC.
+ *
+ *
+ * Size Limitation Triggers Explanation
+ * ------------------------------------
+ *
+ * The GC internally is entirely unaware of the context of the execution of
+ * the mutator. It sees only:
+ *
+ * A) Allocated size: this is the amount of memory currently requested by the
+ * mutator. This quantity is monotonically increasing: i.e. the allocation
+ * rate is always >= 0. It is also easy for the system to track.
+ *
+ * B) Retained size: this is the amount of memory that the mutator can
+ * currently reach. Said another way, it is the size of the heap
+ * immediately after a GC (modulo background sweeping). This size is very
+ * costly to know exactly and also extremely hard to estimate with any
+ * fidelity.
+ *
+ * For reference, a common allocated vs. retained graph might look like:
+ *
+ * | ** **
+ * | ** ** * **
+ * | ** * ** * **
+ * | * ** * ** * **
+ * | ** ** * ** * **
+ * s| * * ** ** + + **
+ * i| * * * + + + + +
+ * z| * * * + + + + +
+ * e| * **+
+ * | * +
+ * | * +
+ * | * +
+ * | * +
+ * | * +
+ * |*+
+ * +--------------------------------------------------
+ * time
+ * *** = allocated
+ * +++ = retained
+ *
+ * Note that this is a bit of a simplification
+ * because in reality we track malloc and GC heap
+ * sizes separately and have a different level of
+ * granularity and accuracy on each heap.
+ *
+ * This presents some obvious implications for Mark-and-Sweep collectors.
+ * Namely:
+ * -> t[marking] ~= size[retained]
+ * -> t[sweeping] ~= size[allocated] - size[retained]
+ *
+ * In a non-incremental collector, maintaining low latency and high
+ * responsiveness requires that total GC times be as low as possible. Thus,
+ * in order to stay responsive when we did not have a fully incremental
+ * collector, our GC triggers were focused on minimizing collection time.
+ * Furthermore, since size[retained] is not under control of the GC, all the
+ * GC could do to control collection times was reduce sweep times by
+ * minimizing size[allocated], per the equation above.
+ *
+ * The result of the above is GC triggers that focus on size[allocated] to
+ * the exclusion of other important factors and default heuristics that are
+ * not optimal for a fully incremental collector. On the other hand, this is
+ * not all bad: minimizing size[allocated] also minimizes the chance of OOM
+ * and sweeping remains one of the hardest areas to further incrementalize.
+ *
+ * EAGER_ALLOC_TRIGGER
+ * -------------------
+ * Occurs when we return to the event loop and find our heap is getting
+ * largish, but before t[marking] OR t[sweeping] is too large for a
+ * responsive non-incremental GC. This is intended to be the common case
+ * in normal web applications: e.g. we just finished an event handler and
+ * the few objects we allocated when computing the new whatzitz have
+ * pushed us slightly over the limit. After this GC we rescale the new
+ * EAGER_ALLOC_TRIGGER trigger to 150% of size[retained] so that our
+ * non-incremental GC times will always be proportional to this size
+ * rather than being dominated by sweeping.
+ *
+ * As a concession to mutators that allocate heavily during their startup
+ * phase, we have a highFrequencyGCMode that ups the growth rate to 300%
+ * of the current size[retained] so that we'll do fewer longer GCs at the
+ * end of the mutator startup rather than more, smaller GCs.
+ *
+ * Assumptions:
+ * -> Responsiveness is proportional to t[marking] + t[sweeping].
+ * -> size[retained] is proportional only to GC allocations.
+ *
+ * ALLOC_TRIGGER (non-incremental)
+ * -------------------------------
+ * If we do not return to the event loop before getting all the way to our
+ * gc trigger bytes then MAYBEGC will never fire. To avoid OOMing, we
+ * succeed the current allocation and set the script interrupt so that we
+ * will (hopefully) do a GC before we overflow our max and have to raise
+ * an OOM exception for the script.
+ *
+ * Assumptions:
+ * -> Common web scripts will return to the event loop before using
+ * 10% of the current gcTriggerBytes worth of GC memory.
+ *
+ * ALLOC_TRIGGER (incremental)
+ * ---------------------------
+ * In practice the above trigger is rough: if a website is just on the
+ * cusp, sometimes it will trigger a non-incremental GC moments before
+ * returning to the event loop, where it could have done an incremental
+ * GC. Thus, we recently added an incremental version of the above with a
+ * substantially lower threshold, so that we have a soft limit here. If
+ * IGC can collect faster than the allocator generates garbage, even if
+ * the allocator does not return to the event loop frequently, we should
+ * not have to fall back to a non-incremental GC.
+ *
+ * INCREMENTAL_TOO_SLOW
+ * --------------------
+ * Do a full, non-incremental GC if we overflow ALLOC_TRIGGER during an
+ * incremental GC. When in the middle of an incremental GC, we suppress
+ * our other triggers, so we need a way to backstop the IGC if the
+ * mutator allocates faster than the IGC can clean things up.
+ *
+ * TOO_MUCH_MALLOC
+ * ---------------
+ * Performs a GC before size[allocated] - size[retained] gets too large
+ * for non-incremental sweeping to be fast in the case that we have
+ * significantly more malloc allocation than GC allocation. This is meant
+ * to complement MAYBEGC triggers. We track this by counting malloced
+ * bytes; the counter gets reset at every GC since we do not always have a
+ * size at the time we call free. Because of this, the malloc heuristic
+ * is, unfortunatly, not usefully able to augment our other GC heap
+ * triggers and is limited to this singular heuristic.
+ *
+ * Assumptions:
+ * -> EITHER size[allocated_by_malloc] ~= size[allocated_by_GC]
+ * OR time[sweeping] ~= size[allocated_by_malloc]
+ * -> size[retained] @ t0 ~= size[retained] @ t1
+ * i.e. That the mutator is in steady-state operation.
+ *
+ * LAST_DITCH_GC
+ * -------------
+ * Does a GC because we are out of memory.
+ *
+ * Assumptions:
+ * -> size[retained] < size[available_memory]
+ */
+class GCSchedulingState
+{
+ /*
+ * Influences how we schedule and run GC's in several subtle ways. The most
+ * important factor is in how it controls the "HeapGrowthFactor". The
+ * growth factor is a measure of how large (as a percentage of the last GC)
+ * the heap is allowed to grow before we try to schedule another GC.
+ */
+ bool inHighFrequencyGCMode_;
+
+ public:
+ GCSchedulingState()
+ : inHighFrequencyGCMode_(false)
+ {}
+
+ bool inHighFrequencyGCMode() const { return inHighFrequencyGCMode_; }
+
+ void updateHighFrequencyMode(uint64_t lastGCTime, uint64_t currentTime,
+ const GCSchedulingTunables& tunables) {
+ inHighFrequencyGCMode_ =
+ tunables.isDynamicHeapGrowthEnabled() && lastGCTime &&
+ lastGCTime + tunables.highFrequencyThresholdUsec() > currentTime;
+ }
+};
+
+template<typename F>
+struct Callback {
+ F op;
+ void* data;
+
+ Callback()
+ : op(nullptr), data(nullptr)
+ {}
+ Callback(F op, void* data)
+ : op(op), data(data)
+ {}
+};
+
+template<typename F>
+using CallbackVector = Vector<Callback<F>, 4, SystemAllocPolicy>;
+
+template <typename T, typename Iter0, typename Iter1>
+class ChainedIter
+{
+ Iter0 iter0_;
+ Iter1 iter1_;
+
+ public:
+ ChainedIter(const Iter0& iter0, const Iter1& iter1)
+ : iter0_(iter0), iter1_(iter1)
+ {}
+
+ bool done() const { return iter0_.done() && iter1_.done(); }
+ void next() {
+ MOZ_ASSERT(!done());
+ if (!iter0_.done()) {
+ iter0_.next();
+ } else {
+ MOZ_ASSERT(!iter1_.done());
+ iter1_.next();
+ }
+ }
+ T get() const {
+ MOZ_ASSERT(!done());
+ if (!iter0_.done())
+ return iter0_.get();
+ MOZ_ASSERT(!iter1_.done());
+ return iter1_.get();
+ }
+
+ operator T() const { return get(); }
+ T operator->() const { return get(); }
+};
+
+typedef HashMap<Value*, const char*, DefaultHasher<Value*>, SystemAllocPolicy> RootedValueMap;
+
+using AllocKinds = mozilla::EnumSet<AllocKind>;
+
+class GCRuntime
+{
+ public:
+ explicit GCRuntime(JSRuntime* rt);
+ MOZ_MUST_USE bool init(uint32_t maxbytes, uint32_t maxNurseryBytes);
+ void finishRoots();
+ void finish();
+
+ inline bool hasZealMode(ZealMode mode);
+ inline void clearZealMode(ZealMode mode);
+ inline bool upcomingZealousGC();
+ inline bool needZealousGC();
+
+ MOZ_MUST_USE bool addRoot(Value* vp, const char* name);
+ void removeRoot(Value* vp);
+ void setMarkStackLimit(size_t limit, AutoLockGC& lock);
+
+ MOZ_MUST_USE bool setParameter(JSGCParamKey key, uint32_t value, AutoLockGC& lock);
+ uint32_t getParameter(JSGCParamKey key, const AutoLockGC& lock);
+
+ MOZ_MUST_USE bool triggerGC(JS::gcreason::Reason reason);
+ void maybeAllocTriggerZoneGC(Zone* zone, const AutoLockGC& lock);
+ // The return value indicates if we were able to do the GC.
+ bool triggerZoneGC(Zone* zone, JS::gcreason::Reason reason);
+ void maybeGC(Zone* zone);
+ void minorGC(JS::gcreason::Reason reason,
+ gcstats::Phase phase = gcstats::PHASE_MINOR_GC) JS_HAZ_GC_CALL;
+ void evictNursery(JS::gcreason::Reason reason = JS::gcreason::EVICT_NURSERY) {
+ minorGC(reason, gcstats::PHASE_EVICT_NURSERY);
+ }
+ // The return value indicates whether a major GC was performed.
+ bool gcIfRequested();
+ void gc(JSGCInvocationKind gckind, JS::gcreason::Reason reason);
+ void startGC(JSGCInvocationKind gckind, JS::gcreason::Reason reason, int64_t millis = 0);
+ void gcSlice(JS::gcreason::Reason reason, int64_t millis = 0);
+ void finishGC(JS::gcreason::Reason reason);
+ void abortGC();
+ void startDebugGC(JSGCInvocationKind gckind, SliceBudget& budget);
+ void debugGCSlice(SliceBudget& budget);
+
+ void triggerFullGCForAtoms() {
+ MOZ_ASSERT(fullGCForAtomsRequested_);
+ fullGCForAtomsRequested_ = false;
+ MOZ_RELEASE_ASSERT(triggerGC(JS::gcreason::ALLOC_TRIGGER));
+ }
+
+ void runDebugGC();
+ inline void poke();
+
+ enum TraceOrMarkRuntime {
+ TraceRuntime,
+ MarkRuntime
+ };
+ void traceRuntime(JSTracer* trc, AutoLockForExclusiveAccess& lock);
+ void traceRuntimeForMinorGC(JSTracer* trc, AutoLockForExclusiveAccess& lock);
+
+ void notifyDidPaint();
+ void shrinkBuffers();
+ void onOutOfMallocMemory();
+ void onOutOfMallocMemory(const AutoLockGC& lock);
+
+#ifdef JS_GC_ZEAL
+ const void* addressOfZealModeBits() { return &zealModeBits; }
+ void getZealBits(uint32_t* zealBits, uint32_t* frequency, uint32_t* nextScheduled);
+ void setZeal(uint8_t zeal, uint32_t frequency);
+ bool parseAndSetZeal(const char* str);
+ void setNextScheduled(uint32_t count);
+ void verifyPreBarriers();
+ void maybeVerifyPreBarriers(bool always);
+ bool selectForMarking(JSObject* object);
+ void clearSelectedForMarking();
+ void setDeterministic(bool enable);
+#endif
+
+ size_t maxMallocBytesAllocated() { return maxMallocBytes; }
+
+ uint64_t nextCellUniqueId() {
+ MOZ_ASSERT(nextCellUniqueId_ > 0);
+ uint64_t uid = ++nextCellUniqueId_;
+ return uid;
+ }
+
+#ifdef DEBUG
+ bool shutdownCollectedEverything() const {
+ return arenasEmptyAtShutdown;
+ }
+#endif
+
+ public:
+ // Internal public interface
+ State state() const { return incrementalState; }
+ bool isHeapCompacting() const { return state() == State::Compact; }
+ bool isForegroundSweeping() const { return state() == State::Sweep; }
+ bool isBackgroundSweeping() { return helperState.isBackgroundSweeping(); }
+ void waitBackgroundSweepEnd() { helperState.waitBackgroundSweepEnd(); }
+ void waitBackgroundSweepOrAllocEnd() {
+ helperState.waitBackgroundSweepEnd();
+ allocTask.cancel(GCParallelTask::CancelAndWait);
+ }
+
+ void requestMinorGC(JS::gcreason::Reason reason);
+
+#ifdef DEBUG
+ bool onBackgroundThread() { return helperState.onBackgroundThread(); }
+#endif // DEBUG
+
+ void lockGC() {
+ lock.lock();
+ }
+
+ void unlockGC() {
+ lock.unlock();
+ }
+
+#ifdef DEBUG
+ bool isAllocAllowed() { return noGCOrAllocationCheck == 0; }
+ void disallowAlloc() { ++noGCOrAllocationCheck; }
+ void allowAlloc() {
+ MOZ_ASSERT(!isAllocAllowed());
+ --noGCOrAllocationCheck;
+ }
+
+ bool isNurseryAllocAllowed() { return noNurseryAllocationCheck == 0; }
+ void disallowNurseryAlloc() { ++noNurseryAllocationCheck; }
+ void allowNurseryAlloc() {
+ MOZ_ASSERT(!isNurseryAllocAllowed());
+ --noNurseryAllocationCheck;
+ }
+
+ bool isStrictProxyCheckingEnabled() { return disableStrictProxyCheckingCount == 0; }
+ void disableStrictProxyChecking() { ++disableStrictProxyCheckingCount; }
+ void enableStrictProxyChecking() {
+ MOZ_ASSERT(disableStrictProxyCheckingCount > 0);
+ --disableStrictProxyCheckingCount;
+ }
+#endif // DEBUG
+
+ bool isInsideUnsafeRegion() { return inUnsafeRegion != 0; }
+ void enterUnsafeRegion() { ++inUnsafeRegion; }
+ void leaveUnsafeRegion() {
+ MOZ_ASSERT(inUnsafeRegion > 0);
+ --inUnsafeRegion;
+ }
+
+ void verifyIsSafeToGC() {
+ MOZ_DIAGNOSTIC_ASSERT(!isInsideUnsafeRegion(),
+ "[AutoAssertNoGC] possible GC in GC-unsafe region");
+ }
+
+ void setAlwaysPreserveCode() { alwaysPreserveCode = true; }
+
+ bool isIncrementalGCAllowed() const { return incrementalAllowed; }
+ void disallowIncrementalGC() { incrementalAllowed = false; }
+
+ bool isIncrementalGCEnabled() const { return mode == JSGC_MODE_INCREMENTAL && incrementalAllowed; }
+ bool isIncrementalGCInProgress() const { return state() != State::NotActive; }
+
+ bool isGenerationalGCEnabled() const { return generationalDisabled == 0; }
+ void disableGenerationalGC();
+ void enableGenerationalGC();
+
+ void disableCompactingGC();
+ void enableCompactingGC();
+ bool isCompactingGCEnabled() const;
+
+ void setGrayRootsTracer(JSTraceDataOp traceOp, void* data);
+ MOZ_MUST_USE bool addBlackRootsTracer(JSTraceDataOp traceOp, void* data);
+ void removeBlackRootsTracer(JSTraceDataOp traceOp, void* data);
+
+ void setMaxMallocBytes(size_t value);
+ int32_t getMallocBytes() const { return mallocBytesUntilGC; }
+ void resetMallocBytes();
+ bool isTooMuchMalloc() const { return mallocBytesUntilGC <= 0; }
+ void updateMallocCounter(JS::Zone* zone, size_t nbytes);
+ void onTooMuchMalloc();
+
+ void setGCCallback(JSGCCallback callback, void* data);
+ void callGCCallback(JSGCStatus status) const;
+ void setObjectsTenuredCallback(JSObjectsTenuredCallback callback,
+ void* data);
+ void callObjectsTenuredCallback();
+ MOZ_MUST_USE bool addFinalizeCallback(JSFinalizeCallback callback, void* data);
+ void removeFinalizeCallback(JSFinalizeCallback func);
+ MOZ_MUST_USE bool addWeakPointerZoneGroupCallback(JSWeakPointerZoneGroupCallback callback,
+ void* data);
+ void removeWeakPointerZoneGroupCallback(JSWeakPointerZoneGroupCallback callback);
+ MOZ_MUST_USE bool addWeakPointerCompartmentCallback(JSWeakPointerCompartmentCallback callback,
+ void* data);
+ void removeWeakPointerCompartmentCallback(JSWeakPointerCompartmentCallback callback);
+ JS::GCSliceCallback setSliceCallback(JS::GCSliceCallback callback);
+ JS::GCNurseryCollectionCallback setNurseryCollectionCallback(
+ JS::GCNurseryCollectionCallback callback);
+ JS::DoCycleCollectionCallback setDoCycleCollectionCallback(JS::DoCycleCollectionCallback callback);
+ void callDoCycleCollectionCallback(JSContext* cx);
+
+ void setFullCompartmentChecks(bool enable);
+
+ bool isManipulatingDeadZones() { return manipulatingDeadZones; }
+ void setManipulatingDeadZones(bool value) { manipulatingDeadZones = value; }
+ unsigned objectsMarkedInDeadZonesCount() { return objectsMarkedInDeadZones; }
+ void incObjectsMarkedInDeadZone() {
+ MOZ_ASSERT(manipulatingDeadZones);
+ ++objectsMarkedInDeadZones;
+ }
+
+ JS::Zone* getCurrentZoneGroup() { return currentZoneGroup; }
+ void setFoundBlackGrayEdges(TenuredCell& target) {
+ AutoEnterOOMUnsafeRegion oomUnsafe;
+ if (!foundBlackGrayEdges.append(&target))
+ oomUnsafe.crash("OOM|small: failed to insert into foundBlackGrayEdges");
+ }
+
+ uint64_t gcNumber() const { return number; }
+
+ uint64_t minorGCCount() const { return minorGCNumber; }
+ void incMinorGcNumber() { ++minorGCNumber; ++number; }
+
+ uint64_t majorGCCount() const { return majorGCNumber; }
+ void incMajorGcNumber() { ++majorGCNumber; ++number; }
+
+ int64_t defaultSliceBudget() const { return defaultTimeBudget_; }
+
+ bool isIncrementalGc() const { return isIncremental; }
+ bool isFullGc() const { return isFull; }
+ bool isCompactingGc() const { return isCompacting; }
+
+ bool minorGCRequested() const { return minorGCTriggerReason != JS::gcreason::NO_REASON; }
+ bool majorGCRequested() const { return majorGCTriggerReason != JS::gcreason::NO_REASON; }
+ bool isGcNeeded() { return minorGCRequested() || majorGCRequested(); }
+
+ bool fullGCForAtomsRequested() const { return fullGCForAtomsRequested_; }
+
+ double computeHeapGrowthFactor(size_t lastBytes);
+ size_t computeTriggerBytes(double growthFactor, size_t lastBytes);
+
+ JSGCMode gcMode() const { return mode; }
+ void setGCMode(JSGCMode m) {
+ mode = m;
+ marker.setGCMode(mode);
+ }
+
+ inline void updateOnFreeArenaAlloc(const ChunkInfo& info);
+ inline void updateOnArenaFree(const ChunkInfo& info);
+
+ ChunkPool& fullChunks(const AutoLockGC& lock) { return fullChunks_; }
+ ChunkPool& availableChunks(const AutoLockGC& lock) { return availableChunks_; }
+ ChunkPool& emptyChunks(const AutoLockGC& lock) { return emptyChunks_; }
+ const ChunkPool& fullChunks(const AutoLockGC& lock) const { return fullChunks_; }
+ const ChunkPool& availableChunks(const AutoLockGC& lock) const { return availableChunks_; }
+ const ChunkPool& emptyChunks(const AutoLockGC& lock) const { return emptyChunks_; }
+ typedef ChainedIter<Chunk*, ChunkPool::Iter, ChunkPool::Iter> NonEmptyChunksIter;
+ NonEmptyChunksIter allNonEmptyChunks() {
+ return NonEmptyChunksIter(ChunkPool::Iter(availableChunks_), ChunkPool::Iter(fullChunks_));
+ }
+
+ Chunk* getOrAllocChunk(const AutoLockGC& lock,
+ AutoMaybeStartBackgroundAllocation& maybeStartBGAlloc);
+ void recycleChunk(Chunk* chunk, const AutoLockGC& lock);
+
+#ifdef JS_GC_ZEAL
+ void startVerifyPreBarriers();
+ void endVerifyPreBarriers();
+ void finishVerifier();
+ bool isVerifyPreBarriersEnabled() const { return !!verifyPreData; }
+#else
+ bool isVerifyPreBarriersEnabled() const { return false; }
+#endif
+
+ // Free certain LifoAlloc blocks when it is safe to do so.
+ void freeUnusedLifoBlocksAfterSweeping(LifoAlloc* lifo);
+ void freeAllLifoBlocksAfterSweeping(LifoAlloc* lifo);
+ void freeAllLifoBlocksAfterMinorGC(LifoAlloc* lifo);
+
+ // Queue a thunk to run after the next minor GC.
+ void callAfterMinorGC(void (*thunk)(void* data), void* data) {
+ nursery.queueSweepAction(thunk, data);
+ }
+
+ // Public here for ReleaseArenaLists and FinalizeTypedArenas.
+ void releaseArena(Arena* arena, const AutoLockGC& lock);
+
+ void releaseHeldRelocatedArenas();
+ void releaseHeldRelocatedArenasWithoutUnlocking(const AutoLockGC& lock);
+
+ // Allocator
+ template <AllowGC allowGC>
+ MOZ_MUST_USE bool checkAllocatorState(JSContext* cx, AllocKind kind);
+ template <AllowGC allowGC>
+ JSObject* tryNewNurseryObject(JSContext* cx, size_t thingSize, size_t nDynamicSlots,
+ const Class* clasp);
+ template <AllowGC allowGC>
+ static JSObject* tryNewTenuredObject(ExclusiveContext* cx, AllocKind kind, size_t thingSize,
+ size_t nDynamicSlots);
+ template <typename T, AllowGC allowGC>
+ static T* tryNewTenuredThing(ExclusiveContext* cx, AllocKind kind, size_t thingSize);
+ static TenuredCell* refillFreeListInGC(Zone* zone, AllocKind thingKind);
+
+ private:
+ enum IncrementalProgress
+ {
+ NotFinished = 0,
+ Finished
+ };
+
+ // For ArenaLists::allocateFromArena()
+ friend class ArenaLists;
+ Chunk* pickChunk(const AutoLockGC& lock,
+ AutoMaybeStartBackgroundAllocation& maybeStartBGAlloc);
+ Arena* allocateArena(Chunk* chunk, Zone* zone, AllocKind kind,
+ ShouldCheckThresholds checkThresholds, const AutoLockGC& lock);
+ void arenaAllocatedDuringGC(JS::Zone* zone, Arena* arena);
+
+ // Allocator internals
+ MOZ_MUST_USE bool gcIfNeededPerAllocation(JSContext* cx);
+ template <typename T>
+ static void checkIncrementalZoneState(ExclusiveContext* cx, T* t);
+ static TenuredCell* refillFreeListFromAnyThread(ExclusiveContext* cx, AllocKind thingKind,
+ size_t thingSize);
+ static TenuredCell* refillFreeListFromMainThread(JSContext* cx, AllocKind thingKind,
+ size_t thingSize);
+ static TenuredCell* refillFreeListOffMainThread(ExclusiveContext* cx, AllocKind thingKind);
+
+ /*
+ * Return the list of chunks that can be released outside the GC lock.
+ * Must be called either during the GC or with the GC lock taken.
+ */
+ friend class BackgroundDecommitTask;
+ ChunkPool expireEmptyChunkPool(const AutoLockGC& lock);
+ void freeEmptyChunks(JSRuntime* rt, const AutoLockGC& lock);
+ void prepareToFreeChunk(ChunkInfo& info);
+
+ friend class BackgroundAllocTask;
+ friend class AutoMaybeStartBackgroundAllocation;
+ bool wantBackgroundAllocation(const AutoLockGC& lock) const;
+ void startBackgroundAllocTaskIfIdle();
+
+ void requestMajorGC(JS::gcreason::Reason reason);
+ SliceBudget defaultBudget(JS::gcreason::Reason reason, int64_t millis);
+ void budgetIncrementalGC(SliceBudget& budget, AutoLockForExclusiveAccess& lock);
+ void resetIncrementalGC(AbortReason reason, AutoLockForExclusiveAccess& lock);
+
+ // Assert if the system state is such that we should never
+ // receive a request to do GC work.
+ void checkCanCallAPI();
+
+ // Check if the system state is such that GC has been supressed
+ // or otherwise delayed.
+ MOZ_MUST_USE bool checkIfGCAllowedInCurrentState(JS::gcreason::Reason reason);
+
+ gcstats::ZoneGCStats scanZonesBeforeGC();
+ void collect(bool nonincrementalByAPI, SliceBudget budget, JS::gcreason::Reason reason) JS_HAZ_GC_CALL;
+ MOZ_MUST_USE bool gcCycle(bool nonincrementalByAPI, SliceBudget& budget,
+ JS::gcreason::Reason reason);
+ void incrementalCollectSlice(SliceBudget& budget, JS::gcreason::Reason reason,
+ AutoLockForExclusiveAccess& lock);
+
+ void pushZealSelectedObjects();
+ void purgeRuntime(AutoLockForExclusiveAccess& lock);
+ MOZ_MUST_USE bool beginMarkPhase(JS::gcreason::Reason reason, AutoLockForExclusiveAccess& lock);
+ bool shouldPreserveJITCode(JSCompartment* comp, int64_t currentTime,
+ JS::gcreason::Reason reason, bool canAllocateMoreCode);
+ void traceRuntimeForMajorGC(JSTracer* trc, AutoLockForExclusiveAccess& lock);
+ void traceRuntimeAtoms(JSTracer* trc, AutoLockForExclusiveAccess& lock);
+ void traceRuntimeCommon(JSTracer* trc, TraceOrMarkRuntime traceOrMark,
+ AutoLockForExclusiveAccess& lock);
+ void bufferGrayRoots();
+ void maybeDoCycleCollection();
+ void markCompartments();
+ IncrementalProgress drainMarkStack(SliceBudget& sliceBudget, gcstats::Phase phase);
+ template <class CompartmentIterT> void markWeakReferences(gcstats::Phase phase);
+ void markWeakReferencesInCurrentGroup(gcstats::Phase phase);
+ template <class ZoneIterT, class CompartmentIterT> void markGrayReferences(gcstats::Phase phase);
+ void markBufferedGrayRoots(JS::Zone* zone);
+ void markGrayReferencesInCurrentGroup(gcstats::Phase phase);
+ void markAllWeakReferences(gcstats::Phase phase);
+ void markAllGrayReferences(gcstats::Phase phase);
+
+ void beginSweepPhase(bool lastGC, AutoLockForExclusiveAccess& lock);
+ void findZoneGroups(AutoLockForExclusiveAccess& lock);
+ MOZ_MUST_USE bool findInterZoneEdges();
+ void getNextZoneGroup();
+ void endMarkingZoneGroup();
+ void beginSweepingZoneGroup(AutoLockForExclusiveAccess& lock);
+ bool shouldReleaseObservedTypes();
+ void endSweepingZoneGroup();
+ IncrementalProgress sweepPhase(SliceBudget& sliceBudget, AutoLockForExclusiveAccess& lock);
+ void endSweepPhase(bool lastGC, AutoLockForExclusiveAccess& lock);
+ void sweepZones(FreeOp* fop, bool lastGC);
+ void decommitAllWithoutUnlocking(const AutoLockGC& lock);
+ void startDecommit();
+ void queueZonesForBackgroundSweep(ZoneList& zones);
+ void sweepBackgroundThings(ZoneList& zones, LifoAlloc& freeBlocks);
+ void assertBackgroundSweepingFinished();
+ bool shouldCompact();
+ void beginCompactPhase();
+ IncrementalProgress compactPhase(JS::gcreason::Reason reason, SliceBudget& sliceBudget,
+ AutoLockForExclusiveAccess& lock);
+ void endCompactPhase(JS::gcreason::Reason reason);
+ void sweepTypesAfterCompacting(Zone* zone);
+ void sweepZoneAfterCompacting(Zone* zone);
+ MOZ_MUST_USE bool relocateArenas(Zone* zone, JS::gcreason::Reason reason,
+ Arena*& relocatedListOut, SliceBudget& sliceBudget);
+ void updateTypeDescrObjects(MovingTracer* trc, Zone* zone);
+ void updateCellPointers(MovingTracer* trc, Zone* zone, AllocKinds kinds, size_t bgTaskCount);
+ void updateAllCellPointers(MovingTracer* trc, Zone* zone);
+ void updatePointersToRelocatedCells(Zone* zone, AutoLockForExclusiveAccess& lock);
+ void protectAndHoldArenas(Arena* arenaList);
+ void unprotectHeldRelocatedArenas();
+ void releaseRelocatedArenas(Arena* arenaList);
+ void releaseRelocatedArenasWithoutUnlocking(Arena* arenaList, const AutoLockGC& lock);
+ void finishCollection(JS::gcreason::Reason reason);
+
+ void computeNonIncrementalMarkingForValidation(AutoLockForExclusiveAccess& lock);
+ void validateIncrementalMarking();
+ void finishMarkingValidation();
+
+#ifdef DEBUG
+ void checkForCompartmentMismatches();
+#endif
+
+ void callFinalizeCallbacks(FreeOp* fop, JSFinalizeStatus status) const;
+ void callWeakPointerZoneGroupCallbacks() const;
+ void callWeakPointerCompartmentCallbacks(JSCompartment* comp) const;
+
+ public:
+ JSRuntime* rt;
+
+ /* Embedders can use this zone however they wish. */
+ JS::Zone* systemZone;
+
+ /* List of compartments and zones (protected by the GC lock). */
+ ZoneVector zones;
+
+ Nursery nursery;
+ StoreBuffer storeBuffer;
+
+ gcstats::Statistics stats;
+
+ GCMarker marker;
+
+ /* Track heap usage for this runtime. */
+ HeapUsage usage;
+
+ /* GC scheduling state and parameters. */
+ GCSchedulingTunables tunables;
+ GCSchedulingState schedulingState;
+
+ MemProfiler mMemProfiler;
+
+ private:
+ // When empty, chunks reside in the emptyChunks pool and are re-used as
+ // needed or eventually expired if not re-used. The emptyChunks pool gets
+ // refilled from the background allocation task heuristically so that empty
+ // chunks should always available for immediate allocation without syscalls.
+ ChunkPool emptyChunks_;
+
+ // Chunks which have had some, but not all, of their arenas allocated live
+ // in the available chunk lists. When all available arenas in a chunk have
+ // been allocated, the chunk is removed from the available list and moved
+ // to the fullChunks pool. During a GC, if all arenas are free, the chunk
+ // is moved back to the emptyChunks pool and scheduled for eventual
+ // release.
+ ChunkPool availableChunks_;
+
+ // When all arenas in a chunk are used, it is moved to the fullChunks pool
+ // so as to reduce the cost of operations on the available lists.
+ ChunkPool fullChunks_;
+
+ RootedValueMap rootsHash;
+
+ size_t maxMallocBytes;
+
+ // An incrementing id used to assign unique ids to cells that require one.
+ mozilla::Atomic<uint64_t, mozilla::ReleaseAcquire> nextCellUniqueId_;
+
+ /*
+ * Number of the committed arenas in all GC chunks including empty chunks.
+ */
+ mozilla::Atomic<uint32_t, mozilla::ReleaseAcquire> numArenasFreeCommitted;
+ VerifyPreTracer* verifyPreData;
+
+ private:
+ bool chunkAllocationSinceLastGC;
+ int64_t lastGCTime;
+
+ JSGCMode mode;
+
+ mozilla::Atomic<size_t, mozilla::ReleaseAcquire> numActiveZoneIters;
+
+ /* During shutdown, the GC needs to clean up every possible object. */
+ bool cleanUpEverything;
+
+ // Gray marking must be done after all black marking is complete. However,
+ // we do not have write barriers on XPConnect roots. Therefore, XPConnect
+ // roots must be accumulated in the first slice of incremental GC. We
+ // accumulate these roots in each zone's gcGrayRoots vector and then mark
+ // them later, after black marking is complete for each compartment. This
+ // accumulation can fail, but in that case we switch to non-incremental GC.
+ enum class GrayBufferState {
+ Unused,
+ Okay,
+ Failed
+ };
+ GrayBufferState grayBufferState;
+ bool hasBufferedGrayRoots() const { return grayBufferState == GrayBufferState::Okay; }
+
+ // Clear each zone's gray buffers, but do not change the current state.
+ void resetBufferedGrayRoots() const;
+
+ // Reset the gray buffering state to Unused.
+ void clearBufferedGrayRoots() {
+ grayBufferState = GrayBufferState::Unused;
+ resetBufferedGrayRoots();
+ }
+
+ mozilla::Atomic<JS::gcreason::Reason, mozilla::Relaxed> majorGCTriggerReason;
+
+ JS::gcreason::Reason minorGCTriggerReason;
+
+ /* Perform full GC if rt->keepAtoms() becomes false. */
+ bool fullGCForAtomsRequested_;
+
+ /* Incremented at the start of every minor GC. */
+ uint64_t minorGCNumber;
+
+ /* Incremented at the start of every major GC. */
+ uint64_t majorGCNumber;
+
+ /* The major GC number at which to release observed type information. */
+ uint64_t jitReleaseNumber;
+
+ /* Incremented on every GC slice. */
+ uint64_t number;
+
+ /* The number at the time of the most recent GC's first slice. */
+ uint64_t startNumber;
+
+ /* Whether the currently running GC can finish in multiple slices. */
+ bool isIncremental;
+
+ /* Whether all zones are being collected in first GC slice. */
+ bool isFull;
+
+ /* Whether the heap will be compacted at the end of GC. */
+ bool isCompacting;
+
+ /* The invocation kind of the current GC, taken from the first slice. */
+ JSGCInvocationKind invocationKind;
+
+ /* The initial GC reason, taken from the first slice. */
+ JS::gcreason::Reason initialReason;
+
+#ifdef DEBUG
+ /*
+ * If this is 0, all cross-compartment proxies must be registered in the
+ * wrapper map. This checking must be disabled temporarily while creating
+ * new wrappers. When non-zero, this records the recursion depth of wrapper
+ * creation.
+ */
+ uintptr_t disableStrictProxyCheckingCount;
+#endif
+
+ /*
+ * The current incremental GC phase. This is also used internally in
+ * non-incremental GC.
+ */
+ State incrementalState;
+
+ /* Indicates that the last incremental slice exhausted the mark stack. */
+ bool lastMarkSlice;
+
+ /* Whether any sweeping will take place in the separate GC helper thread. */
+ bool sweepOnBackgroundThread;
+
+ /* Whether observed type information is being released in the current GC. */
+ bool releaseObservedTypes;
+
+ /* Whether any black->gray edges were found during marking. */
+ BlackGrayEdgeVector foundBlackGrayEdges;
+
+ /* Singly linekd list of zones to be swept in the background. */
+ ZoneList backgroundSweepZones;
+
+ /*
+ * Free LIFO blocks are transferred to this allocator before being freed on
+ * the background GC thread after sweeping.
+ */
+ LifoAlloc blocksToFreeAfterSweeping;
+
+ /*
+ * Free LIFO blocks are transferred to this allocator before being freed
+ * after minor GC.
+ */
+ LifoAlloc blocksToFreeAfterMinorGC;
+
+ /* Index of current zone group (for stats). */
+ unsigned zoneGroupIndex;
+
+ /*
+ * Incremental sweep state.
+ */
+ JS::Zone* zoneGroups;
+ JS::Zone* currentZoneGroup;
+ bool sweepingTypes;
+ unsigned finalizePhase;
+ JS::Zone* sweepZone;
+ AllocKind sweepKind;
+ bool abortSweepAfterCurrentGroup;
+
+ /*
+ * Concurrent sweep infrastructure.
+ */
+ void startTask(GCParallelTask& task, gcstats::Phase phase, AutoLockHelperThreadState& locked);
+ void joinTask(GCParallelTask& task, gcstats::Phase phase, AutoLockHelperThreadState& locked);
+
+ /*
+ * List head of arenas allocated during the sweep phase.
+ */
+ Arena* arenasAllocatedDuringSweep;
+
+ /*
+ * Incremental compacting state.
+ */
+ bool startedCompacting;
+ ZoneList zonesToMaybeCompact;
+ Arena* relocatedArenasToRelease;
+
+#ifdef JS_GC_ZEAL
+ MarkingValidator* markingValidator;
+#endif
+
+ /*
+ * Indicates that a GC slice has taken place in the middle of an animation
+ * frame, rather than at the beginning. In this case, the next slice will be
+ * delayed so that we don't get back-to-back slices.
+ */
+ bool interFrameGC;
+
+ /* Default budget for incremental GC slice. See js/SliceBudget.h. */
+ int64_t defaultTimeBudget_;
+
+ /*
+ * We disable incremental GC if we encounter a Class with a trace hook
+ * that does not implement write barriers.
+ */
+ bool incrementalAllowed;
+
+ /*
+ * GGC can be enabled from the command line while testing.
+ */
+ unsigned generationalDisabled;
+
+ /*
+ * Whether compacting GC can is enabled globally.
+ */
+ bool compactingEnabled;
+
+ /*
+ * Some code cannot tolerate compacting GC so it can be disabled temporarily
+ * with AutoDisableCompactingGC which uses this counter.
+ */
+ unsigned compactingDisabledCount;
+
+ /*
+ * This is true if we are in the middle of a brain transplant (e.g.,
+ * JS_TransplantObject) or some other operation that can manipulate
+ * dead zones.
+ */
+ bool manipulatingDeadZones;
+
+ /*
+ * This field is incremented each time we mark an object inside a
+ * zone with no incoming cross-compartment pointers. Typically if
+ * this happens it signals that an incremental GC is marking too much
+ * stuff. At various times we check this counter and, if it has changed, we
+ * run an immediate, non-incremental GC to clean up the dead
+ * zones. This should happen very rarely.
+ */
+ unsigned objectsMarkedInDeadZones;
+
+ bool poked;
+
+ /*
+ * These options control the zealousness of the GC. At every allocation,
+ * nextScheduled is decremented. When it reaches zero we do a full GC.
+ *
+ * At this point, if zeal_ is one of the types that trigger periodic
+ * collection, then nextScheduled is reset to the value of zealFrequency.
+ * Otherwise, no additional GCs take place.
+ *
+ * You can control these values in several ways:
+ * - Set the JS_GC_ZEAL environment variable
+ * - Call gczeal() or schedulegc() from inside shell-executed JS code
+ * (see the help for details)
+ *
+ * If gcZeal_ == 1 then we perform GCs in select places (during MaybeGC and
+ * whenever a GC poke happens). This option is mainly useful to embedders.
+ *
+ * We use zeal_ == 4 to enable write barrier verification. See the comment
+ * in jsgc.cpp for more information about this.
+ *
+ * zeal_ values from 8 to 10 periodically run different types of
+ * incremental GC.
+ *
+ * zeal_ value 14 performs periodic shrinking collections.
+ */
+#ifdef JS_GC_ZEAL
+ uint32_t zealModeBits;
+ int zealFrequency;
+ int nextScheduled;
+ bool deterministicOnly;
+ int incrementalLimit;
+
+ Vector<JSObject*, 0, SystemAllocPolicy> selectedForMarking;
+#endif
+
+ bool fullCompartmentChecks;
+
+ Callback<JSGCCallback> gcCallback;
+ Callback<JS::DoCycleCollectionCallback> gcDoCycleCollectionCallback;
+ Callback<JSObjectsTenuredCallback> tenuredCallback;
+ CallbackVector<JSFinalizeCallback> finalizeCallbacks;
+ CallbackVector<JSWeakPointerZoneGroupCallback> updateWeakPointerZoneGroupCallbacks;
+ CallbackVector<JSWeakPointerCompartmentCallback> updateWeakPointerCompartmentCallbacks;
+
+ /*
+ * Malloc counter to measure memory pressure for GC scheduling. It runs
+ * from maxMallocBytes down to zero.
+ */
+ mozilla::Atomic<ptrdiff_t, mozilla::ReleaseAcquire> mallocBytesUntilGC;
+
+ /*
+ * Whether a GC has been triggered as a result of mallocBytesUntilGC
+ * falling below zero.
+ */
+ mozilla::Atomic<bool, mozilla::ReleaseAcquire> mallocGCTriggered;
+
+ /*
+ * The trace operations to trace embedding-specific GC roots. One is for
+ * tracing through black roots and the other is for tracing through gray
+ * roots. The black/gray distinction is only relevant to the cycle
+ * collector.
+ */
+ CallbackVector<JSTraceDataOp> blackRootTracers;
+ Callback<JSTraceDataOp> grayRootTracer;
+
+ /* Always preserve JIT code during GCs, for testing. */
+ bool alwaysPreserveCode;
+
+ /*
+ * Some regions of code are hard for the static rooting hazard analysis to
+ * understand. In those cases, we trade the static analysis for a dynamic
+ * analysis. When this is non-zero, we should assert if we trigger, or
+ * might trigger, a GC.
+ */
+ int inUnsafeRegion;
+
+#ifdef DEBUG
+ size_t noGCOrAllocationCheck;
+ size_t noNurseryAllocationCheck;
+
+ bool arenasEmptyAtShutdown;
+#endif
+
+ /* Synchronize GC heap access between main thread and GCHelperState. */
+ friend class js::AutoLockGC;
+ js::Mutex lock;
+
+ BackgroundAllocTask allocTask;
+ BackgroundDecommitTask decommitTask;
+ GCHelperState helperState;
+
+ /*
+ * During incremental sweeping, this field temporarily holds the arenas of
+ * the current AllocKind being swept in order of increasing free space.
+ */
+ SortedArenaList incrementalSweepList;
+
+ friend class js::GCHelperState;
+ friend class MarkingValidator;
+ friend class AutoTraceSession;
+ friend class AutoEnterIteration;
+};
+
+/* Prevent compartments and zones from being collected during iteration. */
+class MOZ_RAII AutoEnterIteration {
+ GCRuntime* gc;
+
+ public:
+ explicit AutoEnterIteration(GCRuntime* gc_) : gc(gc_) {
+ ++gc->numActiveZoneIters;
+ }
+
+ ~AutoEnterIteration() {
+ MOZ_ASSERT(gc->numActiveZoneIters);
+ --gc->numActiveZoneIters;
+ }
+};
+
+// After pulling a Chunk out of the empty chunks pool, we want to run the
+// background allocator to refill it. The code that takes Chunks does so under
+// the GC lock. We need to start the background allocation under the helper
+// threads lock. To avoid lock inversion we have to delay the start until after
+// we are outside the GC lock. This class handles that delay automatically.
+class MOZ_RAII AutoMaybeStartBackgroundAllocation
+{
+ GCRuntime* gc;
+
+ public:
+ AutoMaybeStartBackgroundAllocation()
+ : gc(nullptr)
+ {}
+
+ void tryToStartBackgroundAllocation(GCRuntime& gc) {
+ this->gc = &gc;
+ }
+
+ ~AutoMaybeStartBackgroundAllocation() {
+ if (gc)
+ gc->startBackgroundAllocTaskIfIdle();
+ }
+};
+
+#ifdef JS_GC_ZEAL
+
+inline bool
+GCRuntime::hasZealMode(ZealMode mode)
+{
+ static_assert(size_t(ZealMode::Limit) < sizeof(zealModeBits) * 8,
+ "Zeal modes must fit in zealModeBits");
+ return zealModeBits & (1 << uint32_t(mode));
+}
+
+inline void
+GCRuntime::clearZealMode(ZealMode mode)
+{
+ zealModeBits &= ~(1 << uint32_t(mode));
+ MOZ_ASSERT(!hasZealMode(mode));
+}
+
+inline bool
+GCRuntime::upcomingZealousGC() {
+ return nextScheduled == 1;
+}
+
+inline bool
+GCRuntime::needZealousGC() {
+ if (nextScheduled > 0 && --nextScheduled == 0) {
+ if (hasZealMode(ZealMode::Alloc) ||
+ hasZealMode(ZealMode::GenerationalGC) ||
+ hasZealMode(ZealMode::IncrementalRootsThenFinish) ||
+ hasZealMode(ZealMode::IncrementalMarkAllThenFinish) ||
+ hasZealMode(ZealMode::IncrementalMultipleSlices) ||
+ hasZealMode(ZealMode::Compact))
+ {
+ nextScheduled = zealFrequency;
+ }
+ return true;
+ }
+ return false;
+}
+#else
+inline bool GCRuntime::hasZealMode(ZealMode mode) { return false; }
+inline void GCRuntime::clearZealMode(ZealMode mode) { }
+inline bool GCRuntime::upcomingZealousGC() { return false; }
+inline bool GCRuntime::needZealousGC() { return false; }
+#endif
+
+} /* namespace gc */
+
+} /* namespace js */
+
+#endif