blob: 8454dbed1d33b2a80a8e6d2a4c8a439392e81a54 [file] [log] [blame]
/*
* Copyright (C) 2003-2009, 2011, 2013-2016 Apple Inc. All rights reserved.
* Copyright (C) 2007 Eric Seidel <eric@webkit.org>
*
* This library is free software; you can redistribute it and/or
* modify it under the terms of the GNU Lesser General Public
* License as published by the Free Software Foundation; either
* version 2 of the License, or (at your option) any later version.
*
* This library is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
* Lesser General Public License for more details.
*
* You should have received a copy of the GNU Lesser General Public
* License along with this library; if not, write to the Free Software
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
*
*/
#include "config.h"
#include "Heap.h"
#include "CodeBlock.h"
#include "CodeBlockSet.h"
#include "ConservativeRoots.h"
#include "DFGWorklist.h"
#include "EdenGCActivityCallback.h"
#include "FullGCActivityCallback.h"
#include "GCActivityCallback.h"
#include "GCIncomingRefCountedSetInlines.h"
#include "GCSegmentedArrayInlines.h"
#include "GCTypeMap.h"
#include "HasOwnPropertyCache.h"
#include "HeapHelperPool.h"
#include "HeapIterationScope.h"
#include "HeapProfiler.h"
#include "HeapRootVisitor.h"
#include "HeapSnapshot.h"
#include "HeapStatistics.h"
#include "HeapVerifier.h"
#include "HelpingGCScope.h"
#include "IncrementalSweeper.h"
#include "Interpreter.h"
#include "JITStubRoutineSet.h"
#include "JITWorklist.h"
#include "JSCInlines.h"
#include "JSGlobalObject.h"
#include "JSLock.h"
#include "JSVirtualMachineInternal.h"
#include "MarkedSpaceInlines.h"
#include "SamplingProfiler.h"
#include "ShadowChicken.h"
#include "SuperSampler.h"
#include "StopIfNecessaryTimer.h"
#include "TypeProfilerLog.h"
#include "UnlinkedCodeBlock.h"
#include "VM.h"
#include "WeakSetInlines.h"
#include <algorithm>
#include <wtf/CurrentTime.h>
#include <wtf/MainThread.h>
#include <wtf/ParallelVectorIterator.h>
#include <wtf/ProcessID.h>
#include <wtf/RAMSize.h>
#include <wtf/SimpleStats.h>
#if USE(FOUNDATION)
#if __has_include(<objc/objc-internal.h>)
#include <objc/objc-internal.h>
#else
extern "C" void* objc_autoreleasePoolPush(void);
extern "C" void objc_autoreleasePoolPop(void *context);
#endif
#endif // USE(FOUNDATION)
using namespace std;
namespace JSC {
namespace {
double maxPauseMS(double thisPauseMS)
{
static double maxPauseMS = std::max(thisPauseMS, maxPauseMS);
return maxPauseMS;
}
} // anonymous namespace
class Heap::ResumeTheWorldScope {
public:
ResumeTheWorldScope(Heap& heap)
: m_heap(heap)
{
if (!Options::useConcurrentGC())
return;
if (Options::logGC()) {
double thisPauseMS = (MonotonicTime::now() - m_heap.m_stopTime).milliseconds();
dataLog("p=", thisPauseMS, " ms (max ", maxPauseMS(thisPauseMS), ")...]\n");
}
m_heap.resumeTheWorld();
}
~ResumeTheWorldScope()
{
if (!Options::useConcurrentGC())
return;
m_heap.stopTheWorld();
if (Options::logGC())
dataLog("[GC: ");
}
private:
Heap& m_heap;
};
namespace {
size_t minHeapSize(HeapType heapType, size_t ramSize)
{
if (heapType == LargeHeap) {
double result = min(
static_cast<double>(Options::largeHeapSize()),
ramSize * Options::smallHeapRAMFraction());
return static_cast<size_t>(result);
}
return Options::smallHeapSize();
}
size_t proportionalHeapSize(size_t heapSize, size_t ramSize)
{
if (heapSize < ramSize * Options::smallHeapRAMFraction())
return Options::smallHeapGrowthFactor() * heapSize;
if (heapSize < ramSize * Options::mediumHeapRAMFraction())
return Options::mediumHeapGrowthFactor() * heapSize;
return Options::largeHeapGrowthFactor() * heapSize;
}
bool isValidSharedInstanceThreadState(VM* vm)
{
return vm->currentThreadIsHoldingAPILock();
}
bool isValidThreadState(VM* vm)
{
if (vm->atomicStringTable() != wtfThreadData().atomicStringTable())
return false;
if (vm->isSharedInstance() && !isValidSharedInstanceThreadState(vm))
return false;
return true;
}
void recordType(TypeCountSet& set, JSCell* cell)
{
const char* typeName = "[unknown]";
const ClassInfo* info = cell->classInfo();
if (info && info->className)
typeName = info->className;
set.add(typeName);
}
bool measurePhaseTiming()
{
return false;
}
HashMap<const char*, GCTypeMap<SimpleStats>>& timingStats()
{
static HashMap<const char*, GCTypeMap<SimpleStats>>* result;
static std::once_flag once;
std::call_once(
once,
[] {
result = new HashMap<const char*, GCTypeMap<SimpleStats>>();
});
return *result;
}
SimpleStats& timingStats(const char* name, CollectionScope scope)
{
return timingStats().add(name, GCTypeMap<SimpleStats>()).iterator->value[scope];
}
class TimingScope {
public:
TimingScope(std::optional<CollectionScope> scope, const char* name)
: m_scope(scope)
, m_name(name)
{
if (measurePhaseTiming())
m_before = monotonicallyIncreasingTimeMS();
}
TimingScope(Heap& heap, const char* name)
: TimingScope(heap.collectionScope(), name)
{
}
void setScope(std::optional<CollectionScope> scope)
{
m_scope = scope;
}
void setScope(Heap& heap)
{
setScope(heap.collectionScope());
}
~TimingScope()
{
if (measurePhaseTiming()) {
double after = monotonicallyIncreasingTimeMS();
double timing = after - m_before;
SimpleStats& stats = timingStats(m_name, *m_scope);
stats.add(timing);
dataLog("[GC:", *m_scope, "] ", m_name, " took: ", timing, " ms (average ", stats.mean(), " ms).\n");
}
}
private:
std::optional<CollectionScope> m_scope;
double m_before;
const char* m_name;
};
} // anonymous namespace
class Heap::Thread : public AutomaticThread {
public:
Thread(const LockHolder& locker, Heap& heap)
: AutomaticThread(locker, heap.m_threadLock, heap.m_threadCondition)
, m_heap(heap)
{
}
protected:
PollResult poll(const LockHolder& locker) override
{
if (m_heap.m_threadShouldStop) {
m_heap.notifyThreadStopping(locker);
return PollResult::Stop;
}
if (m_heap.shouldCollectInThread(locker))
return PollResult::Work;
return PollResult::Wait;
}
WorkResult work() override
{
m_heap.collectInThread();
return WorkResult::Continue;
}
void threadDidStart() override
{
WTF::registerGCThread(GCThreadType::Main);
}
private:
Heap& m_heap;
};
Heap::Heap(VM* vm, HeapType heapType)
: m_heapType(heapType)
, m_ramSize(Options::forceRAMSize() ? Options::forceRAMSize() : ramSize())
, m_minBytesPerCycle(minHeapSize(m_heapType, m_ramSize))
, m_sizeAfterLastCollect(0)
, m_sizeAfterLastFullCollect(0)
, m_sizeBeforeLastFullCollect(0)
, m_sizeAfterLastEdenCollect(0)
, m_sizeBeforeLastEdenCollect(0)
, m_bytesAllocatedThisCycle(0)
, m_bytesAbandonedSinceLastFullCollect(0)
, m_maxEdenSize(m_minBytesPerCycle)
, m_maxHeapSize(m_minBytesPerCycle)
, m_shouldDoFullCollection(false)
, m_totalBytesVisited(0)
, m_objectSpace(this)
, m_extraMemorySize(0)
, m_deprecatedExtraMemorySize(0)
, m_machineThreads(this)
, m_collectorSlotVisitor(std::make_unique<SlotVisitor>(*this))
, m_mutatorMarkStack(std::make_unique<MarkStackArray>())
, m_handleSet(vm)
, m_codeBlocks(std::make_unique<CodeBlockSet>())
, m_jitStubRoutines(std::make_unique<JITStubRoutineSet>())
, m_isSafeToCollect(false)
, m_writeBarrierBuffer(256)
, m_vm(vm)
// We seed with 10ms so that GCActivityCallback::didAllocate doesn't continuously
// schedule the timer if we've never done a collection.
, m_lastFullGCLength(0.01)
, m_lastEdenGCLength(0.01)
#if USE(CF)
, m_runLoop(CFRunLoopGetCurrent())
#endif // USE(CF)
, m_fullActivityCallback(GCActivityCallback::createFullTimer(this))
, m_edenActivityCallback(GCActivityCallback::createEdenTimer(this))
, m_sweeper(adoptRef(new IncrementalSweeper(this)))
, m_stopIfNecessaryTimer(adoptRef(new StopIfNecessaryTimer(vm)))
, m_deferralDepth(0)
#if USE(FOUNDATION)
, m_delayedReleaseRecursionCount(0)
#endif
, m_sharedCollectorMarkStack(std::make_unique<MarkStackArray>())
, m_sharedMutatorMarkStack(std::make_unique<MarkStackArray>())
, m_helperClient(&heapHelperPool())
, m_threadLock(Box<Lock>::create())
, m_threadCondition(AutomaticThreadCondition::create())
{
m_worldState.store(0);
if (Options::verifyHeap())
m_verifier = std::make_unique<HeapVerifier>(this, Options::numberOfGCCyclesToRecordForVerification());
LockHolder locker(*m_threadLock);
m_thread = adoptRef(new Thread(locker, *this));
}
Heap::~Heap()
{
for (auto& slotVisitor : m_parallelSlotVisitors)
slotVisitor->clearMarkStacks();
m_collectorSlotVisitor->clearMarkStacks();
m_mutatorMarkStack->clear();
for (WeakBlock* block : m_logicallyEmptyWeakBlocks)
WeakBlock::destroy(*this, block);
}
bool Heap::isPagedOut(double deadline)
{
return m_objectSpace.isPagedOut(deadline);
}
// The VM is being destroyed and the collector will never run again.
// Run all pending finalizers now because we won't get another chance.
void Heap::lastChanceToFinalize()
{
RELEASE_ASSERT(!m_vm->entryScope);
RELEASE_ASSERT(m_mutatorState == MutatorState::Running);
// Carefully bring the thread down. We need to use waitForCollector() until we know that there
// won't be any other collections.
bool stopped = false;
{
LockHolder locker(*m_threadLock);
stopped = m_thread->tryStop(locker);
if (!stopped) {
m_threadShouldStop = true;
m_threadCondition->notifyOne(locker);
}
}
if (!stopped) {
waitForCollector(
[&] (const LockHolder&) -> bool {
return m_threadIsStopping;
});
// It's now safe to join the thread, since we know that there will not be any more collections.
m_thread->join();
}
m_arrayBuffers.lastChanceToFinalize();
m_codeBlocks->lastChanceToFinalize();
m_objectSpace.lastChanceToFinalize();
releaseDelayedReleasedObjects();
sweepAllLogicallyEmptyWeakBlocks();
}
void Heap::releaseDelayedReleasedObjects()
{
#if USE(FOUNDATION)
// We need to guard against the case that releasing an object can create more objects due to the
// release calling into JS. When those JS call(s) exit and all locks are being dropped we end up
// back here and could try to recursively release objects. We guard that with a recursive entry
// count. Only the initial call will release objects, recursive calls simple return and let the
// the initial call to the function take care of any objects created during release time.
// This also means that we need to loop until there are no objects in m_delayedReleaseObjects
// and use a temp Vector for the actual releasing.
if (!m_delayedReleaseRecursionCount++) {
while (!m_delayedReleaseObjects.isEmpty()) {
ASSERT(m_vm->currentThreadIsHoldingAPILock());
Vector<RetainPtr<CFTypeRef>> objectsToRelease = WTFMove(m_delayedReleaseObjects);
{
// We need to drop locks before calling out to arbitrary code.
JSLock::DropAllLocks dropAllLocks(m_vm);
void* context = objc_autoreleasePoolPush();
objectsToRelease.clear();
objc_autoreleasePoolPop(context);
}
}
}
m_delayedReleaseRecursionCount--;
#endif
}
void Heap::reportExtraMemoryAllocatedSlowCase(size_t size)
{
didAllocate(size);
collectIfNecessaryOrDefer();
}
void Heap::deprecatedReportExtraMemorySlowCase(size_t size)
{
m_deprecatedExtraMemorySize += size;
reportExtraMemoryAllocatedSlowCase(size);
}
void Heap::reportAbandonedObjectGraph()
{
// Our clients don't know exactly how much memory they
// are abandoning so we just guess for them.
size_t abandonedBytes = static_cast<size_t>(0.1 * capacity());
// We want to accelerate the next collection. Because memory has just
// been abandoned, the next collection has the potential to
// be more profitable. Since allocation is the trigger for collection,
// we hasten the next collection by pretending that we've allocated more memory.
if (m_fullActivityCallback) {
m_fullActivityCallback->didAllocate(
m_sizeAfterLastCollect - m_sizeAfterLastFullCollect + m_bytesAllocatedThisCycle + m_bytesAbandonedSinceLastFullCollect);
}
m_bytesAbandonedSinceLastFullCollect += abandonedBytes;
}
void Heap::protect(JSValue k)
{
ASSERT(k);
ASSERT(m_vm->currentThreadIsHoldingAPILock());
if (!k.isCell())
return;
m_protectedValues.add(k.asCell());
}
bool Heap::unprotect(JSValue k)
{
ASSERT(k);
ASSERT(m_vm->currentThreadIsHoldingAPILock());
if (!k.isCell())
return false;
return m_protectedValues.remove(k.asCell());
}
void Heap::addReference(JSCell* cell, ArrayBuffer* buffer)
{
if (m_arrayBuffers.addReference(cell, buffer)) {
collectIfNecessaryOrDefer();
didAllocate(buffer->gcSizeEstimateInBytes());
}
}
void Heap::harvestWeakReferences()
{
m_collectorSlotVisitor->harvestWeakReferences();
}
void Heap::finalizeUnconditionalFinalizers()
{
m_collectorSlotVisitor->finalizeUnconditionalFinalizers();
}
void Heap::willStartIterating()
{
m_objectSpace.willStartIterating();
}
void Heap::didFinishIterating()
{
m_objectSpace.didFinishIterating();
}
void Heap::completeAllJITPlans()
{
#if ENABLE(JIT)
JITWorklist::instance()->completeAllForVM(*m_vm);
#endif // ENABLE(JIT)
DFG::completeAllPlansForVM(*m_vm);
}
void Heap::markToFixpoint(double gcStartTime)
{
TimingScope markToFixpointTimingScope(*this, "Heap::markToFixpoint");
HeapRootVisitor heapRootVisitor(*m_collectorSlotVisitor);
if (m_collectionScope == CollectionScope::Full) {
m_opaqueRoots.clear();
m_collectorSlotVisitor->clearMarkStacks();
m_mutatorMarkStack->clear();
}
beginMarking();
m_parallelMarkersShouldExit = false;
m_helperClient.setFunction(
[this] () {
SlotVisitor* slotVisitor;
{
LockHolder locker(m_parallelSlotVisitorLock);
if (m_availableParallelSlotVisitors.isEmpty()) {
std::unique_ptr<SlotVisitor> newVisitor =
std::make_unique<SlotVisitor>(*this);
slotVisitor = newVisitor.get();
m_parallelSlotVisitors.append(WTFMove(newVisitor));
} else
slotVisitor = m_availableParallelSlotVisitors.takeLast();
}
WTF::registerGCThread(GCThreadType::Helper);
{
ParallelModeEnabler parallelModeEnabler(*slotVisitor);
slotVisitor->didStartMarking();
slotVisitor->drainFromShared(SlotVisitor::SlaveDrain);
}
{
LockHolder locker(m_parallelSlotVisitorLock);
m_availableParallelSlotVisitors.append(slotVisitor);
}
});
m_collectorSlotVisitor->didStartMarking();
MonotonicTime initialTime = MonotonicTime::now();
const Seconds period = Seconds::fromMilliseconds(Options::concurrentGCPeriodMS());
const double bytesAllocatedThisCycleAtTheBeginning = m_bytesAllocatedThisCycle;
const double bytesAllocatedThisCycleAtTheEnd =
Options::concurrentGCMaxHeadroom() *
std::max(
bytesAllocatedThisCycleAtTheBeginning,
static_cast<double>(m_maxEdenSize));
auto targetMutatorUtilization = [&] () -> double {
double headroomFullness =
(m_bytesAllocatedThisCycle - bytesAllocatedThisCycleAtTheBeginning) /
(bytesAllocatedThisCycleAtTheEnd - bytesAllocatedThisCycleAtTheBeginning);
// headroomFullness can be NaN and other interesting things if
// bytesAllocatedThisCycleAtTheBeginning is zero. We see that in debug tests. This code
// defends against all floating point dragons.
if (!(headroomFullness >= 0))
headroomFullness = 0;
if (!(headroomFullness <= 1))
headroomFullness = 1;
double mutatorUtilization = 1 - headroomFullness;
// Scale the mutator utilization into the permitted window.
mutatorUtilization =
Options::minimumMutatorUtilization() +
mutatorUtilization * (
Options::maximumMutatorUtilization() -
Options::minimumMutatorUtilization());
return mutatorUtilization;
};
auto targetCollectorUtilization = [&] () -> double {
return 1 - targetMutatorUtilization();
};
auto elapsedInPeriod = [&] (MonotonicTime now) -> Seconds {
return (now - initialTime) % period;
};
auto phase = [&] (MonotonicTime now) -> double {
return elapsedInPeriod(now) / period;
};
auto shouldBeResumed = [&] (MonotonicTime now) -> bool {
if (Options::collectorShouldResumeFirst())
return phase(now) <= targetMutatorUtilization();
return phase(now) > targetCollectorUtilization();
};
auto timeToResume = [&] (MonotonicTime now) -> MonotonicTime {
ASSERT(!shouldBeResumed(now));
if (Options::collectorShouldResumeFirst())
return now - elapsedInPeriod(now) + period;
return now - elapsedInPeriod(now) + period * targetCollectorUtilization();
};
auto timeToStop = [&] (MonotonicTime now) -> MonotonicTime {
ASSERT(shouldBeResumed(now));
if (Options::collectorShouldResumeFirst())
return now - elapsedInPeriod(now) + period * targetMutatorUtilization();
return now - - elapsedInPeriod(now) + period;
};
// Adjust the target extra pause ratio as necessary.
double rateOfCollection =
(m_lastGCEndTime - m_lastGCStartTime) /
(m_currentGCStartTime - m_lastGCStartTime);
if (Options::logGC())
dataLog("cr=", rateOfCollection, " ");
// FIXME: Determine if this is useful or get rid of it.
// https://bugs.webkit.org/show_bug.cgi?id=164940
double extraPauseRatio = Options::initialExtraPauseRatio();
for (unsigned iteration = 1; ; ++iteration) {
if (Options::logGC())
dataLog("i#", iteration, " ");
MonotonicTime topOfLoop = MonotonicTime::now();
{
TimingScope preConvergenceTimingScope(*this, "Heap::markToFixpoint conservative scan");
ConservativeRoots conservativeRoots(*this);
SuperSamplerScope superSamplerScope(false);
gatherStackRoots(conservativeRoots);
gatherJSStackRoots(conservativeRoots);
gatherScratchBufferRoots(conservativeRoots);
visitConservativeRoots(conservativeRoots);
}
// Now we visit roots that don't get barriered, so each fixpoint iteration just revisits
// all of them.
#if JSC_OBJC_API_ENABLED
scanExternalRememberedSet(*m_vm, *m_collectorSlotVisitor);
#endif
if (m_vm->smallStrings.needsToBeVisited(*m_collectionScope))
m_vm->smallStrings.visitStrongReferences(*m_collectorSlotVisitor);
for (auto& pair : m_protectedValues)
heapRootVisitor.visit(&pair.key);
if (m_markListSet && m_markListSet->size())
MarkedArgumentBuffer::markLists(heapRootVisitor, *m_markListSet);
if (m_vm->exception())
heapRootVisitor.visit(m_vm->addressOfException());
if (m_vm->lastException())
heapRootVisitor.visit(m_vm->addressOfLastException());
m_handleSet.visitStrongHandles(heapRootVisitor);
m_handleStack.visit(heapRootVisitor);
#if ENABLE(SAMPLING_PROFILER)
if (SamplingProfiler* samplingProfiler = m_vm->samplingProfiler()) {
LockHolder locker(samplingProfiler->getLock());
samplingProfiler->processUnverifiedStackTraces();
samplingProfiler->visit(*m_collectorSlotVisitor);
if (Options::logGC() == GCLogging::Verbose)
dataLog("Sampling Profiler data:\n", *m_collectorSlotVisitor);
}
#endif // ENABLE(SAMPLING_PROFILER)
if (m_vm->typeProfiler())
m_vm->typeProfilerLog()->visit(*m_collectorSlotVisitor);
m_vm->shadowChicken().visitChildren(*m_collectorSlotVisitor);
m_jitStubRoutines->traceMarkedStubRoutines(*m_collectorSlotVisitor);
m_collectorSlotVisitor->mergeOpaqueRootsIfNecessary();
for (auto& parallelVisitor : m_parallelSlotVisitors)
parallelVisitor->mergeOpaqueRootsIfNecessary();
m_objectSpace.visitWeakSets(heapRootVisitor);
harvestWeakReferences();
visitCompilerWorklistWeakReferences();
DFG::markCodeBlocks(*m_vm, *m_collectorSlotVisitor);
bool shouldTerminate = m_collectorSlotVisitor->isEmpty() && m_mutatorMarkStack->isEmpty();
if (Options::logGC())
dataLog(m_collectorSlotVisitor->collectorMarkStack().size(), "+", m_mutatorMarkStack->size() + m_collectorSlotVisitor->mutatorMarkStack().size(), ", a=", m_bytesAllocatedThisCycle / 1024, " kb, b=", m_barriersExecuted, ", mu=", targetMutatorUtilization(), " ");
// We want to do this to conservatively ensure that we rescan any code blocks that are
// running right now. However, we need to be sure to do it *after* we mark the code block
// so that we know for sure if it really needs a barrier. Also, this has to happen after the
// fixpoint check - otherwise we might loop forever. Incidentally, we also want to do this
// at the end of GC so that anything at the end of the last GC gets barriered in the next
// GC.
m_codeBlocks->writeBarrierCurrentlyExecuting(this);
DFG::rememberCodeBlocks(*m_vm);
if (shouldTerminate)
break;
// The SlotVisitor's mark stacks are accessed by the collector thread (i.e. this thread)
// without locks. That's why we double-buffer.
m_mutatorMarkStack->transferTo(m_collectorSlotVisitor->mutatorMarkStack());
if (Options::logGC() == GCLogging::Verbose)
dataLog("Live Weak Handles:\n", *m_collectorSlotVisitor);
MonotonicTime beforeConvergence = MonotonicTime::now();
{
TimingScope traceTimingScope(*this, "Heap::markToFixpoint tracing");
ParallelModeEnabler enabler(*m_collectorSlotVisitor);
if (Options::useCollectorTimeslicing()) {
// Before we yield to the mutator, we should do GC work proportional to the time we
// spent paused. We initialize the timeslicer to start after this "mandatory" pause
// completes.
SlotVisitor::SharedDrainResult drainResult;
Seconds extraPause = (beforeConvergence - topOfLoop) * extraPauseRatio;
initialTime = beforeConvergence + extraPause;
drainResult = m_collectorSlotVisitor->drainInParallel(initialTime);
while (drainResult != SlotVisitor::SharedDrainResult::Done) {
MonotonicTime now = MonotonicTime::now();
if (shouldBeResumed(now)) {
ResumeTheWorldScope resumeTheWorldScope(*this);
drainResult = m_collectorSlotVisitor->drainInParallel(timeToStop(now));
} else
drainResult = m_collectorSlotVisitor->drainInParallel(timeToResume(now));
}
} else {
// Disabling collector timeslicing is meant to be used together with
// --collectContinuously=true to maximize the opportunity for harmful races.
ResumeTheWorldScope resumeTheWorldScope(*this);
m_collectorSlotVisitor->drainInParallel();
}
}
extraPauseRatio *= Options::extraPauseRatioIterationGrowthRate();
}
{
std::lock_guard<Lock> lock(m_markingMutex);
m_parallelMarkersShouldExit = true;
m_markingConditionVariable.notifyAll();
}
m_helperClient.finish();
updateObjectCounts(gcStartTime);
endMarking();
}
void Heap::gatherStackRoots(ConservativeRoots& roots)
{
m_machineThreads.gatherConservativeRoots(roots, *m_jitStubRoutines, *m_codeBlocks);
}
void Heap::gatherJSStackRoots(ConservativeRoots& roots)
{
#if !ENABLE(JIT)
m_vm->interpreter->cloopStack().gatherConservativeRoots(roots, *m_jitStubRoutines, *m_codeBlocks);
#else
UNUSED_PARAM(roots);
#endif
}
void Heap::gatherScratchBufferRoots(ConservativeRoots& roots)
{
#if ENABLE(DFG_JIT)
m_vm->gatherConservativeRoots(roots);
#else
UNUSED_PARAM(roots);
#endif
}
void Heap::beginMarking()
{
TimingScope timingScope(*this, "Heap::beginMarking");
if (m_collectionScope == CollectionScope::Full)
m_codeBlocks->clearMarksForFullCollection();
m_jitStubRoutines->clearMarks();
m_objectSpace.beginMarking();
m_mutatorShouldBeFenced = true;
m_barrierThreshold = tautologicalThreshold;
m_barriersExecuted = 0;
}
void Heap::visitConservativeRoots(ConservativeRoots& roots)
{
m_collectorSlotVisitor->append(roots);
if (Options::logGC() == GCLogging::Verbose)
dataLog("Conservative Roots:\n", *m_collectorSlotVisitor);
}
void Heap::visitCompilerWorklistWeakReferences()
{
#if ENABLE(DFG_JIT)
for (unsigned i = DFG::numberOfWorklists(); i--;)
DFG::existingWorklistForIndex(i).visitWeakReferences(*m_collectorSlotVisitor);
if (Options::logGC() == GCLogging::Verbose)
dataLog("DFG Worklists:\n", *m_collectorSlotVisitor);
#endif
}
void Heap::removeDeadCompilerWorklistEntries()
{
#if ENABLE(DFG_JIT)
for (unsigned i = DFG::numberOfWorklists(); i--;)
DFG::existingWorklistForIndex(i).removeDeadPlans(*m_vm);
#endif
}
bool Heap::isHeapSnapshotting() const
{
HeapProfiler* heapProfiler = m_vm->heapProfiler();
if (UNLIKELY(heapProfiler))
return heapProfiler->activeSnapshotBuilder();
return false;
}
struct GatherHeapSnapshotData : MarkedBlock::CountFunctor {
GatherHeapSnapshotData(HeapSnapshotBuilder& builder)
: m_builder(builder)
{
}
IterationStatus operator()(HeapCell* heapCell, HeapCell::Kind kind) const
{
if (kind == HeapCell::JSCell) {
JSCell* cell = static_cast<JSCell*>(heapCell);
cell->methodTable()->heapSnapshot(cell, m_builder);
}
return IterationStatus::Continue;
}
HeapSnapshotBuilder& m_builder;
};
void Heap::gatherExtraHeapSnapshotData(HeapProfiler& heapProfiler)
{
if (HeapSnapshotBuilder* builder = heapProfiler.activeSnapshotBuilder()) {
HeapIterationScope heapIterationScope(*this);
GatherHeapSnapshotData functor(*builder);
m_objectSpace.forEachLiveCell(heapIterationScope, functor);
}
}
struct RemoveDeadHeapSnapshotNodes : MarkedBlock::CountFunctor {
RemoveDeadHeapSnapshotNodes(HeapSnapshot& snapshot)
: m_snapshot(snapshot)
{
}
IterationStatus operator()(HeapCell* cell, HeapCell::Kind kind) const
{
if (kind == HeapCell::JSCell)
m_snapshot.sweepCell(static_cast<JSCell*>(cell));
return IterationStatus::Continue;
}
HeapSnapshot& m_snapshot;
};
void Heap::removeDeadHeapSnapshotNodes(HeapProfiler& heapProfiler)
{
if (HeapSnapshot* snapshot = heapProfiler.mostRecentSnapshot()) {
HeapIterationScope heapIterationScope(*this);
RemoveDeadHeapSnapshotNodes functor(*snapshot);
m_objectSpace.forEachDeadCell(heapIterationScope, functor);
snapshot->shrinkToFit();
}
}
void Heap::updateObjectCounts(double gcStartTime)
{
if (Options::logGC() == GCLogging::Verbose) {
size_t visitCount = m_collectorSlotVisitor->visitCount();
visitCount += threadVisitCount();
dataLogF("\nNumber of live Objects after GC %lu, took %.6f secs\n", static_cast<unsigned long>(visitCount), WTF::monotonicallyIncreasingTime() - gcStartTime);
}
if (m_collectionScope == CollectionScope::Full)
m_totalBytesVisited = 0;
m_totalBytesVisitedThisCycle =
m_collectorSlotVisitor->bytesVisited() +
threadBytesVisited();
m_totalBytesVisited += m_totalBytesVisitedThisCycle;
}
void Heap::endMarking()
{
m_collectorSlotVisitor->reset();
for (auto& parallelVisitor : m_parallelSlotVisitors)
parallelVisitor->reset();
RELEASE_ASSERT(m_sharedCollectorMarkStack->isEmpty());
RELEASE_ASSERT(m_sharedMutatorMarkStack->isEmpty());
m_weakReferenceHarvesters.removeAll();
m_objectSpace.endMarking();
m_mutatorShouldBeFenced = Options::forceFencedBarrier();
m_barrierThreshold = Options::forceFencedBarrier() ? tautologicalThreshold : blackThreshold;
}
size_t Heap::objectCount()
{
return m_objectSpace.objectCount();
}
size_t Heap::extraMemorySize()
{
return m_extraMemorySize + m_deprecatedExtraMemorySize + m_arrayBuffers.size();
}
size_t Heap::size()
{
return m_objectSpace.size() + extraMemorySize();
}
size_t Heap::capacity()
{
return m_objectSpace.capacity() + extraMemorySize();
}
size_t Heap::protectedGlobalObjectCount()
{
size_t result = 0;
forEachProtectedCell(
[&] (JSCell* cell) {
if (cell->isObject() && asObject(cell)->isGlobalObject())
result++;
});
return result;
}
size_t Heap::globalObjectCount()
{
HeapIterationScope iterationScope(*this);
size_t result = 0;
m_objectSpace.forEachLiveCell(
iterationScope,
[&] (HeapCell* heapCell, HeapCell::Kind kind) -> IterationStatus {
if (kind != HeapCell::JSCell)
return IterationStatus::Continue;
JSCell* cell = static_cast<JSCell*>(heapCell);
if (cell->isObject() && asObject(cell)->isGlobalObject())
result++;
return IterationStatus::Continue;
});
return result;
}
size_t Heap::protectedObjectCount()
{
size_t result = 0;
forEachProtectedCell(
[&] (JSCell*) {
result++;
});
return result;
}
std::unique_ptr<TypeCountSet> Heap::protectedObjectTypeCounts()
{
std::unique_ptr<TypeCountSet> result = std::make_unique<TypeCountSet>();
forEachProtectedCell(
[&] (JSCell* cell) {
recordType(*result, cell);
});
return result;
}
std::unique_ptr<TypeCountSet> Heap::objectTypeCounts()
{
std::unique_ptr<TypeCountSet> result = std::make_unique<TypeCountSet>();
HeapIterationScope iterationScope(*this);
m_objectSpace.forEachLiveCell(
iterationScope,
[&] (HeapCell* cell, HeapCell::Kind kind) -> IterationStatus {
if (kind == HeapCell::JSCell)
recordType(*result, static_cast<JSCell*>(cell));
return IterationStatus::Continue;
});
return result;
}
void Heap::deleteAllCodeBlocks()
{
// If JavaScript is running, it's not safe to delete all JavaScript code, since
// we'll end up returning to deleted code.
RELEASE_ASSERT(!m_vm->entryScope);
ASSERT(!m_collectionScope);
completeAllJITPlans();
for (ExecutableBase* executable : m_executables)
executable->clearCode();
}
void Heap::deleteAllUnlinkedCodeBlocks()
{
for (ExecutableBase* current : m_executables) {
if (!current->isFunctionExecutable())
continue;
static_cast<FunctionExecutable*>(current)->unlinkedExecutable()->clearCode();
}
}
void Heap::clearUnmarkedExecutables()
{
for (unsigned i = m_executables.size(); i--;) {
ExecutableBase* current = m_executables[i];
if (isMarked(current))
continue;
// Eagerly dereference the Executable's JITCode in order to run watchpoint
// destructors. Otherwise, watchpoints might fire for deleted CodeBlocks.
current->clearCode();
std::swap(m_executables[i], m_executables.last());
m_executables.removeLast();
}
m_executables.shrinkToFit();
}
void Heap::deleteUnmarkedCompiledCode()
{
clearUnmarkedExecutables();
m_codeBlocks->deleteUnmarkedAndUnreferenced(*m_lastCollectionScope);
m_jitStubRoutines->deleteUnmarkedJettisonedStubRoutines();
}
void Heap::addToRememberedSet(const JSCell* cell)
{
ASSERT(cell);
ASSERT(!Options::useConcurrentJIT() || !isCompilationThread());
m_barriersExecuted++;
if (!Heap::isMarkedConcurrently(cell)) {
// During a full collection a store into an unmarked object that had surivived past
// collections will manifest as a store to an unmarked black object. If the object gets
// marked at some time after this then it will go down the normal marking path. We can
// safely ignore these stores.
return;
}
// It could be that the object was *just* marked. This means that the collector may set the
// state to Grey and then to AnthraciteOrBlack at any time. It's OK for us to race with the
// collector here. If we win then this is accurate because the object _will_ get scanned again.
// If we lose then someone else will barrier the object again. That would be unfortunate but not
// the end of the world.
cell->setCellState(CellState::Grey);
m_mutatorMarkStack->append(cell);
}
void Heap::collectAllGarbage()
{
if (!m_isSafeToCollect)
return;
collectSync(CollectionScope::Full);
DeferGCForAWhile deferGC(*this);
if (UNLIKELY(Options::useImmortalObjects()))
sweeper()->willFinishSweeping();
else {
double before = 0;
if (Options::logGC()) {
dataLog("[Full sweep: ", capacity() / 1024, " kb ");
before = currentTimeMS();
}
m_objectSpace.sweep();
m_objectSpace.shrink();
if (Options::logGC()) {
double after = currentTimeMS();
dataLog("=> ", capacity() / 1024, " kb, ", after - before, " ms]\n");
}
}
m_objectSpace.assertNoUnswept();
sweepAllLogicallyEmptyWeakBlocks();
}
void Heap::collectAsync(std::optional<CollectionScope> scope)
{
if (!m_isSafeToCollect)
return;
bool alreadyRequested = false;
{
LockHolder locker(*m_threadLock);
for (std::optional<CollectionScope> request : m_requests) {
if (scope) {
if (scope == CollectionScope::Eden) {
alreadyRequested = true;
break;
} else {
RELEASE_ASSERT(scope == CollectionScope::Full);
if (request == CollectionScope::Full) {
alreadyRequested = true;
break;
}
}
} else {
if (!request || request == CollectionScope::Full) {
alreadyRequested = true;
break;
}
}
}
}
if (alreadyRequested)
return;
requestCollection(scope);
}
void Heap::collectSync(std::optional<CollectionScope> scope)
{
if (!m_isSafeToCollect)
return;
waitForCollection(requestCollection(scope));
}
bool Heap::shouldCollectInThread(const LockHolder&)
{
RELEASE_ASSERT(m_requests.isEmpty() == (m_lastServedTicket == m_lastGrantedTicket));
RELEASE_ASSERT(m_lastServedTicket <= m_lastGrantedTicket);
return !m_requests.isEmpty();
}
void Heap::collectInThread()
{
m_currentGCStartTime = MonotonicTime::now();
std::optional<CollectionScope> scope;
{
LockHolder locker(*m_threadLock);
RELEASE_ASSERT(!m_requests.isEmpty());
scope = m_requests.first();
}
SuperSamplerScope superSamplerScope(false);
TimingScope collectImplTimingScope(scope, "Heap::collectInThread");
#if ENABLE(ALLOCATION_LOGGING)
dataLogF("JSC GC starting collection.\n");
#endif
stopTheWorld();
MonotonicTime before;
if (Options::logGC()) {
dataLog("[GC: START ", capacity() / 1024, " kb ");
before = MonotonicTime::now();
}
double gcStartTime;
ASSERT(m_isSafeToCollect);
if (m_collectionScope) {
dataLog("Collection scope already set during GC: ", *m_collectionScope, "\n");
RELEASE_ASSERT_NOT_REACHED();
}
willStartCollection(scope);
collectImplTimingScope.setScope(*this);
gcStartTime = WTF::monotonicallyIncreasingTime();
if (m_verifier) {
// Verify that live objects from the last GC cycle haven't been corrupted by
// mutators before we begin this new GC cycle.
m_verifier->verify(HeapVerifier::Phase::BeforeGC);
m_verifier->initializeGCCycle();
m_verifier->gatherLiveObjects(HeapVerifier::Phase::BeforeMarking);
}
prepareForMarking();
markToFixpoint(gcStartTime);
if (m_verifier) {
m_verifier->gatherLiveObjects(HeapVerifier::Phase::AfterMarking);
m_verifier->verify(HeapVerifier::Phase::AfterMarking);
}
if (vm()->typeProfiler())
vm()->typeProfiler()->invalidateTypeSetCache();
reapWeakHandles();
pruneStaleEntriesFromWeakGCMaps();
sweepArrayBuffers();
snapshotUnswept();
finalizeUnconditionalFinalizers();
removeDeadCompilerWorklistEntries();
notifyIncrementalSweeper();
m_codeBlocks->writeBarrierCurrentlyExecuting(this);
m_codeBlocks->clearCurrentlyExecuting();
prepareForAllocation();
updateAllocationLimits();
didFinishCollection(gcStartTime);
if (m_verifier) {
m_verifier->trimDeadObjects();
m_verifier->verify(HeapVerifier::Phase::AfterGC);
}
if (false) {
dataLog("Heap state after GC:\n");
m_objectSpace.dumpBits();
}
if (Options::logGC()) {
MonotonicTime after = MonotonicTime::now();
double thisPauseMS = (after - m_stopTime).milliseconds();
dataLog("p=", thisPauseMS, " ms (max ", maxPauseMS(thisPauseMS), "), cycle ", (after - before).milliseconds(), " ms END]\n");
}
{
LockHolder locker(*m_threadLock);
m_requests.removeFirst();
m_lastServedTicket++;
clearMutatorWaiting();
}
ParkingLot::unparkAll(&m_worldState);
setNeedFinalize();
resumeTheWorld();
m_lastGCStartTime = m_currentGCStartTime;
m_lastGCEndTime = MonotonicTime::now();
}
void Heap::stopTheWorld()
{
RELEASE_ASSERT(!m_collectorBelievesThatTheWorldIsStopped);
waitWhileNeedFinalize();
stopTheMutator();
suspendCompilerThreads();
m_collectorBelievesThatTheWorldIsStopped = true;
#if ENABLE(JIT)
{
DeferGCForAWhile awhile(*this);
if (JITWorklist::instance()->completeAllForVM(*m_vm))
setGCDidJIT();
}
#endif // ENABLE(JIT)
vm()->shadowChicken().update(*vm(), vm()->topCallFrame);
flushWriteBarrierBuffer();
m_structureIDTable.flushOldTables();
m_objectSpace.stopAllocating();
m_stopTime = MonotonicTime::now();
}
void Heap::resumeTheWorld()
{
// Calling resumeAllocating does the Right Thing depending on whether this is the end of a
// collection cycle or this is just a concurrent phase within a collection cycle:
// - At end of collection cycle: it's a no-op because prepareForAllocation already cleared the
// last active block.
// - During collection cycle: it reinstates the last active block.
m_objectSpace.resumeAllocating();
RELEASE_ASSERT(m_collectorBelievesThatTheWorldIsStopped);
m_collectorBelievesThatTheWorldIsStopped = false;
resumeCompilerThreads();
resumeTheMutator();
}
void Heap::stopTheMutator()
{
for (;;) {
unsigned oldState = m_worldState.load();
if ((oldState & stoppedBit)
&& (oldState & shouldStopBit))
return;
// Note: We could just have the mutator stop in-place like we do when !hasAccessBit. We could
// switch to that if it turned out to be less confusing, but then it would not give the
// mutator the opportunity to react to the world being stopped.
if (oldState & mutatorWaitingBit) {
if (m_worldState.compareExchangeWeak(oldState, oldState & ~mutatorWaitingBit))
ParkingLot::unparkAll(&m_worldState);
continue;
}
if (!(oldState & hasAccessBit)
|| (oldState & stoppedBit)) {
// We can stop the world instantly.
if (m_worldState.compareExchangeWeak(oldState, oldState | stoppedBit | shouldStopBit))
return;
continue;
}
RELEASE_ASSERT(oldState & hasAccessBit);
RELEASE_ASSERT(!(oldState & stoppedBit));
m_worldState.compareExchangeStrong(oldState, oldState | shouldStopBit);
m_stopIfNecessaryTimer->scheduleSoon();
ParkingLot::compareAndPark(&m_worldState, oldState | shouldStopBit);
}
}
void Heap::resumeTheMutator()
{
for (;;) {
unsigned oldState = m_worldState.load();
RELEASE_ASSERT(oldState & shouldStopBit);
if (!(oldState & hasAccessBit)) {
// We can resume the world instantly.
if (m_worldState.compareExchangeWeak(oldState, oldState & ~(stoppedBit | shouldStopBit))) {
ParkingLot::unparkAll(&m_worldState);
return;
}
continue;
}
// We can tell the world to resume.
if (m_worldState.compareExchangeWeak(oldState, oldState & ~shouldStopBit)) {
ParkingLot::unparkAll(&m_worldState);
return;
}
}
}
void Heap::stopIfNecessarySlow()
{
while (stopIfNecessarySlow(m_worldState.load())) { }
handleGCDidJIT();
}
bool Heap::stopIfNecessarySlow(unsigned oldState)
{
RELEASE_ASSERT(oldState & hasAccessBit);
if (handleNeedFinalize(oldState))
return true;
if (!(oldState & shouldStopBit)) {
if (!(oldState & stoppedBit))
return false;
m_worldState.compareExchangeStrong(oldState, oldState & ~stoppedBit);
return true;
}
m_worldState.compareExchangeStrong(oldState, oldState | stoppedBit);
ParkingLot::unparkAll(&m_worldState);
ParkingLot::compareAndPark(&m_worldState, oldState | stoppedBit);
return true;
}
template<typename Func>
void Heap::waitForCollector(const Func& func)
{
for (;;) {
bool done;
{
LockHolder locker(*m_threadLock);
done = func(locker);
if (!done) {
setMutatorWaiting();
// At this point, the collector knows that we intend to wait, and he will clear the
// waiting bit and then unparkAll when the GC cycle finishes. Clearing the bit
// prevents us from parking except if there is also stop-the-world. Unparking after
// clearing means that if the clearing happens after we park, then we will unpark.
}
}
// If we're in a stop-the-world scenario, we need to wait for that even if done is true.
unsigned oldState = m_worldState.load();
if (stopIfNecessarySlow(oldState))
continue;
if (done) {
clearMutatorWaiting(); // Clean up just in case.
return;
}
// If mutatorWaitingBit is still set then we want to wait.
ParkingLot::compareAndPark(&m_worldState, oldState | mutatorWaitingBit);
}
}
void Heap::acquireAccessSlow()
{
for (;;) {
unsigned oldState = m_worldState.load();
RELEASE_ASSERT(!(oldState & hasAccessBit));
if (oldState & shouldStopBit) {
RELEASE_ASSERT(oldState & stoppedBit);
// Wait until we're not stopped anymore.
ParkingLot::compareAndPark(&m_worldState, oldState);
continue;
}
RELEASE_ASSERT(!(oldState & stoppedBit));
unsigned newState = oldState | hasAccessBit;
if (m_worldState.compareExchangeWeak(oldState, newState)) {
handleGCDidJIT();
handleNeedFinalize();
return;
}
}
}
void Heap::releaseAccessSlow()
{
for (;;) {
unsigned oldState = m_worldState.load();
RELEASE_ASSERT(oldState & hasAccessBit);
RELEASE_ASSERT(!(oldState & stoppedBit));
if (handleNeedFinalize(oldState))
continue;
if (oldState & shouldStopBit) {
unsigned newState = (oldState & ~hasAccessBit) | stoppedBit;
if (m_worldState.compareExchangeWeak(oldState, newState)) {
ParkingLot::unparkAll(&m_worldState);
return;
}
continue;
}
RELEASE_ASSERT(!(oldState & shouldStopBit));
if (m_worldState.compareExchangeWeak(oldState, oldState & ~hasAccessBit))
return;
}
}
bool Heap::handleGCDidJIT(unsigned oldState)
{
RELEASE_ASSERT(oldState & hasAccessBit);
if (!(oldState & gcDidJITBit))
return false;
if (m_worldState.compareExchangeWeak(oldState, oldState & ~gcDidJITBit)) {
WTF::crossModifyingCodeFence();
return true;
}
return true;
}
bool Heap::handleNeedFinalize(unsigned oldState)
{
RELEASE_ASSERT(oldState & hasAccessBit);
if (!(oldState & needFinalizeBit))
return false;
if (m_worldState.compareExchangeWeak(oldState, oldState & ~needFinalizeBit)) {
finalize();
// Wake up anyone waiting for us to finalize. Note that they may have woken up already, in
// which case they would be waiting for us to release heap access.
ParkingLot::unparkAll(&m_worldState);
return true;
}
return true;
}
void Heap::handleGCDidJIT()
{
while (handleGCDidJIT(m_worldState.load())) { }
}
void Heap::handleNeedFinalize()
{
while (handleNeedFinalize(m_worldState.load())) { }
}
void Heap::setGCDidJIT()
{
m_worldState.transaction(
[&] (unsigned& state) {
RELEASE_ASSERT(state & stoppedBit);
state |= gcDidJITBit;
});
}
void Heap::setNeedFinalize()
{
m_worldState.exchangeOr(needFinalizeBit);
ParkingLot::unparkAll(&m_worldState);
m_stopIfNecessaryTimer->scheduleSoon();
}
void Heap::waitWhileNeedFinalize()
{
for (;;) {
unsigned oldState = m_worldState.load();
if (!(oldState & needFinalizeBit)) {
// This means that either there was no finalize request or the main thread will finalize
// with heap access, so a subsequent call to stopTheWorld() will return only when
// finalize finishes.
return;
}
ParkingLot::compareAndPark(&m_worldState, oldState);
}
}
void Heap::setMutatorWaiting()
{
m_worldState.exchangeOr(mutatorWaitingBit);
}
void Heap::clearMutatorWaiting()
{
m_worldState.exchangeAnd(~mutatorWaitingBit);
}
void Heap::notifyThreadStopping(const LockHolder&)
{
m_threadIsStopping = true;
clearMutatorWaiting();
ParkingLot::unparkAll(&m_worldState);
}
void Heap::finalize()
{
{
HelpingGCScope helpingGCScope(*this);
deleteUnmarkedCompiledCode();
deleteSourceProviderCaches();
sweepLargeAllocations();
}
if (Options::collectContinuously())
collectAsync();
if (HasOwnPropertyCache* cache = vm()->hasOwnPropertyCache())
cache->clear();
}
Heap::Ticket Heap::requestCollection(std::optional<CollectionScope> scope)
{
stopIfNecessary();
ASSERT(vm()->currentThreadIsHoldingAPILock());
RELEASE_ASSERT(vm()->atomicStringTable() == wtfThreadData().atomicStringTable());
sanitizeStackForVM(m_vm);
LockHolder locker(*m_threadLock);
m_requests.append(scope);
m_lastGrantedTicket++;
m_threadCondition->notifyOne(locker);
return m_lastGrantedTicket;
}
void Heap::waitForCollection(Ticket ticket)
{
waitForCollector(
[&] (const LockHolder&) -> bool {
return m_lastServedTicket >= ticket;
});
}
void Heap::sweepLargeAllocations()
{
m_objectSpace.sweepLargeAllocations();
}
void Heap::suspendCompilerThreads()
{
#if ENABLE(DFG_JIT)
// We ensure the worklists so that it's not possible for the mutator to start a new worklist
// after we have suspended the ones that he had started before. That's not very expensive since
// the worklists use AutomaticThreads anyway.
for (unsigned i = DFG::numberOfWorklists(); i--;)
DFG::ensureWorklistForIndex(i).suspendAllThreads();
#endif
}
void Heap::willStartCollection(std::optional<CollectionScope> scope)
{
if (Options::logGC())
dataLog("=> ");
if (shouldDoFullCollection(scope)) {
m_collectionScope = CollectionScope::Full;
m_shouldDoFullCollection = false;
if (Options::logGC())
dataLog("FullCollection, ");
} else {
m_collectionScope = CollectionScope::Eden;
if (Options::logGC())
dataLog("EdenCollection, ");
}
if (m_collectionScope == CollectionScope::Full) {
m_sizeBeforeLastFullCollect = m_sizeAfterLastCollect + m_bytesAllocatedThisCycle;
m_extraMemorySize = 0;
m_deprecatedExtraMemorySize = 0;
#if ENABLE(RESOURCE_USAGE)
m_externalMemorySize = 0;
#endif
if (m_fullActivityCallback)
m_fullActivityCallback->willCollect();
} else {
ASSERT(m_collectionScope == CollectionScope::Eden);
m_sizeBeforeLastEdenCollect = m_sizeAfterLastCollect + m_bytesAllocatedThisCycle;
}
if (m_edenActivityCallback)
m_edenActivityCallback->willCollect();
for (auto* observer : m_observers)
observer->willGarbageCollect();
}
void Heap::flushWriteBarrierBuffer()
{
m_writeBarrierBuffer.flush(*this);
}
void Heap::prepareForMarking()
{
m_objectSpace.prepareForMarking();
}
void Heap::reapWeakHandles()
{
m_objectSpace.reapWeakSets();
}
void Heap::pruneStaleEntriesFromWeakGCMaps()
{
if (m_collectionScope != CollectionScope::Full)
return;
for (auto& pruneCallback : m_weakGCMaps.values())
pruneCallback();
}
void Heap::sweepArrayBuffers()
{
m_arrayBuffers.sweep();
}
void Heap::snapshotUnswept()
{
TimingScope timingScope(*this, "Heap::snapshotUnswept");
m_objectSpace.snapshotUnswept();
}
void Heap::deleteSourceProviderCaches()
{
m_vm->clearSourceProviderCaches();
}
void Heap::notifyIncrementalSweeper()
{
if (m_collectionScope == CollectionScope::Full) {
if (!m_logicallyEmptyWeakBlocks.isEmpty())
m_indexOfNextLogicallyEmptyWeakBlockToSweep = 0;
}
m_sweeper->startSweeping();
}
void Heap::prepareForAllocation()
{
m_objectSpace.prepareForAllocation();
}
void Heap::updateAllocationLimits()
{
static const bool verbose = false;
if (verbose) {
dataLog("\n");
dataLog("bytesAllocatedThisCycle = ", m_bytesAllocatedThisCycle, "\n");
}
// Calculate our current heap size threshold for the purpose of figuring out when we should
// run another collection. This isn't the same as either size() or capacity(), though it should
// be somewhere between the two. The key is to match the size calculations involved calls to
// didAllocate(), while never dangerously underestimating capacity(). In extreme cases of
// fragmentation, we may have size() much smaller than capacity().
size_t currentHeapSize = 0;
// For marked space, we use the total number of bytes visited. This matches the logic for
// MarkedAllocator's calls to didAllocate(), which effectively accounts for the total size of
// objects allocated rather than blocks used. This will underestimate capacity(), and in case
// of fragmentation, this may be substantial. Fortunately, marked space rarely fragments because
// cells usually have a narrow range of sizes. So, the underestimation is probably OK.
currentHeapSize += m_totalBytesVisited;
if (verbose)
dataLog("totalBytesVisited = ", m_totalBytesVisited, ", currentHeapSize = ", currentHeapSize, "\n");
// It's up to the user to ensure that extraMemorySize() ends up corresponding to allocation-time
// extra memory reporting.
currentHeapSize += extraMemorySize();
if (verbose)
dataLog("extraMemorySize() = ", extraMemorySize(), ", currentHeapSize = ", currentHeapSize, "\n");
if (Options::gcMaxHeapSize() && currentHeapSize > Options::gcMaxHeapSize())
HeapStatistics::exitWithFailure();
if (m_collectionScope == CollectionScope::Full) {
// To avoid pathological GC churn in very small and very large heaps, we set
// the new allocation limit based on the current size of the heap, with a
// fixed minimum.
m_maxHeapSize = max(minHeapSize(m_heapType, m_ramSize), proportionalHeapSize(currentHeapSize, m_ramSize));
if (verbose)
dataLog("Full: maxHeapSize = ", m_maxHeapSize, "\n");
m_maxEdenSize = m_maxHeapSize - currentHeapSize;
if (verbose)
dataLog("Full: maxEdenSize = ", m_maxEdenSize, "\n");
m_sizeAfterLastFullCollect = currentHeapSize;
if (verbose)
dataLog("Full: sizeAfterLastFullCollect = ", currentHeapSize, "\n");
m_bytesAbandonedSinceLastFullCollect = 0;
if (verbose)
dataLog("Full: bytesAbandonedSinceLastFullCollect = ", 0, "\n");
} else {
ASSERT(currentHeapSize >= m_sizeAfterLastCollect);
// Theoretically, we shouldn't ever scan more memory than the heap size we planned to have.
// But we are sloppy, so we have to defend against the overflow.
m_maxEdenSize = currentHeapSize > m_maxHeapSize ? 0 : m_maxHeapSize - currentHeapSize;
if (verbose)
dataLog("Eden: maxEdenSize = ", m_maxEdenSize, "\n");
m_sizeAfterLastEdenCollect = currentHeapSize;
if (verbose)
dataLog("Eden: sizeAfterLastEdenCollect = ", currentHeapSize, "\n");
double edenToOldGenerationRatio = (double)m_maxEdenSize / (double)m_maxHeapSize;
double minEdenToOldGenerationRatio = 1.0 / 3.0;
if (edenToOldGenerationRatio < minEdenToOldGenerationRatio)
m_shouldDoFullCollection = true;
// This seems suspect at first, but what it does is ensure that the nursery size is fixed.
m_maxHeapSize += currentHeapSize - m_sizeAfterLastCollect;
if (verbose)
dataLog("Eden: maxHeapSize = ", m_maxHeapSize, "\n");
m_maxEdenSize = m_maxHeapSize - currentHeapSize;
if (verbose)
dataLog("Eden: maxEdenSize = ", m_maxEdenSize, "\n");
if (m_fullActivityCallback) {
ASSERT(currentHeapSize >= m_sizeAfterLastFullCollect);
m_fullActivityCallback->didAllocate(currentHeapSize - m_sizeAfterLastFullCollect);
}
}
m_sizeAfterLastCollect = currentHeapSize;
if (verbose)
dataLog("sizeAfterLastCollect = ", m_sizeAfterLastCollect, "\n");
m_bytesAllocatedThisCycle = 0;
if (Options::logGC())
dataLog("=> ", currentHeapSize / 1024, " kb, ");
}
void Heap::didFinishCollection(double gcStartTime)
{
double gcEndTime = WTF::monotonicallyIncreasingTime();
CollectionScope scope = *m_collectionScope;
if (scope == CollectionScope::Full)
m_lastFullGCLength = gcEndTime - gcStartTime;
else
m_lastEdenGCLength = gcEndTime - gcStartTime;
#if ENABLE(RESOURCE_USAGE)
ASSERT(externalMemorySize() <= extraMemorySize());
#endif
if (Options::recordGCPauseTimes())
HeapStatistics::recordGCPauseTime(gcStartTime, gcEndTime);
if (Options::useZombieMode())
zombifyDeadObjects();
if (Options::dumpObjectStatistics())
HeapStatistics::dumpObjectStatistics(this);
if (HeapProfiler* heapProfiler = m_vm->heapProfiler()) {
gatherExtraHeapSnapshotData(*heapProfiler);
removeDeadHeapSnapshotNodes(*heapProfiler);
}
RELEASE_ASSERT(m_collectionScope);
m_lastCollectionScope = m_collectionScope;
m_collectionScope = std::nullopt;
for (auto* observer : m_observers)
observer->didGarbageCollect(scope);
}
void Heap::resumeCompilerThreads()
{
#if ENABLE(DFG_JIT)
for (unsigned i = DFG::numberOfWorklists(); i--;)
DFG::existingWorklistForIndex(i).resumeAllThreads();
#endif
}
GCActivityCallback* Heap::fullActivityCallback()
{
return m_fullActivityCallback.get();
}
GCActivityCallback* Heap::edenActivityCallback()
{
return m_edenActivityCallback.get();
}
IncrementalSweeper* Heap::sweeper()
{
return m_sweeper.get();
}
void Heap::setGarbageCollectionTimerEnabled(bool enable)
{
if (m_fullActivityCallback)
m_fullActivityCallback->setEnabled(enable);
if (m_edenActivityCallback)
m_edenActivityCallback->setEnabled(enable);
}
void Heap::didAllocate(size_t bytes)
{
if (m_edenActivityCallback)
m_edenActivityCallback->didAllocate(m_bytesAllocatedThisCycle + m_bytesAbandonedSinceLastFullCollect);
m_bytesAllocatedThisCycle += bytes;
}
bool Heap::isValidAllocation(size_t)
{
if (!isValidThreadState(m_vm))
return false;
if (isCurrentThreadBusy())
return false;
return true;
}
void Heap::addFinalizer(JSCell* cell, Finalizer finalizer)
{
WeakSet::allocate(cell, &m_finalizerOwner, reinterpret_cast<void*>(finalizer)); // Balanced by FinalizerOwner::finalize().
}
void Heap::FinalizerOwner::finalize(Handle<Unknown> handle, void* context)
{
HandleSlot slot = handle.slot();
Finalizer finalizer = reinterpret_cast<Finalizer>(context);
finalizer(slot->asCell());
WeakSet::deallocate(WeakImpl::asWeakImpl(slot));
}
void Heap::addExecutable(ExecutableBase* executable)
{
m_executables.append(executable);
}
void Heap::collectAllGarbageIfNotDoneRecently()
{
if (!m_fullActivityCallback) {
collectAllGarbage();
return;
}
if (m_fullActivityCallback->didSyncGCRecently()) {
// A synchronous GC was already requested recently so we merely accelerate next collection.
reportAbandonedObjectGraph();
return;
}
m_fullActivityCallback->setDidSyncGCRecently();
collectAllGarbage();
}
class Zombify : public MarkedBlock::VoidFunctor {
public:
inline void visit(HeapCell* cell) const
{
void** current = reinterpret_cast_ptr<void**>(cell);
// We want to maintain zapped-ness because that's how we know if we've called
// the destructor.
if (cell->isZapped())
current++;
void* limit = static_cast<void*>(reinterpret_cast<char*>(cell) + cell->cellSize());
for (; current < limit; current++)
*current = zombifiedBits;
}
IterationStatus operator()(HeapCell* cell, HeapCell::Kind) const
{
visit(cell);
return IterationStatus::Continue;
}
};
void Heap::zombifyDeadObjects()
{
// Sweep now because destructors will crash once we're zombified.
m_objectSpace.sweep();
HeapIterationScope iterationScope(*this);
m_objectSpace.forEachDeadCell(iterationScope, Zombify());
}
void Heap::flushWriteBarrierBuffer(JSCell* cell)
{
m_writeBarrierBuffer.flush(*this);
m_writeBarrierBuffer.add(cell);
}
bool Heap::shouldDoFullCollection(std::optional<CollectionScope> scope) const
{
if (!Options::useGenerationalGC())
return true;
if (!scope)
return m_shouldDoFullCollection;
return *scope == CollectionScope::Full;
}
void Heap::addLogicallyEmptyWeakBlock(WeakBlock* block)
{
m_logicallyEmptyWeakBlocks.append(block);
}
void Heap::sweepAllLogicallyEmptyWeakBlocks()
{
if (m_logicallyEmptyWeakBlocks.isEmpty())
return;
m_indexOfNextLogicallyEmptyWeakBlockToSweep = 0;
while (sweepNextLogicallyEmptyWeakBlock()) { }
}
bool Heap::sweepNextLogicallyEmptyWeakBlock()
{
if (m_indexOfNextLogicallyEmptyWeakBlockToSweep == WTF::notFound)
return false;
WeakBlock* block = m_logicallyEmptyWeakBlocks[m_indexOfNextLogicallyEmptyWeakBlockToSweep];
block->sweep();
if (block->isEmpty()) {
std::swap(m_logicallyEmptyWeakBlocks[m_indexOfNextLogicallyEmptyWeakBlockToSweep], m_logicallyEmptyWeakBlocks.last());
m_logicallyEmptyWeakBlocks.removeLast();
WeakBlock::destroy(*this, block);
} else
m_indexOfNextLogicallyEmptyWeakBlockToSweep++;
if (m_indexOfNextLogicallyEmptyWeakBlockToSweep >= m_logicallyEmptyWeakBlocks.size()) {
m_indexOfNextLogicallyEmptyWeakBlockToSweep = WTF::notFound;
return false;
}
return true;
}
size_t Heap::threadVisitCount()
{
unsigned long result = 0;
for (auto& parallelVisitor : m_parallelSlotVisitors)
result += parallelVisitor->visitCount();
return result;
}
size_t Heap::threadBytesVisited()
{
size_t result = 0;
for (auto& parallelVisitor : m_parallelSlotVisitors)
result += parallelVisitor->bytesVisited();
return result;
}
void Heap::forEachCodeBlockImpl(const ScopedLambda<bool(CodeBlock*)>& func)
{
// We don't know the full set of CodeBlocks until compilation has terminated.
completeAllJITPlans();
return m_codeBlocks->iterate(func);
}
void Heap::writeBarrierSlowPath(const JSCell* from)
{
if (UNLIKELY(mutatorShouldBeFenced())) {
// In this case, the barrierThreshold is the tautological threshold, so from could still be
// not black. But we can't know for sure until we fire off a fence.
WTF::storeLoadFence();
if (from->cellState() != CellState::AnthraciteOrBlack)
return;
}
addToRememberedSet(from);
}
bool Heap::canCollect()
{
if (isDeferred())
return false;
if (!m_isSafeToCollect)
return false;
if (mutatorState() == MutatorState::HelpingGC)
return false;
return true;
}
bool Heap::shouldCollectHeuristic()
{
if (Options::gcMaxHeapSize())
return m_bytesAllocatedThisCycle > Options::gcMaxHeapSize();
return m_bytesAllocatedThisCycle > m_maxEdenSize;
}
bool Heap::shouldCollect()
{
return canCollect() && shouldCollectHeuristic();
}
bool Heap::isCurrentThreadBusy()
{
return mayBeGCThread() || mutatorState() != MutatorState::Running;
}
void Heap::reportExtraMemoryVisited(size_t size)
{
size_t* counter = &m_extraMemorySize;
for (;;) {
size_t oldSize = *counter;
if (WTF::atomicCompareExchangeWeakRelaxed(counter, oldSize, oldSize + size))
return;
}
}
#if ENABLE(RESOURCE_USAGE)
void Heap::reportExternalMemoryVisited(size_t size)
{
size_t* counter = &m_externalMemorySize;
for (;;) {
size_t oldSize = *counter;
if (WTF::atomicCompareExchangeWeakRelaxed(counter, oldSize, oldSize + size))
return;
}
}
#endif
bool Heap::collectIfNecessaryOrDefer(GCDeferralContext* deferralContext)
{
if (!canCollect())
return false;
if (deferralContext) {
deferralContext->m_shouldGC |=
!!(m_worldState.load() & (shouldStopBit | needFinalizeBit | gcDidJITBit));
} else
stopIfNecessary();
if (!shouldCollectHeuristic())
return false;
if (deferralContext)
deferralContext->m_shouldGC = true;
else
collectAsync();
return true;
}
void Heap::collectAccordingToDeferGCProbability()
{
if (isDeferred() || !m_isSafeToCollect || collectionScope() || mutatorState() == MutatorState::HelpingGC)
return;
if (randomNumber() < Options::deferGCProbability()) {
collectAsync();
return;
}
// If our coin flip told us not to GC, we still might GC,
// but we GC according to our memory pressure markers.
collectIfNecessaryOrDefer();
}
void Heap::decrementDeferralDepthAndGCIfNeeded()
{
decrementDeferralDepth();
if (UNLIKELY(Options::deferGCShouldCollectWithProbability()))
collectAccordingToDeferGCProbability();
else
collectIfNecessaryOrDefer();
}
void Heap::registerWeakGCMap(void* weakGCMap, std::function<void()> pruningCallback)
{
m_weakGCMaps.add(weakGCMap, WTFMove(pruningCallback));
}
void Heap::unregisterWeakGCMap(void* weakGCMap)
{
m_weakGCMaps.remove(weakGCMap);
}
void Heap::didAllocateBlock(size_t capacity)
{
#if ENABLE(RESOURCE_USAGE)
m_blockBytesAllocated += capacity;
#else
UNUSED_PARAM(capacity);
#endif
}
void Heap::didFreeBlock(size_t capacity)
{
#if ENABLE(RESOURCE_USAGE)
m_blockBytesAllocated -= capacity;
#else
UNUSED_PARAM(capacity);
#endif
}
#if USE(CF)
void Heap::setRunLoop(CFRunLoopRef runLoop)
{
m_runLoop = runLoop;
m_fullActivityCallback->setRunLoop(runLoop);
m_edenActivityCallback->setRunLoop(runLoop);
m_sweeper->setRunLoop(runLoop);
}
#endif // USE(CF)
void Heap::notifyIsSafeToCollect()
{
m_isSafeToCollect = true;
if (Options::collectContinuously())
collectAsync();
}
} // namespace JSC