| /* |
| * Copyright (C) 2013, 2014 Apple Inc. All rights reserved. |
| * |
| * Redistribution and use in source and binary forms, with or without |
| * modification, are permitted provided that the following conditions |
| * are met: |
| * 1. Redistributions of source code must retain the above copyright |
| * notice, this list of conditions and the following disclaimer. |
| * 2. Redistributions in binary form must reproduce the above copyright |
| * notice, this list of conditions and the following disclaimer in the |
| * documentation and/or other materials provided with the distribution. |
| * |
| * THIS SOFTWARE IS PROVIDED BY APPLE INC. ``AS IS'' AND ANY |
| * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE |
| * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR |
| * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL APPLE INC. OR |
| * CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, |
| * EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, |
| * PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR |
| * PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY |
| * OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT |
| * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE |
| * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
| */ |
| |
| #include "config.h" |
| #include "DFGWorklist.h" |
| |
| #if ENABLE(DFG_JIT) |
| |
| #include "CodeBlock.h" |
| #include "DeferGC.h" |
| #include "DFGLongLivedState.h" |
| #include "DFGSafepoint.h" |
| #include "JSCInlines.h" |
| #include <mutex> |
| |
| namespace JSC { namespace DFG { |
| |
| Worklist::Worklist(CString worklistName) |
| : m_threadName(toCString(worklistName, " Worker Thread")) |
| , m_numberOfActiveThreads(0) |
| { |
| } |
| |
| Worklist::~Worklist() |
| { |
| { |
| LockHolder locker(m_lock); |
| for (unsigned i = m_threads.size(); i--;) |
| m_queue.append(nullptr); // Use null plan to indicate that we want the thread to terminate. |
| m_planEnqueued.notifyAll(); |
| } |
| for (unsigned i = m_threads.size(); i--;) |
| waitForThreadCompletion(m_threads[i]->m_identifier); |
| ASSERT(!m_numberOfActiveThreads); |
| } |
| |
| void Worklist::finishCreation(unsigned numberOfThreads, int relativePriority) |
| { |
| RELEASE_ASSERT(numberOfThreads); |
| for (unsigned i = numberOfThreads; i--;) { |
| std::unique_ptr<ThreadData> data = std::make_unique<ThreadData>(this); |
| data->m_identifier = createThread(threadFunction, data.get(), m_threadName.data()); |
| if (relativePriority) |
| changeThreadPriority(data->m_identifier, relativePriority); |
| m_threads.append(WTF::move(data)); |
| } |
| } |
| |
| Ref<Worklist> Worklist::create(CString worklistName, unsigned numberOfThreads, int relativePriority) |
| { |
| Ref<Worklist> result = adoptRef(*new Worklist(worklistName)); |
| result->finishCreation(numberOfThreads, relativePriority); |
| return result; |
| } |
| |
| bool Worklist::isActiveForVM(VM& vm) const |
| { |
| LockHolder locker(m_lock); |
| PlanMap::const_iterator end = m_plans.end(); |
| for (PlanMap::const_iterator iter = m_plans.begin(); iter != end; ++iter) { |
| if (&iter->value->vm == &vm) |
| return true; |
| } |
| return false; |
| } |
| |
| void Worklist::enqueue(PassRefPtr<Plan> passedPlan) |
| { |
| RefPtr<Plan> plan = passedPlan; |
| LockHolder locker(m_lock); |
| if (Options::verboseCompilationQueue()) { |
| dump(locker, WTF::dataFile()); |
| dataLog(": Enqueueing plan to optimize ", plan->key(), "\n"); |
| } |
| ASSERT(m_plans.find(plan->key()) == m_plans.end()); |
| m_plans.add(plan->key(), plan); |
| m_queue.append(plan); |
| m_planEnqueued.notifyOne(); |
| } |
| |
| Worklist::State Worklist::compilationState(CompilationKey key) |
| { |
| LockHolder locker(m_lock); |
| PlanMap::iterator iter = m_plans.find(key); |
| if (iter == m_plans.end()) |
| return NotKnown; |
| return iter->value->stage == Plan::Ready ? Compiled : Compiling; |
| } |
| |
| void Worklist::waitUntilAllPlansForVMAreReady(VM& vm) |
| { |
| DeferGC deferGC(vm.heap); |
| // Wait for all of the plans for the given VM to complete. The idea here |
| // is that we want all of the caller VM's plans to be done. We don't care |
| // about any other VM's plans, and we won't attempt to wait on those. |
| // After we release this lock, we know that although other VMs may still |
| // be adding plans, our VM will not be. |
| |
| LockHolder locker(m_lock); |
| |
| if (Options::verboseCompilationQueue()) { |
| dump(locker, WTF::dataFile()); |
| dataLog(": Waiting for all in VM to complete.\n"); |
| } |
| |
| for (;;) { |
| bool allAreCompiled = true; |
| PlanMap::iterator end = m_plans.end(); |
| for (PlanMap::iterator iter = m_plans.begin(); iter != end; ++iter) { |
| if (&iter->value->vm != &vm) |
| continue; |
| if (iter->value->stage != Plan::Ready) { |
| allAreCompiled = false; |
| break; |
| } |
| } |
| |
| if (allAreCompiled) |
| break; |
| |
| m_planCompiled.wait(m_lock); |
| } |
| } |
| |
| void Worklist::removeAllReadyPlansForVM(VM& vm, Vector<RefPtr<Plan>, 8>& myReadyPlans) |
| { |
| DeferGC deferGC(vm.heap); |
| LockHolder locker(m_lock); |
| for (size_t i = 0; i < m_readyPlans.size(); ++i) { |
| RefPtr<Plan> plan = m_readyPlans[i]; |
| if (&plan->vm != &vm) |
| continue; |
| if (plan->stage != Plan::Ready) |
| continue; |
| myReadyPlans.append(plan); |
| m_readyPlans[i--] = m_readyPlans.last(); |
| m_readyPlans.removeLast(); |
| m_plans.remove(plan->key()); |
| } |
| } |
| |
| void Worklist::removeAllReadyPlansForVM(VM& vm) |
| { |
| Vector<RefPtr<Plan>, 8> myReadyPlans; |
| removeAllReadyPlansForVM(vm, myReadyPlans); |
| } |
| |
| Worklist::State Worklist::completeAllReadyPlansForVM(VM& vm, CompilationKey requestedKey) |
| { |
| DeferGC deferGC(vm.heap); |
| Vector<RefPtr<Plan>, 8> myReadyPlans; |
| |
| removeAllReadyPlansForVM(vm, myReadyPlans); |
| |
| State resultingState = NotKnown; |
| |
| while (!myReadyPlans.isEmpty()) { |
| RefPtr<Plan> plan = myReadyPlans.takeLast(); |
| CompilationKey currentKey = plan->key(); |
| |
| if (Options::verboseCompilationQueue()) |
| dataLog(*this, ": Completing ", currentKey, "\n"); |
| |
| RELEASE_ASSERT(plan->stage == Plan::Ready); |
| |
| plan->finalizeAndNotifyCallback(); |
| |
| if (currentKey == requestedKey) |
| resultingState = Compiled; |
| } |
| |
| if (!!requestedKey && resultingState == NotKnown) { |
| LockHolder locker(m_lock); |
| if (m_plans.contains(requestedKey)) |
| resultingState = Compiling; |
| } |
| |
| return resultingState; |
| } |
| |
| void Worklist::completeAllPlansForVM(VM& vm) |
| { |
| DeferGC deferGC(vm.heap); |
| waitUntilAllPlansForVMAreReady(vm); |
| completeAllReadyPlansForVM(vm); |
| } |
| |
| void Worklist::rememberCodeBlocks(VM& vm) |
| { |
| LockHolder locker(m_lock); |
| for (PlanMap::iterator iter = m_plans.begin(); iter != m_plans.end(); ++iter) { |
| Plan* plan = iter->value.get(); |
| if (&plan->vm != &vm) |
| continue; |
| plan->rememberCodeBlocks(); |
| } |
| } |
| |
| void Worklist::suspendAllThreads() |
| { |
| m_suspensionLock.lock(); |
| for (unsigned i = m_threads.size(); i--;) |
| m_threads[i]->m_rightToRun.lock(); |
| } |
| |
| void Worklist::resumeAllThreads() |
| { |
| for (unsigned i = m_threads.size(); i--;) |
| m_threads[i]->m_rightToRun.unlock(); |
| m_suspensionLock.unlock(); |
| } |
| |
| void Worklist::visitWeakReferences(SlotVisitor& visitor) |
| { |
| VM* vm = visitor.heap()->vm(); |
| { |
| LockHolder locker(m_lock); |
| for (PlanMap::iterator iter = m_plans.begin(); iter != m_plans.end(); ++iter) { |
| Plan* plan = iter->value.get(); |
| if (&plan->vm != vm) |
| continue; |
| plan->checkLivenessAndVisitChildren(visitor); |
| } |
| } |
| // This loop doesn't need locking because: |
| // (1) no new threads can be added to m_threads. Hence, it is immutable and needs no locks. |
| // (2) ThreadData::m_safepoint is protected by that thread's m_rightToRun which we must be |
| // holding here because of a prior call to suspendAllThreads(). |
| for (unsigned i = m_threads.size(); i--;) { |
| ThreadData* data = m_threads[i].get(); |
| Safepoint* safepoint = data->m_safepoint; |
| if (safepoint && &safepoint->vm() == vm) |
| safepoint->checkLivenessAndVisitChildren(visitor); |
| } |
| } |
| |
| void Worklist::removeDeadPlans(VM& vm) |
| { |
| { |
| LockHolder locker(m_lock); |
| HashSet<CompilationKey> deadPlanKeys; |
| for (PlanMap::iterator iter = m_plans.begin(); iter != m_plans.end(); ++iter) { |
| Plan* plan = iter->value.get(); |
| if (&plan->vm != &vm) |
| continue; |
| if (plan->isKnownToBeLiveDuringGC()) |
| continue; |
| RELEASE_ASSERT(plan->stage != Plan::Cancelled); // Should not be cancelled, yet. |
| ASSERT(!deadPlanKeys.contains(plan->key())); |
| deadPlanKeys.add(plan->key()); |
| } |
| if (!deadPlanKeys.isEmpty()) { |
| for (HashSet<CompilationKey>::iterator iter = deadPlanKeys.begin(); iter != deadPlanKeys.end(); ++iter) |
| m_plans.take(*iter)->cancel(); |
| Deque<RefPtr<Plan>> newQueue; |
| while (!m_queue.isEmpty()) { |
| RefPtr<Plan> plan = m_queue.takeFirst(); |
| if (plan->stage != Plan::Cancelled) |
| newQueue.append(plan); |
| } |
| m_queue.swap(newQueue); |
| for (unsigned i = 0; i < m_readyPlans.size(); ++i) { |
| if (m_readyPlans[i]->stage != Plan::Cancelled) |
| continue; |
| m_readyPlans[i] = m_readyPlans.last(); |
| m_readyPlans.removeLast(); |
| } |
| } |
| } |
| |
| // No locking needed for this part, see comment in visitWeakReferences(). |
| for (unsigned i = m_threads.size(); i--;) { |
| ThreadData* data = m_threads[i].get(); |
| Safepoint* safepoint = data->m_safepoint; |
| if (!safepoint) |
| continue; |
| if (&safepoint->vm() != &vm) |
| continue; |
| if (safepoint->isKnownToBeLiveDuringGC()) |
| continue; |
| safepoint->cancel(); |
| } |
| } |
| |
| size_t Worklist::queueLength() |
| { |
| LockHolder locker(m_lock); |
| return m_queue.size(); |
| } |
| |
| void Worklist::dump(PrintStream& out) const |
| { |
| LockHolder locker(m_lock); |
| dump(locker, out); |
| } |
| |
| void Worklist::dump(const LockHolder&, PrintStream& out) const |
| { |
| out.print( |
| "Worklist(", RawPointer(this), ")[Queue Length = ", m_queue.size(), |
| ", Map Size = ", m_plans.size(), ", Num Ready = ", m_readyPlans.size(), |
| ", Num Active Threads = ", m_numberOfActiveThreads, "/", m_threads.size(), "]"); |
| } |
| |
| void Worklist::runThread(ThreadData* data) |
| { |
| CompilationScope compilationScope; |
| |
| if (Options::verboseCompilationQueue()) |
| dataLog(*this, ": Thread started\n"); |
| |
| LongLivedState longLivedState; |
| |
| for (;;) { |
| RefPtr<Plan> plan; |
| { |
| LockHolder locker(m_lock); |
| while (m_queue.isEmpty()) |
| m_planEnqueued.wait(m_lock); |
| |
| plan = m_queue.takeFirst(); |
| if (plan) |
| m_numberOfActiveThreads++; |
| } |
| |
| if (!plan) { |
| if (Options::verboseCompilationQueue()) |
| dataLog(*this, ": Thread shutting down\n"); |
| return; |
| } |
| |
| { |
| LockHolder locker(data->m_rightToRun); |
| { |
| LockHolder locker(m_lock); |
| if (plan->stage == Plan::Cancelled) { |
| m_numberOfActiveThreads--; |
| continue; |
| } |
| plan->notifyCompiling(); |
| } |
| |
| if (Options::verboseCompilationQueue()) |
| dataLog(*this, ": Compiling ", plan->key(), " asynchronously\n"); |
| |
| RELEASE_ASSERT(!plan->vm.heap.isCollecting()); |
| plan->compileInThread(longLivedState, data); |
| RELEASE_ASSERT(!plan->vm.heap.isCollecting()); |
| |
| { |
| LockHolder locker(m_lock); |
| if (plan->stage == Plan::Cancelled) { |
| m_numberOfActiveThreads--; |
| continue; |
| } |
| plan->notifyCompiled(); |
| } |
| RELEASE_ASSERT(!plan->vm.heap.isCollecting()); |
| } |
| |
| { |
| LockHolder locker(m_lock); |
| |
| // We could have been cancelled between releasing rightToRun and acquiring m_lock. |
| // This would mean that we might be in the middle of GC right now. |
| if (plan->stage == Plan::Cancelled) { |
| m_numberOfActiveThreads--; |
| continue; |
| } |
| |
| plan->notifyReady(); |
| |
| if (Options::verboseCompilationQueue()) { |
| dump(locker, WTF::dataFile()); |
| dataLog(": Compiled ", plan->key(), " asynchronously\n"); |
| } |
| |
| m_readyPlans.append(plan); |
| |
| m_planCompiled.notifyAll(); |
| m_numberOfActiveThreads--; |
| } |
| } |
| } |
| |
| void Worklist::threadFunction(void* argument) |
| { |
| ThreadData* data = static_cast<ThreadData*>(argument); |
| data->m_worklist->runThread(data); |
| } |
| |
| static Worklist* theGlobalDFGWorklist; |
| |
| Worklist* ensureGlobalDFGWorklist() |
| { |
| static std::once_flag initializeGlobalWorklistOnceFlag; |
| std::call_once(initializeGlobalWorklistOnceFlag, [] { |
| theGlobalDFGWorklist = &Worklist::create("DFG Worklist", Options::numberOfDFGCompilerThreads(), Options::priorityDeltaOfDFGCompilerThreads()).leakRef(); |
| }); |
| return theGlobalDFGWorklist; |
| } |
| |
| Worklist* existingGlobalDFGWorklistOrNull() |
| { |
| return theGlobalDFGWorklist; |
| } |
| |
| static Worklist* theGlobalFTLWorklist; |
| |
| Worklist* ensureGlobalFTLWorklist() |
| { |
| static std::once_flag initializeGlobalWorklistOnceFlag; |
| std::call_once(initializeGlobalWorklistOnceFlag, [] { |
| theGlobalFTLWorklist = &Worklist::create("FTL Worklist", Options::numberOfFTLCompilerThreads(), Options::priorityDeltaOfFTLCompilerThreads()).leakRef(); |
| }); |
| return theGlobalFTLWorklist; |
| } |
| |
| Worklist* existingGlobalFTLWorklistOrNull() |
| { |
| return theGlobalFTLWorklist; |
| } |
| |
| Worklist* ensureGlobalWorklistFor(CompilationMode mode) |
| { |
| switch (mode) { |
| case InvalidCompilationMode: |
| RELEASE_ASSERT_NOT_REACHED(); |
| return 0; |
| case DFGMode: |
| return ensureGlobalDFGWorklist(); |
| case FTLMode: |
| case FTLForOSREntryMode: |
| return ensureGlobalFTLWorklist(); |
| } |
| RELEASE_ASSERT_NOT_REACHED(); |
| return 0; |
| } |
| |
| void completeAllPlansForVM(VM& vm) |
| { |
| for (unsigned i = DFG::numberOfWorklists(); i--;) { |
| if (DFG::Worklist* worklist = DFG::worklistForIndexOrNull(i)) |
| worklist->completeAllPlansForVM(vm); |
| } |
| } |
| |
| void rememberCodeBlocks(VM& vm) |
| { |
| for (unsigned i = DFG::numberOfWorklists(); i--;) { |
| if (DFG::Worklist* worklist = DFG::worklistForIndexOrNull(i)) |
| worklist->rememberCodeBlocks(vm); |
| } |
| } |
| |
| } } // namespace JSC::DFG |
| |
| #endif // ENABLE(DFG_JIT) |
| |