| /* |
| * Copyright (C) 2017-2019 Apple Inc. All rights reserved. |
| * |
| * Redistribution and use in source and binary forms, with or without |
| * modification, are permitted provided that the following conditions |
| * are met: |
| * 1. Redistributions of source code must retain the above copyright |
| * notice, this list of conditions and the following disclaimer. |
| * 2. Redistributions in binary form must reproduce the above copyright |
| * notice, this list of conditions and the following disclaimer in the |
| * documentation and/or other materials provided with the distribution. |
| * |
| * THIS SOFTWARE IS PROVIDED BY APPLE INC. ``AS IS'' AND ANY |
| * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE |
| * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR |
| * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL APPLE INC. OR |
| * CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, |
| * EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, |
| * PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR |
| * PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY |
| * OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT |
| * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE |
| * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
| */ |
| |
| #pragma once |
| |
| #include "BMalloced.h" |
| #include "IsoAllocator.h" |
| #include "IsoDirectoryPage.h" |
| #include "IsoTLSAllocatorEntry.h" |
| #include "PhysicalPageMap.h" |
| |
| namespace bmalloc { |
| |
| class AllIsoHeaps; |
| |
| class BEXPORT IsoHeapImplBase { |
| MAKE_BMALLOCED; |
| IsoHeapImplBase(const IsoHeapImplBase&) = delete; |
| IsoHeapImplBase& operator=(const IsoHeapImplBase&) = delete; |
| public: |
| static constexpr unsigned maxAllocationFromShared = 8; |
| static constexpr unsigned maxAllocationFromSharedMask = (1U << maxAllocationFromShared) - 1U; |
| static_assert(maxAllocationFromShared <= bmalloc::alignment, ""); |
| static_assert(isPowerOfTwo(maxAllocationFromShared), ""); |
| |
| virtual ~IsoHeapImplBase(); |
| |
| virtual void scavenge(Vector<DeferredDecommit>&) = 0; |
| #if BPLATFORM(MAC) |
| virtual void scavengeToHighWatermark(Vector<DeferredDecommit>&) = 0; |
| #endif |
| virtual size_t freeableMemory() = 0; |
| virtual size_t footprint() = 0; |
| |
| void scavengeNow(); |
| static void finishScavenging(Vector<DeferredDecommit>&); |
| |
| void addToAllIsoHeaps(); |
| |
| protected: |
| IsoHeapImplBase(); |
| |
| friend class IsoSharedPage; |
| friend class AllIsoHeaps; |
| |
| IsoHeapImplBase* m_next { nullptr }; |
| std::chrono::steady_clock::time_point m_lastSlowPathTime; |
| std::array<void*, maxAllocationFromShared> m_sharedCells { }; |
| unsigned m_numberOfAllocationsFromSharedInOneCycle { 0 }; |
| unsigned m_availableShared { maxAllocationFromSharedMask }; |
| AllocationMode m_allocationMode { AllocationMode::Init }; |
| |
| static_assert(sizeof(m_availableShared) * 8 >= maxAllocationFromShared, ""); |
| }; |
| |
| template<typename Config> |
| class IsoHeapImpl final : public IsoHeapImplBase { |
| // Pick a size that makes us most efficiently use the bitvectors. |
| static constexpr unsigned numPagesInInlineDirectory = 32; |
| |
| public: |
| IsoHeapImpl(); |
| |
| EligibilityResult<Config> takeFirstEligible(); |
| |
| // Callbacks from directory. |
| void didBecomeEligibleOrDecommited(IsoDirectory<Config, numPagesInInlineDirectory>*); |
| void didBecomeEligibleOrDecommited(IsoDirectory<Config, IsoDirectoryPage<Config>::numPages>*); |
| |
| void scavenge(Vector<DeferredDecommit>&) override; |
| #if BPLATFORM(MAC) |
| void scavengeToHighWatermark(Vector<DeferredDecommit>&) override; |
| #endif |
| |
| size_t freeableMemory() override; |
| |
| size_t footprint() override; |
| |
| unsigned allocatorOffset(); |
| unsigned deallocatorOffset(); |
| |
| // White-box testing functions. |
| unsigned numLiveObjects(); |
| unsigned numCommittedPages(); |
| |
| template<typename Func> |
| void forEachDirectory(const Func&); |
| |
| template<typename Func> |
| void forEachCommittedPage(const Func&); |
| |
| // This is only accurate when all threads are scavenged. Otherwise it will overestimate. |
| template<typename Func> |
| void forEachLiveObject(const Func&); |
| |
| void didCommit(void* ptr, size_t bytes); |
| void didDecommit(void* ptr, size_t bytes); |
| |
| void isNowFreeable(void* ptr, size_t bytes); |
| void isNoLongerFreeable(void* ptr, size_t bytes); |
| |
| AllocationMode updateAllocationMode(); |
| void* allocateFromShared(const std::lock_guard<Mutex>&, bool abortOnFailure); |
| |
| // It's almost always the caller's responsibility to grab the lock. This lock comes from the |
| // (*PerProcess<IsoTLSEntryHolder<IsoTLSDeallocatorEntry<Config>>>::get())->lock. That's pretty weird, and we don't |
| // try to disguise the fact that it's weird. We only do that because heaps in the same size class |
| // share the same deallocator log, so it makes sense for them to also share the same lock to |
| // amortize lock acquisition costs. |
| Mutex& lock; |
| |
| private: |
| IsoDirectory<Config, numPagesInInlineDirectory> m_inlineDirectory; |
| IsoDirectoryPage<Config>* m_headDirectory { nullptr }; |
| IsoDirectoryPage<Config>* m_tailDirectory { nullptr }; |
| size_t m_footprint { 0 }; |
| size_t m_freeableMemory { 0 }; |
| #if ENABLE_PHYSICAL_PAGE_MAP |
| PhysicalPageMap m_physicalPageMap; |
| #endif |
| unsigned m_nextDirectoryPageIndex { 1 }; // We start at 1 so that the high water mark being zero means we've only allocated in the inline directory since the last scavenge. |
| unsigned m_directoryHighWatermark { 0 }; |
| |
| bool m_isInlineDirectoryEligibleOrDecommitted { true }; |
| IsoDirectoryPage<Config>* m_firstEligibleOrDecommitedDirectory { nullptr }; |
| |
| IsoTLSEntryHolder<IsoTLSAllocatorEntry<Config>> m_allocator; |
| }; |
| |
| } // namespace bmalloc |
| |
| |