| /* |
| * Copyright (C) 2013-2021 Apple Inc. All rights reserved. |
| * |
| * Redistribution and use in source and binary forms, with or without |
| * modification, are permitted provided that the following conditions |
| * are met: |
| * 1. Redistributions of source code must retain the above copyright |
| * notice, this list of conditions and the following disclaimer. |
| * 2. Redistributions in binary form must reproduce the above copyright |
| * notice, this list of conditions and the following disclaimer in the |
| * documentation and/or other materials provided with the distribution. |
| * |
| * THIS SOFTWARE IS PROVIDED BY APPLE INC. AND ITS CONTRIBUTORS ``AS IS'' |
| * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, |
| * THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR |
| * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL APPLE INC. OR ITS CONTRIBUTORS |
| * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR |
| * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF |
| * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS |
| * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN |
| * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) |
| * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF |
| * THE POSSIBILITY OF SUCH DAMAGE. |
| */ |
| |
| #pragma once |
| |
| #include "JSCConfig.h" |
| #include "MarkedBlock.h" |
| #include <wtf/HashTraits.h> |
| #include <wtf/StdIntExtras.h> |
| |
| namespace JSC { |
| |
| class Structure; |
| |
| #if CPU(ADDRESS64) |
| |
| // We would like to define this value in PlatformEnable.h, but it is not possible since the following is relying on MACH_VM_MAX_ADDRESS. |
| #if CPU(ARM64) && OS(DARWIN) |
| #if MACH_VM_MAX_ADDRESS_RAW < (1ULL << 36) |
| #define ENABLE_STRUCTURE_ID_WITH_SHIFT 1 |
| static_assert(MACH_VM_MAX_ADDRESS_RAW == MACH_VM_MAX_ADDRESS); |
| #endif |
| #endif |
| |
| #if !ENABLE(STRUCTURE_ID_WITH_SHIFT) |
| #if defined(STRUCTURE_HEAP_ADDRESS_SIZE_IN_MB) && STRUCTURE_HEAP_ADDRESS_SIZE_IN_MB > 0 |
| constexpr uintptr_t structureHeapAddressSize = STRUCTURE_HEAP_ADDRESS_SIZE_IN_MB * MB; |
| #elif PLATFORM(IOS_FAMILY) && CPU(ARM64) && !CPU(ARM64E) |
| constexpr uintptr_t structureHeapAddressSize = 512 * MB; |
| #else |
| constexpr uintptr_t structureHeapAddressSize = 4 * GB; |
| #endif |
| #endif // !ENABLE(STRUCTURE_ID_WITH_SHIFT) |
| |
| #endif // CPU(ADDRESS64) |
| |
| class StructureID { |
| public: |
| static constexpr uint32_t nukedStructureIDBit = 1; |
| |
| #if ENABLE(STRUCTURE_ID_WITH_SHIFT) |
| // ENABLE(STRUCTURE_ID_WITH_SHIFT) is used when our virtual memory space is limited (specifically, less than or equal to 36 bit) while pointer is 64 bit. |
| // In that case, we round up Structures size with 32 bytes instead of 16 bytes. This ensures that lower 5 bit become zero for Structure. |
| // By shifting this address with 4, we can encode 36 bit address into 32 bit StructureID. And we can ensure that StructureID's lowest bit is still zero |
| // because we round Structure size with 32 bytes. This lowest bit is used for nuke bit. |
| static constexpr unsigned encodeShiftAmount = 4; |
| #elif CPU(ADDRESS64) |
| static constexpr CPURegister structureIDMask = structureHeapAddressSize - 1; |
| #endif |
| |
| StructureID() = default; |
| StructureID(StructureID const&) = default; |
| StructureID& operator=(StructureID const&) = default; |
| |
| StructureID nuke() const { return StructureID(m_bits | nukedStructureIDBit); } |
| bool isNuked() const { return m_bits & nukedStructureIDBit; } |
| StructureID decontaminate() const { return StructureID(m_bits & ~nukedStructureIDBit); } |
| |
| inline Structure* decode() const; |
| inline Structure* tryDecode() const; |
| static StructureID encode(const Structure*); |
| |
| explicit operator bool() const { return !!m_bits; } |
| bool operator==(StructureID const& other) const { return m_bits == other.m_bits; } |
| bool operator!=(StructureID const& other) const { return m_bits != other.m_bits; } |
| constexpr uint32_t bits() const { return m_bits; } |
| |
| StructureID(WTF::HashTableDeletedValueType) : m_bits(nukedStructureIDBit) { } |
| bool isHashTableDeletedValue() const { return *this == StructureID(WTF::HashTableDeletedValue); } |
| |
| private: |
| explicit StructureID(uint32_t bits) : m_bits(bits) { } |
| |
| uint32_t m_bits { 0 }; |
| }; |
| static_assert(sizeof(StructureID) == sizeof(uint32_t)); |
| |
| #if ENABLE(STRUCTURE_ID_WITH_SHIFT) |
| |
| ALWAYS_INLINE Structure* StructureID::decode() const |
| { |
| ASSERT(decontaminate()); |
| return reinterpret_cast<Structure*>(static_cast<uintptr_t>(decontaminate().m_bits) << encodeShiftAmount); |
| } |
| |
| ALWAYS_INLINE Structure* StructureID::tryDecode() const |
| { |
| // Take care to only use the bits from m_bits in the structure's address reservation. |
| uintptr_t address = static_cast<uintptr_t>(decontaminate().m_bits) << encodeShiftAmount; |
| if (address < MarkedBlock::blockSize) |
| return nullptr; |
| return reinterpret_cast<Structure*>(address); |
| } |
| |
| ALWAYS_INLINE StructureID StructureID::encode(const Structure* structure) |
| { |
| ASSERT(structure); |
| auto result = StructureID(reinterpret_cast<uintptr_t>(structure) >> encodeShiftAmount); |
| ASSERT(result.decode() == structure); |
| return result; |
| } |
| |
| #elif CPU(ADDRESS64) |
| |
| ALWAYS_INLINE Structure* StructureID::decode() const |
| { |
| // Take care to only use the bits from m_bits in the structure's address reservation. |
| ASSERT(decontaminate()); |
| return reinterpret_cast<Structure*>((static_cast<uintptr_t>(decontaminate().m_bits) & structureIDMask) + g_jscConfig.startOfStructureHeap); |
| } |
| |
| ALWAYS_INLINE Structure* StructureID::tryDecode() const |
| { |
| // Take care to only use the bits from m_bits in the structure's address reservation. |
| uintptr_t offset = static_cast<uintptr_t>(decontaminate().m_bits); |
| if (offset < MarkedBlock::blockSize || offset >= g_jscConfig.sizeOfStructureHeap) |
| return nullptr; |
| return reinterpret_cast<Structure*>((offset & structureIDMask) + g_jscConfig.startOfStructureHeap); |
| } |
| |
| ALWAYS_INLINE StructureID StructureID::encode(const Structure* structure) |
| { |
| ASSERT(structure); |
| ASSERT(g_jscConfig.startOfStructureHeap <= reinterpret_cast<uintptr_t>(structure) && reinterpret_cast<uintptr_t>(structure) < g_jscConfig.startOfStructureHeap + structureHeapAddressSize); |
| auto result = StructureID(reinterpret_cast<uintptr_t>(structure) & structureIDMask); |
| ASSERT(result.decode() == structure); |
| return result; |
| } |
| |
| #else // CPU(ADDRESS64) |
| |
| ALWAYS_INLINE Structure* StructureID::decode() const |
| { |
| ASSERT(decontaminate()); |
| return reinterpret_cast<Structure*>(decontaminate().m_bits); |
| } |
| |
| ALWAYS_INLINE Structure* StructureID::tryDecode() const |
| { |
| return reinterpret_cast<Structure*>(decontaminate().m_bits); |
| } |
| |
| ALWAYS_INLINE StructureID StructureID::encode(const Structure* structure) |
| { |
| ASSERT(structure); |
| return StructureID(reinterpret_cast<uint32_t>(structure)); |
| } |
| |
| #endif |
| |
| struct StructureIDHash { |
| static unsigned hash(const StructureID& key) { return key.bits(); } |
| static bool equal(const StructureID& a, const StructureID& b) { return a == b; } |
| static constexpr bool safeToCompareToEmptyOrDeleted = true; |
| }; |
| |
| } // namespace JSC |
| |
| namespace WTF { |
| |
| template<typename T> struct DefaultHash; |
| template<> struct DefaultHash<JSC::StructureID> : JSC::StructureIDHash { }; |
| |
| template<typename T> struct HashTraits; |
| template<> struct HashTraits<JSC::StructureID> : SimpleClassHashTraits<JSC::StructureID> { |
| static constexpr bool emptyValueIsZero = true; |
| }; |
| |
| } |