blob: 8f532347a63fed930c565f7f690ef19da69d2d68 [file] [log] [blame]
/*
* Copyright (C) 2021 Apple Inc. All rights reserved.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
* 1. Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
* 2. Redistributions in binary form must reproduce the above copyright
* notice, this list of conditions and the following disclaimer in the
* documentation and/or other materials provided with the distribution.
*
* THIS SOFTWARE IS PROVIDED BY APPLE INC. ``AS IS'' AND ANY
* EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
* PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL APPLE INC. OR
* CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL,
* EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO,
* PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR
* PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY
* OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
*/
#pragma once
#include "ExceptionExpectation.h"
#include "HashMapImpl.h"
#include "JSCJSValueInlines.h"
#include "VMTrapsInlines.h"
namespace JSC {
ALWAYS_INLINE bool areKeysEqual(JSGlobalObject* globalObject, JSValue a, JSValue b)
{
// We want +0 and -0 to be compared to true here. sameValue() itself doesn't
// guarantee that, however, we normalize all keys before comparing and storing
// them in the map. The normalization will convert -0.0 and 0.0 to the integer
// representation for 0.
return sameValue(globalObject, a, b);
}
// Note that normalization is inlined in DFG's NormalizeMapKey.
// Keep in sync with the implementation of DFG and FTL normalization.
ALWAYS_INLINE JSValue normalizeMapKey(JSValue key)
{
if (!key.isNumber()) {
if (key.isHeapBigInt())
return tryConvertToBigInt32(key.asHeapBigInt());
return key;
}
if (key.isInt32())
return key;
double d = key.asDouble();
if (std::isnan(d))
return jsNaN();
int i = static_cast<int>(d);
if (i == d) {
// When a key is -0, we convert it to positive zero.
// When a key is the double representation for an integer, we convert it to an integer.
return jsNumber(i);
}
// This means key is definitely not negative zero, and it's definitely not a double representation of an integer.
return key;
}
ALWAYS_INLINE uint32_t wangsInt64Hash(uint64_t key)
{
key += ~(key << 32);
key ^= (key >> 22);
key += ~(key << 13);
key ^= (key >> 8);
key += (key << 3);
key ^= (key >> 15);
key += ~(key << 27);
key ^= (key >> 31);
return static_cast<unsigned>(key);
}
ALWAYS_INLINE uint32_t jsMapHash(JSBigInt* bigInt)
{
return bigInt->hash();
}
template<ExceptionExpectation expection>
ALWAYS_INLINE uint32_t jsMapHashImpl(JSGlobalObject* globalObject, VM& vm, JSValue value)
{
ASSERT_WITH_MESSAGE(normalizeMapKey(value) == value, "We expect normalized values flowing into this function.");
if (value.isString()) {
auto scope = DECLARE_THROW_SCOPE(vm);
const String& wtfString = asString(value)->value(globalObject);
if constexpr (expection == ExceptionExpectation::CanThrow)
RETURN_IF_EXCEPTION(scope, UINT_MAX);
else
EXCEPTION_ASSERT_UNUSED(scope, !scope.exception());
return wtfString.impl()->hash();
}
if (value.isHeapBigInt())
return jsMapHash(value.asHeapBigInt());
return wangsInt64Hash(JSValue::encode(value));
}
ALWAYS_INLINE uint32_t jsMapHash(JSGlobalObject* globalObject, VM& vm, JSValue value)
{
return jsMapHashImpl<ExceptionExpectation::CanThrow>(globalObject, vm, value);
}
ALWAYS_INLINE uint32_t jsMapHashForAlreadyHashedValue(JSGlobalObject* globalObject, VM& vm, JSValue value)
{
return jsMapHashImpl<ExceptionExpectation::ShouldNotThrow>(globalObject, vm, value);
}
ALWAYS_INLINE std::optional<uint32_t> concurrentJSMapHash(JSValue key)
{
key = normalizeMapKey(key);
if (key.isString()) {
JSString* string = asString(key);
if (string->length() > 10 * 1024)
return std::nullopt;
const StringImpl* impl = string->tryGetValueImpl();
if (!impl)
return std::nullopt;
return impl->concurrentHash();
}
if (key.isHeapBigInt())
return key.asHeapBigInt()->concurrentHash();
uint64_t rawValue = JSValue::encode(key);
return wangsInt64Hash(rawValue);
}
ALWAYS_INLINE uint32_t shouldShrink(uint32_t capacity, uint32_t keyCount)
{
return 8 * keyCount <= capacity && capacity > 4;
}
ALWAYS_INLINE uint32_t shouldRehashAfterAdd(uint32_t capacity, uint32_t keyCount, uint32_t deleteCount)
{
return 2 * (keyCount + deleteCount) >= capacity;
}
ALWAYS_INLINE uint32_t nextCapacity(uint32_t capacity, uint32_t keyCount)
{
if (shouldShrink(capacity, keyCount)) {
ASSERT((capacity / 2) >= 4);
return capacity / 2;
}
if (3 * keyCount <= capacity && capacity > 64) {
// We stay at the same size if rehashing would cause us to be no more than
// 1/3rd full. This comes up for programs like this:
// Say the hash table grew to a key count of 64, causing it to grow to a capacity of 256.
// Then, the table added 63 items. The load is now 127. Then, 63 items are deleted.
// The load is still 127. Then, another item is added. The load is now 128, and we
// decide that we need to rehash. The key count is 65, almost exactly what it was
// when we grew to a capacity of 256. We don't really need to grow to a capacity
// of 512 in this situation. Instead, we choose to rehash at the same size. This
// will bring the load down to 65. We rehash into the same size when we determine
// that the new load ratio will be under 1/3rd. (We also pick a minumum capacity
// at which this rule kicks in because otherwise we will be too sensitive to rehashing
// at the same capacity).
return capacity;
}
return Checked<uint32_t>(capacity) * 2;
}
template <typename HashMapBucketType>
void HashMapImpl<HashMapBucketType>::finishCreation(JSGlobalObject* globalObject, VM& vm)
{
ASSERT_WITH_MESSAGE(HashMapBucket<HashMapBucketDataKey>::offsetOfKey() == HashMapBucket<HashMapBucketDataKeyValue>::offsetOfKey(), "We assume this to be true in the DFG and FTL JIT.");
auto scope = DECLARE_THROW_SCOPE(vm);
Base::finishCreation(vm);
makeAndSetNewBuffer(globalObject, vm);
RETURN_IF_EXCEPTION(scope, void());
setUpHeadAndTail(globalObject, vm);
}
template <typename HashMapBucketType>
void HashMapImpl<HashMapBucketType>::finishCreation(JSGlobalObject* globalObject, VM& vm, HashMapImpl* base)
{
auto scope = DECLARE_THROW_SCOPE(vm);
Base::finishCreation(vm);
// This size should be the same to the case when you clone the map by calling add() repeatedly.
uint32_t capacity = (Checked<uint32_t>(base->m_keyCount) * 2) + 1;
RELEASE_ASSERT(capacity <= (1U << 31));
capacity = std::max<uint32_t>(WTF::roundUpToPowerOfTwo(capacity), 4U);
m_capacity = capacity;
makeAndSetNewBuffer(globalObject, vm);
RETURN_IF_EXCEPTION(scope, void());
setUpHeadAndTail(globalObject, vm);
HashMapBucketType* bucket = base->m_head.get()->next();
while (bucket) {
if (!bucket->deleted()) {
addNormalizedNonExistingForCloning(globalObject, bucket->key(), HashMapBucketType::extractValue(*bucket));
RETURN_IF_EXCEPTION(scope, void());
}
bucket = bucket->next();
}
checkConsistency();
}
template <typename HashMapBucketType>
ALWAYS_INLINE HashMapBucketType** HashMapImpl<HashMapBucketType>::findBucket(JSGlobalObject* globalObject, JSValue key)
{
VM& vm = getVM(globalObject);
auto scope = DECLARE_THROW_SCOPE(vm);
key = normalizeMapKey(key);
uint32_t hash = jsMapHash(globalObject, vm, key);
RETURN_IF_EXCEPTION(scope, nullptr);
return findBucket(globalObject, key, hash);
}
template <typename HashMapBucketType>
ALWAYS_INLINE HashMapBucketType** HashMapImpl<HashMapBucketType>::findBucket(JSGlobalObject* globalObject, JSValue key, uint32_t hash)
{
ASSERT_WITH_MESSAGE(normalizeMapKey(key) == key, "We expect normalized values flowing into this function.");
return findBucketAlreadyHashedAndNormalized(globalObject, key, hash);
}
template <typename HashMapBucketType>
template <typename T>
ALWAYS_INLINE typename std::enable_if<std::is_same<T, HashMapBucket<HashMapBucketDataKeyValue>>::value, JSValue>::type HashMapImpl<HashMapBucketType>::get(JSGlobalObject* globalObject, JSValue key)
{
if (HashMapBucketType** bucket = findBucket(globalObject, key))
return (*bucket)->value();
return jsUndefined();
}
template <typename HashMapBucketType>
ALWAYS_INLINE bool HashMapImpl<HashMapBucketType>::has(JSGlobalObject* globalObject, JSValue key)
{
return !!findBucket(globalObject, key);
}
template <typename HashMapBucketType>
ALWAYS_INLINE void HashMapImpl<HashMapBucketType>::add(JSGlobalObject* globalObject, JSValue key, JSValue value)
{
VM& vm = getVM(globalObject);
auto scope = DECLARE_THROW_SCOPE(vm);
key = normalizeMapKey(key);
addNormalizedInternal(globalObject, key, value, [&] (HashMapBucketType* bucket) {
return !isDeleted(bucket) && areKeysEqual(globalObject, key, bucket->key());
});
RETURN_IF_EXCEPTION(scope, void());
scope.release();
if (shouldRehashAfterAdd())
rehash(globalObject);
}
template <typename HashMapBucketType>
ALWAYS_INLINE HashMapBucketType* HashMapImpl<HashMapBucketType>::addNormalized(JSGlobalObject* globalObject, JSValue key, JSValue value, uint32_t hash)
{
VM& vm = getVM(globalObject);
ASSERT_WITH_MESSAGE(normalizeMapKey(key) == key, "We expect normalized values flowing into this function.");
DEFER_TERMINATION_AND_ASSERT_WITH_MESSAGE(vm, jsMapHash(globalObject, getVM(globalObject), key) == hash, "We expect hash value is what we expect.");
auto* bucket = addNormalizedInternal(vm, key, value, hash, [&] (HashMapBucketType* bucket) {
return !isDeleted(bucket) && areKeysEqual(globalObject, key, bucket->key());
});
if (shouldRehashAfterAdd())
rehash(globalObject);
return bucket;
}
template <typename HashMapBucketType>
ALWAYS_INLINE bool HashMapImpl<HashMapBucketType>::remove(JSGlobalObject* globalObject, JSValue key)
{
HashMapBucketType** bucket = findBucket(globalObject, key);
if (!bucket)
return false;
VM& vm = getVM(globalObject);
HashMapBucketType* impl = *bucket;
impl->next()->setPrev(vm, impl->prev());
impl->prev()->setNext(vm, impl->next());
impl->makeDeleted(vm);
*bucket = deletedValue();
++m_deleteCount;
ASSERT(m_keyCount > 0);
--m_keyCount;
if (shouldShrink())
rehash(globalObject);
return true;
}
template <typename HashMapBucketType>
ALWAYS_INLINE void HashMapImpl<HashMapBucketType>::clear(JSGlobalObject* globalObject)
{
VM& vm = getVM(globalObject);
m_keyCount = 0;
m_deleteCount = 0;
HashMapBucketType* head = m_head.get();
HashMapBucketType* bucket = m_head->next();
HashMapBucketType* tail = m_tail.get();
while (bucket != tail) {
HashMapBucketType* next = bucket->next();
// We restart each iterator by pointing it to the head of the list.
bucket->setNext(vm, head);
bucket->makeDeleted(vm);
bucket = next;
}
m_head->setNext(vm, m_tail.get());
m_tail->setPrev(vm, m_head.get());
m_capacity = 4;
makeAndSetNewBuffer(globalObject, vm);
checkConsistency();
}
template <typename HashMapBucketType>
ALWAYS_INLINE void HashMapImpl<HashMapBucketType>::setUpHeadAndTail(JSGlobalObject*, VM& vm)
{
m_head.set(vm, this, HashMapBucketType::create(vm));
m_tail.set(vm, this, HashMapBucketType::create(vm));
m_head->setNext(vm, m_tail.get());
m_tail->setPrev(vm, m_head.get());
ASSERT(m_head->deleted());
ASSERT(m_tail->deleted());
}
template <typename HashMapBucketType>
ALWAYS_INLINE void HashMapImpl<HashMapBucketType>::addNormalizedNonExistingForCloning(JSGlobalObject* globalObject, JSValue key, JSValue value)
{
addNormalizedInternal(globalObject, key, value, [&] (HashMapBucketType*) {
return false;
});
}
template <typename HashMapBucketType>
template<typename CanUseBucket>
ALWAYS_INLINE void HashMapImpl<HashMapBucketType>::addNormalizedInternal(JSGlobalObject* globalObject, JSValue key, JSValue value, const CanUseBucket& canUseBucket)
{
VM& vm = getVM(globalObject);
auto scope = DECLARE_THROW_SCOPE(vm);
uint32_t hash = jsMapHash(globalObject, vm, key);
RETURN_IF_EXCEPTION(scope, void());
scope.release();
addNormalizedInternal(vm, key, value, hash, canUseBucket);
}
template <typename HashMapBucketType>
template<typename CanUseBucket>
ALWAYS_INLINE HashMapBucketType* HashMapImpl<HashMapBucketType>::addNormalizedInternal(VM& vm, JSValue key, JSValue value, uint32_t hash, const CanUseBucket& canUseBucket)
{
ASSERT_WITH_MESSAGE(normalizeMapKey(key) == key, "We expect normalized values flowing into this function.");
const uint32_t mask = m_capacity - 1;
uint32_t index = hash & mask;
HashMapBucketType** buffer = this->buffer();
HashMapBucketType* bucket = buffer[index];
while (!isEmpty(bucket)) {
if (canUseBucket(bucket)) {
bucket->setValue(vm, value);
return bucket;
}
index = (index + 1) & mask;
bucket = buffer[index];
}
HashMapBucketType* newEntry = m_tail.get();
buffer[index] = newEntry;
newEntry->setKey(vm, key);
newEntry->setValue(vm, value);
ASSERT(!newEntry->deleted());
HashMapBucketType* newTail = HashMapBucketType::create(vm);
m_tail.set(vm, this, newTail);
newTail->setPrev(vm, newEntry);
ASSERT(newTail->deleted());
newEntry->setNext(vm, newTail);
++m_keyCount;
return newEntry;
}
template <typename HashMapBucketType>
ALWAYS_INLINE HashMapBucketType** HashMapImpl<HashMapBucketType>::findBucketAlreadyHashedAndNormalized(JSGlobalObject* globalObject, JSValue key, uint32_t hash)
{
const uint32_t mask = m_capacity - 1;
uint32_t index = hash & mask;
HashMapBucketType** buffer = this->buffer();
HashMapBucketType* bucket = buffer[index];
while (!isEmpty(bucket)) {
if (!isDeleted(bucket) && areKeysEqual(globalObject, key, bucket->key()))
return buffer + index;
index = (index + 1) & mask;
bucket = buffer[index];
}
return nullptr;
}
template <typename HashMapBucketType>
void HashMapImpl<HashMapBucketType>::rehash(JSGlobalObject* globalObject)
{
VM& vm = getVM(globalObject);
auto scope = DECLARE_THROW_SCOPE(vm);
uint32_t oldCapacity = m_capacity;
m_capacity = nextCapacity(m_capacity, m_keyCount);
if (m_capacity != oldCapacity) {
makeAndSetNewBuffer(globalObject, vm);
RETURN_IF_EXCEPTION(scope, void());
} else {
m_buffer->reset(m_capacity);
assertBufferIsEmpty();
}
HashMapBucketType* iter = m_head->next();
HashMapBucketType* end = m_tail.get();
const uint32_t mask = m_capacity - 1;
RELEASE_ASSERT(!(m_capacity & (m_capacity - 1)));
HashMapBucketType** buffer = this->buffer();
while (iter != end) {
uint32_t index = jsMapHashForAlreadyHashedValue(globalObject, vm, iter->key()) & mask;
EXCEPTION_ASSERT_WITH_MESSAGE(!scope.exception(), "All keys should already be hashed before, so this should not throw because it won't resolve ropes.");
{
HashMapBucketType* bucket = buffer[index];
while (!isEmpty(bucket)) {
index = (index + 1) & mask;
bucket = buffer[index];
}
}
buffer[index] = iter;
iter = iter->next();
}
m_deleteCount = 0;
checkConsistency();
}
template <typename HashMapBucketType>
ALWAYS_INLINE void HashMapImpl<HashMapBucketType>::checkConsistency() const
{
if (ASSERT_ENABLED) {
HashMapBucketType* iter = m_head->next();
HashMapBucketType* end = m_tail.get();
uint32_t size = 0;
while (iter != end) {
++size;
iter = iter->next();
}
ASSERT(size == m_keyCount);
}
}
template <typename HashMapBucketType>
void HashMapImpl<HashMapBucketType>::makeAndSetNewBuffer(JSGlobalObject* globalObject, VM& vm)
{
ASSERT(!(m_capacity & (m_capacity - 1)));
HashMapBufferType* buffer = HashMapBufferType::create(globalObject, vm, this, m_capacity);
if (UNLIKELY(!buffer))
return;
m_buffer.set(vm, this, buffer);
assertBufferIsEmpty();
}
template <typename HashMapBucketType>
ALWAYS_INLINE void HashMapImpl<HashMapBucketType>::assertBufferIsEmpty() const
{
if (ASSERT_ENABLED) {
for (unsigned i = 0; i < m_capacity; i++)
ASSERT(isEmpty(buffer()[i]));
}
}
} // namespace JSC