blob: c0e4ff657c7deb263390deca5399770a89d64a01 [file] [log] [blame]
/*
* Copyright (C) 2016-2021 Apple Inc. All Rights Reserved.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
* 1. Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
* 2. Redistributions in binary form must reproduce the above copyright
* notice, this list of conditions and the following disclaimer in the
* documentation and/or other materials provided with the distribution.
*
* THIS SOFTWARE IS PROVIDED BY APPLE INC. ``AS IS'' AND ANY
* EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
* PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL APPLE INC. OR
* CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL,
* EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO,
* PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR
* PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY
* OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
*/
#pragma once
#include <wtf/Assertions.h>
#include <wtf/FastMalloc.h>
#include <wtf/HashFunctions.h>
#include <wtf/Noncopyable.h>
namespace WTF {
DECLARE_ALLOCATOR_WITH_HEAP_IDENTIFIER(SmallSet);
// Functionally, this class is very similar to std::variant<Vector<T, SmallArraySize>, HashSet<T>>
// It is optimized primarily for space, but is also quite fast
// Its main limitation is that it has no way to remove elements once they have been added to it
// Also, instead of being fully parameterized by a HashTrait parameter, it always uses -1 (all ones) as its empty value
// Relatedly, it can only store objects of up to 64 bit size (but that particular limitation should be fairly easy to lift if needed)
// Use it whenever you need to store an unbounded but probably small number of unsigned integers or pointers.
template<typename T, typename Hash = PtrHashBase<T, false /* isSmartPtr */>, unsigned SmallArraySize = 8>
class SmallSet {
WTF_MAKE_FAST_ALLOCATED;
WTF_MAKE_NONCOPYABLE(SmallSet);
static_assert(std::is_trivially_destructible<T>::value, "We currently don't support non-trivially destructible types.");
static_assert(!(SmallArraySize & (SmallArraySize - 1)), "Inline size must be a power of two.");
static_assert(sizeof(T*) <= SmallArraySize * sizeof(T), "This class has not been tested for m_inline.buffer larger than m_inline.smallStorage");
public:
SmallSet()
: m_inline()
{
initialize();
}
// We take care to have SmallSet have partial move semantics allowable through
// memcpy. It's partial move semantics because our destructor should not be called
// on the SmallPtrObject in the old memory we were moved from (otherwise, we might free m_buffer twice)
// unless that old memory is reset to be isSmall(). See move constructor below.
// To maintain these semantics, we determine if we're small by checking our size
// and not our m_buffer pointer. And when we're small, we don't do operations on
// m_buffer, instead, we perform operations on m_smallStorage directly. The reason we want
// these semantics is that it's beneficial to have a Vector that contains SmallSet
// (or an object with SmallSet as a field) be allowed to use memcpy for its move operation.
SmallSet(SmallSet&& other)
{
memcpy(static_cast<void*>(this), static_cast<void*>(&other), sizeof(SmallSet));
other.initialize();
}
SmallSet& operator=(SmallSet&& other)
{
this->~SmallSet();
new (this) SmallSet(WTFMove(other));
return *this;
}
~SmallSet()
{
if (!isSmall())
SmallSetMalloc::free(m_inline.buffer);
}
// We could easily include an iterator in this to fully match the HashSet interface, but currently none of our clients require it.
struct AddResult {
bool isNewEntry;
};
inline AddResult add(T value)
{
ASSERT(isValidEntry(value));
if (isSmall()) {
for (unsigned i = 0; i < m_size; i++) {
if (m_inline.smallStorage[i] == value)
return { false };
}
if (m_size < SmallArraySize) {
m_inline.smallStorage[m_size] = value;
++m_size;
return { true };
}
grow(std::max(64u, SmallArraySize * 2));
// Fall through. We're no longer small :(
}
// If we're more than 3/4ths full we grow.
if (UNLIKELY(m_size * 4 >= m_capacity * 3)) {
grow(m_capacity * 2);
ASSERT(!(m_capacity & (m_capacity - 1)));
}
T* bucket = this->bucket(value);
if (*bucket != value) {
*bucket = value;
++m_size;
return { true };
}
return { false };
}
inline bool contains(T value) const
{
ASSERT(isValidEntry(value));
if (isSmall()) {
// We only need to search up to m_size because we store things linearly inside m_smallStorage.
for (unsigned i = 0; i < m_size; i++) {
if (m_inline.smallStorage[i] == value)
return true;
}
return false;
}
T* bucket = this->bucket(value);
return *bucket == value;
}
class iterator {
WTF_MAKE_FAST_ALLOCATED;
public:
iterator& operator++()
{
m_index++;
ASSERT(m_index <= m_capacity);
while (m_index < m_capacity && m_buffer[m_index] == emptyValue())
m_index++;
return *this;
}
T operator*() const { ASSERT(m_index < m_capacity); return static_cast<T>(m_buffer[m_index]); }
bool operator==(const iterator& other) const { ASSERT(m_buffer == other.m_buffer); return m_index == other.m_index; }
bool operator!=(const iterator& other) const { ASSERT(m_buffer == other.m_buffer); return !(*this == other); }
private:
template<typename U, typename H, unsigned S> friend class WTF::SmallSet;
unsigned m_index;
unsigned m_capacity;
T* m_buffer;
};
iterator begin() const
{
iterator it;
it.m_index = std::numeric_limits<unsigned>::max();
it.m_capacity = m_capacity;
if (isSmall())
it.m_buffer = const_cast<T*>(m_inline.smallStorage);
else
it.m_buffer = m_inline.buffer;
++it;
return it;
}
iterator end() const
{
iterator it;
it.m_index = m_capacity;
it.m_capacity = m_capacity;
if (isSmall())
it.m_buffer = const_cast<T*>(m_inline.smallStorage);
else
it.m_buffer = m_inline.buffer;
return it;
}
inline unsigned size() const { return m_size; }
inline bool isEmpty() const { return !size(); }
unsigned memoryUse() const
{
unsigned memory = sizeof(SmallSet);
if (!isSmall())
memory += m_capacity * sizeof(T);
return memory;
}
private:
constexpr static T emptyValue()
{
if constexpr (std::is_pointer<T>::value)
return static_cast<T>(bitwise_cast<void*>(std::numeric_limits<uintptr_t>::max()));
return std::numeric_limits<T>::max();
}
bool isValidEntry(const T value) const
{
return value != emptyValue();
}
inline bool isSmall() const
{
return m_capacity == SmallArraySize;
}
inline void initialize()
{
m_size = 0;
m_capacity = SmallArraySize;
memset(static_cast<void*>(m_inline.smallStorage), -1, sizeof(T) * SmallArraySize);
ASSERT(isSmall());
}
inline void grow(unsigned size)
{
// We memset the new buffer with -1, so for consistency emptyValue() must return something which is all 1s.
#if !defined(NDEBUG)
if constexpr (std::is_pointer<T>::value)
ASSERT(bitwise_cast<intptr_t>(emptyValue()) == -1ll);
else if constexpr (sizeof(T) == 8)
ASSERT(bitwise_cast<int64_t>(emptyValue()) == -1ll);
else if constexpr (sizeof(T) == 4)
ASSERT(bitwise_cast<int32_t>(emptyValue()) == -1);
else if constexpr (sizeof(T) == 2)
ASSERT(bitwise_cast<int16_t>(emptyValue()) == -1);
else if constexpr (sizeof(T) == 1)
ASSERT(bitwise_cast<int8_t>(emptyValue()) == -1);
else
RELEASE_ASSERT_NOT_REACHED();
#endif
size_t allocationSize = sizeof(T) * size;
bool wasSmall = isSmall();
T* oldBuffer = wasSmall ? m_inline.smallStorage : m_inline.buffer;
unsigned oldCapacity = m_capacity;
T* newBuffer = static_cast<T*>(SmallSetMalloc::malloc(allocationSize));
memset(static_cast<void*>(newBuffer), -1, allocationSize);
m_capacity = size;
for (unsigned i = 0; i < oldCapacity; i++) {
if (oldBuffer[i] != emptyValue()) {
T* ptr = bucketInBuffer(newBuffer, static_cast<T>(oldBuffer[i]));
*ptr = oldBuffer[i];
}
}
if (!wasSmall)
SmallSetMalloc::free(oldBuffer);
m_inline.buffer = newBuffer;
}
inline T* bucket(T target) const
{
ASSERT(!isSmall());
return bucketInBuffer(m_inline.buffer, target);
}
inline T* bucketInBuffer(T* buffer, T target) const
{
ASSERT(!(m_capacity & (m_capacity - 1)));
unsigned bucket = Hash::hash(target) & (m_capacity - 1);
unsigned index = 0;
while (true) {
T* ptr = buffer + bucket;
if (*ptr == emptyValue())
return ptr;
if (*ptr == target)
return ptr;
index++;
bucket = (bucket + index) & (m_capacity - 1);
}
}
unsigned m_size;
unsigned m_capacity;
union U {
T* buffer;
T smallStorage[SmallArraySize];
U() { };
} m_inline;
};
} // namespace WTF
using WTF::SmallSet;