blob: c23de6505e07d706b13d2a71a0ac3b42887db869 [file] [log] [blame]
//
// Copyright 2019 The ANGLE Project Authors. All rights reserved.
// Use of this source code is governed by a BSD-style license that can be
// found in the LICENSE file.
//
// mtl_buffer_pool.mm:
// Implements the class methods for BufferPool.
//
#include "libANGLE/renderer/metal/mtl_buffer_pool.h"
#include "libANGLE/renderer/metal/ContextMtl.h"
#include "libANGLE/renderer/metal/DisplayMtl.h"
namespace rx
{
namespace mtl
{
// BufferPool implementation.
BufferPool::BufferPool() : BufferPool(false) {}
BufferPool::BufferPool(bool alwaysAllocNewBuffer)
: BufferPool(alwaysAllocNewBuffer, BufferPoolMemPolicy::Auto)
{}
BufferPool::BufferPool(bool alwaysAllocNewBuffer, BufferPoolMemPolicy policy)
: mInitialSize(0),
mBuffer(nullptr),
mNextAllocationOffset(0),
mLastFlushOffset(0),
mSize(0),
mAlignment(1),
mBuffersAllocated(0),
mMaxBuffers(0),
mMemPolicy(policy),
mAlwaysAllocateNewBuffer(alwaysAllocNewBuffer)
{}
angle::Result BufferPool::reset(ContextMtl *contextMtl,
size_t initialSize,
size_t alignment,
size_t maxBuffers)
{
ANGLE_TRY(finalizePendingBuffer(contextMtl));
releaseInFlightBuffers(contextMtl);
mSize = 0;
if (mBufferFreeList.size() && mInitialSize <= mBufferFreeList.front()->size())
{
// Instead of deleteing old buffers, we should reset them to avoid excessive
// memory re-allocations
if (maxBuffers && mBufferFreeList.size() > maxBuffers)
{
mBufferFreeList.resize(maxBuffers);
mBuffersAllocated = maxBuffers;
}
mSize = mBufferFreeList.front()->size();
for (size_t i = 0; i < mBufferFreeList.size(); ++i)
{
BufferRef &buffer = mBufferFreeList[i];
if (!buffer->isBeingUsedByGPU(contextMtl))
{
// If buffer is not used by GPU, re-use it immediately.
continue;
}
bool useSharedMem = shouldAllocateInSharedMem(contextMtl);
if (IsError(buffer->resetWithSharedMemOpt(contextMtl, useSharedMem, mSize, nullptr)))
{
mBufferFreeList.clear();
mBuffersAllocated = 0;
mSize = 0;
break;
}
}
}
else
{
mBufferFreeList.clear();
mBuffersAllocated = 0;
}
mInitialSize = initialSize;
mMaxBuffers = maxBuffers;
updateAlignment(contextMtl, alignment);
return angle::Result::Continue;
}
void BufferPool::initialize(Context *context,
size_t initialSize,
size_t alignment,
size_t maxBuffers)
{
if (mBuffersAllocated)
{
// Invalid call, must call destroy() first.
UNREACHABLE();
}
mInitialSize = initialSize;
mMaxBuffers = maxBuffers;
updateAlignment(context, alignment);
}
BufferPool::~BufferPool() {}
bool BufferPool::shouldAllocateInSharedMem(ContextMtl *contextMtl) const
{
if (ANGLE_UNLIKELY(contextMtl->getDisplay()->getFeatures().forceBufferGPUStorage.enabled))
{
return false;
}
switch (mMemPolicy)
{
case BufferPoolMemPolicy::AlwaysSharedMem:
return true;
case BufferPoolMemPolicy::AlwaysGPUMem:
return false;
default:
return mSize <= kSharedMemBufferMaxBufSizeHint;
}
}
angle::Result BufferPool::allocateNewBuffer(ContextMtl *contextMtl)
{
if (mMaxBuffers > 0 && mBuffersAllocated >= mMaxBuffers)
{
// We reach the max number of buffers allowed.
// Try to deallocate old and smaller size inflight buffers.
releaseInFlightBuffers(contextMtl);
}
if (mMaxBuffers > 0 && mBuffersAllocated >= mMaxBuffers)
{
// If we reach this point, it means there was no buffer deallocated inside
// releaseInFlightBuffers() thus, the number of buffers allocated still exceeds number
// allowed.
ASSERT(!mBufferFreeList.empty());
// Reuse the buffer in free list:
if (mBufferFreeList.front()->isBeingUsedByGPU(contextMtl))
{
contextMtl->flushCommandBuffer(mtl::NoWait);
// Force the GPU to finish its rendering and make the old buffer available.
contextMtl->cmdQueue().ensureResourceReadyForCPU(mBufferFreeList.front());
}
mBuffer = mBufferFreeList.front();
mBufferFreeList.erase(mBufferFreeList.begin());
return angle::Result::Continue;
}
bool useSharedMem = shouldAllocateInSharedMem(contextMtl);
ANGLE_TRY(
Buffer::MakeBufferWithSharedMemOpt(contextMtl, useSharedMem, mSize, nullptr, &mBuffer));
ASSERT(mBuffer);
mBuffersAllocated++;
return angle::Result::Continue;
}
angle::Result BufferPool::allocate(ContextMtl *contextMtl,
size_t sizeInBytes,
uint8_t **ptrOut,
BufferRef *bufferOut,
size_t *offsetOut,
bool *newBufferAllocatedOut)
{
size_t sizeToAllocate = roundUp(sizeInBytes, mAlignment);
angle::base::CheckedNumeric<size_t> checkedNextWriteOffset = mNextAllocationOffset;
checkedNextWriteOffset += sizeToAllocate;
if (!mBuffer || !checkedNextWriteOffset.IsValid() ||
checkedNextWriteOffset.ValueOrDie() >= mSize ||
// If the current buffer has been modified by GPU, do not reuse it:
mBuffer->isCPUReadMemNeedSync() || mAlwaysAllocateNewBuffer)
{
if (mBuffer)
{
ANGLE_TRY(finalizePendingBuffer(contextMtl));
}
if (sizeToAllocate > mSize)
{
mSize = std::max(mInitialSize, sizeToAllocate);
// Clear the free list since the free buffers are now too small.
destroyBufferList(contextMtl, &mBufferFreeList);
}
// The front of the free list should be the oldest. Thus if it is in use the rest of the
// free list should be in use as well.
if (mBufferFreeList.empty() || mBufferFreeList.front()->isBeingUsedByGPU(contextMtl))
{
ANGLE_TRY(allocateNewBuffer(contextMtl));
}
else
{
mBuffer = mBufferFreeList.front();
mBufferFreeList.erase(mBufferFreeList.begin());
}
ASSERT(mBuffer->size() == mSize);
mNextAllocationOffset = 0;
mLastFlushOffset = 0;
if (newBufferAllocatedOut != nullptr)
{
*newBufferAllocatedOut = true;
}
}
else if (newBufferAllocatedOut != nullptr)
{
*newBufferAllocatedOut = false;
}
ASSERT(mBuffer != nullptr);
if (bufferOut != nullptr)
{
*bufferOut = mBuffer;
}
// Optionally map() the buffer if possible
if (ptrOut)
{
// We don't need to synchronize with GPU access, since allocation should return a
// non-overlapped region each time.
*ptrOut = mBuffer->mapWithOpt(contextMtl, /** readOnly */ false, /** noSync */ true) +
mNextAllocationOffset;
}
if (offsetOut)
{
*offsetOut = static_cast<size_t>(mNextAllocationOffset);
}
mNextAllocationOffset += static_cast<uint32_t>(sizeToAllocate);
return angle::Result::Continue;
}
angle::Result BufferPool::commit(ContextMtl *contextMtl, bool flushEntireBuffer)
{
if (mBuffer && mNextAllocationOffset > mLastFlushOffset)
{
if (flushEntireBuffer)
{
mBuffer->flush(contextMtl, 0, mLastFlushOffset);
}
else
{
mBuffer->flush(contextMtl, mLastFlushOffset, mNextAllocationOffset - mLastFlushOffset);
}
mLastFlushOffset = mNextAllocationOffset;
}
return angle::Result::Continue;
}
angle::Result BufferPool::finalizePendingBuffer(ContextMtl *contextMtl)
{
if (mBuffer)
{
ANGLE_TRY(commit(contextMtl));
// commit() already flushes so no need to flush here.
mBuffer->unmapNoFlush(contextMtl);
mInFlightBuffers.push_back(mBuffer);
mBuffer = nullptr;
}
mNextAllocationOffset = 0;
mLastFlushOffset = 0;
return angle::Result::Continue;
}
void BufferPool::releaseInFlightBuffers(ContextMtl *contextMtl)
{
for (auto &toRelease : mInFlightBuffers)
{
// If the dynamic buffer was resized we cannot reuse the retained buffer.
if (toRelease->size() < mSize
#if TARGET_OS_OSX || TARGET_OS_MACCATALYST
// Also release buffer if it was allocated in different policy
|| toRelease->useSharedMem() != shouldAllocateInSharedMem(contextMtl)
#endif
)
{
toRelease = nullptr;
mBuffersAllocated--;
}
else
{
mBufferFreeList.push_back(toRelease);
}
}
mInFlightBuffers.clear();
}
void BufferPool::destroyBufferList(ContextMtl *contextMtl, std::deque<BufferRef> *buffers)
{
ASSERT(mBuffersAllocated >= buffers->size());
mBuffersAllocated -= buffers->size();
buffers->clear();
}
void BufferPool::destroy(ContextMtl *contextMtl)
{
destroyBufferList(contextMtl, &mInFlightBuffers);
destroyBufferList(contextMtl, &mBufferFreeList);
reset();
if (mBuffer)
{
mBuffer->unmap(contextMtl);
mBuffer = nullptr;
}
}
void BufferPool::updateAlignment(Context *context, size_t alignment)
{
ASSERT(alignment > 0);
// NOTE(hqle): May check additional platform limits.
// If alignment has changed, make sure the next allocation is done at an aligned offset.
if (alignment != mAlignment)
{
mNextAllocationOffset = roundUp(mNextAllocationOffset, static_cast<uint32_t>(alignment));
mAlignment = alignment;
}
}
void BufferPool::reset()
{
mSize = 0;
mNextAllocationOffset = 0;
mLastFlushOffset = 0;
mMaxBuffers = 0;
mAlwaysAllocateNewBuffer = false;
mBuffersAllocated = 0;
}
}
}