blob: 42bb0c4d4b91c6e745afa9a803ffeee4dba9fec4 [file] [log] [blame]
// Copyright 2020 The Dawn & Tint Authors
//
// Redistribution and use in source and binary forms, with or without
// modification, are permitted provided that the following conditions are met:
//
// 1. Redistributions of source code must retain the above copyright notice, this
// list of conditions and the following disclaimer.
//
// 2. Redistributions in binary form must reproduce the above copyright notice,
// this list of conditions and the following disclaimer in the documentation
// and/or other materials provided with the distribution.
//
// 3. Neither the name of the copyright holder nor the names of its
// contributors may be used to endorse or promote products derived from
// this software without specific prior written permission.
//
// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
// AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
// IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
// DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE
// FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
// DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
// SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER
// CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
// OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
#include "dawn/native/vulkan/DescriptorSetAllocator.h"
#include <utility>
#include "dawn/native/vulkan/BindGroupLayoutVk.h"
#include "dawn/native/vulkan/DeviceVk.h"
#include "dawn/native/vulkan/FencedDeleter.h"
#include "dawn/native/vulkan/VulkanError.h"
namespace dawn::native::vulkan {
// TODO(enga): Figure out this value.
static constexpr uint32_t kMaxDescriptorsPerPool = 512;
// static
Ref<DescriptorSetAllocator> DescriptorSetAllocator::Create(
BindGroupLayout* layout,
std::map<VkDescriptorType, uint32_t> descriptorCountPerType) {
return AcquireRef(new DescriptorSetAllocator(layout, descriptorCountPerType));
}
DescriptorSetAllocator::DescriptorSetAllocator(
BindGroupLayout* layout,
std::map<VkDescriptorType, uint32_t> descriptorCountPerType)
: ObjectBase(layout->GetDevice()), mLayout(layout) {
DAWN_ASSERT(layout != nullptr);
// Compute the total number of descriptors for this layout.
uint32_t totalDescriptorCount = 0;
mPoolSizes.reserve(descriptorCountPerType.size());
for (const auto& [type, count] : descriptorCountPerType) {
DAWN_ASSERT(count > 0);
totalDescriptorCount += count;
mPoolSizes.push_back(VkDescriptorPoolSize{type, count});
}
if (totalDescriptorCount == 0) {
// Vulkan requires that valid usage of vkCreateDescriptorPool must have a non-zero
// number of pools, each of which has non-zero descriptor counts.
// Since the descriptor set layout is empty, we should be able to allocate
// |kMaxDescriptorsPerPool| sets from this 1-sized descriptor pool.
// The type of this descriptor pool doesn't matter because it is never used.
mPoolSizes.push_back(VkDescriptorPoolSize{VK_DESCRIPTOR_TYPE_UNIFORM_BUFFER, 1});
mMaxSets = kMaxDescriptorsPerPool;
} else {
DAWN_ASSERT(totalDescriptorCount <= kMaxBindingsPerPipelineLayout);
static_assert(kMaxBindingsPerPipelineLayout <= kMaxDescriptorsPerPool);
// Compute the total number of descriptors sets that fits given the max.
mMaxSets = kMaxDescriptorsPerPool / totalDescriptorCount;
DAWN_ASSERT(mMaxSets > 0);
// Grow the number of desciptors in the pool to fit the computed |mMaxSets|.
for (auto& poolSize : mPoolSizes) {
poolSize.descriptorCount *= mMaxSets;
}
}
}
DescriptorSetAllocator::~DescriptorSetAllocator() {
for (auto& pool : mDescriptorPools) {
DAWN_ASSERT(pool.freeSetIndices.size() == mMaxSets);
if (pool.vkPool != VK_NULL_HANDLE) {
Device* device = ToBackend(GetDevice());
device->GetFencedDeleter()->DeleteWhenUnused(pool.vkPool);
}
}
}
ResultOrError<DescriptorSetAllocation> DescriptorSetAllocator::Allocate() {
if (mAvailableDescriptorPoolIndices.empty()) {
DAWN_TRY(AllocateDescriptorPool());
}
DAWN_ASSERT(!mAvailableDescriptorPoolIndices.empty());
const PoolIndex poolIndex = mAvailableDescriptorPoolIndices.back();
DescriptorPool* pool = &mDescriptorPools[poolIndex];
DAWN_ASSERT(!pool->freeSetIndices.empty());
SetIndex setIndex = pool->freeSetIndices.back();
pool->freeSetIndices.pop_back();
if (pool->freeSetIndices.empty()) {
mAvailableDescriptorPoolIndices.pop_back();
}
return DescriptorSetAllocation{pool->sets[setIndex], poolIndex, setIndex};
}
void DescriptorSetAllocator::Deallocate(DescriptorSetAllocation* allocationInfo) {
DAWN_ASSERT(allocationInfo != nullptr);
DAWN_ASSERT(allocationInfo->set != VK_NULL_HANDLE);
// We can't reuse the descriptor set right away because the Vulkan spec says in the
// documentation for vkCmdBindDescriptorSets that the set may be consumed any time between
// host execution of the command and the end of the draw/dispatch.
Device* device = ToBackend(GetDevice());
const ExecutionSerial serial = device->GetPendingCommandSerial();
mPendingDeallocations.Enqueue({allocationInfo->poolIndex, allocationInfo->setIndex}, serial);
if (mLastDeallocationSerial != serial) {
device->EnqueueDeferredDeallocation(this);
mLastDeallocationSerial = serial;
}
// Clear the content of allocation so that use after frees are more visible.
*allocationInfo = {};
}
void DescriptorSetAllocator::FinishDeallocation(ExecutionSerial completedSerial) {
for (const Deallocation& dealloc : mPendingDeallocations.IterateUpTo(completedSerial)) {
DAWN_ASSERT(dealloc.poolIndex < mDescriptorPools.size());
auto& freeSetIndices = mDescriptorPools[dealloc.poolIndex].freeSetIndices;
if (freeSetIndices.empty()) {
mAvailableDescriptorPoolIndices.emplace_back(dealloc.poolIndex);
}
freeSetIndices.emplace_back(dealloc.setIndex);
}
mPendingDeallocations.ClearUpTo(completedSerial);
}
MaybeError DescriptorSetAllocator::AllocateDescriptorPool() {
VkDescriptorPoolCreateInfo createInfo;
createInfo.sType = VK_STRUCTURE_TYPE_DESCRIPTOR_POOL_CREATE_INFO;
createInfo.pNext = nullptr;
createInfo.flags = 0;
createInfo.maxSets = mMaxSets;
createInfo.poolSizeCount = static_cast<PoolIndex>(mPoolSizes.size());
createInfo.pPoolSizes = mPoolSizes.data();
Device* device = ToBackend(GetDevice());
VkDescriptorPool descriptorPool;
DAWN_TRY(CheckVkSuccess(device->fn.CreateDescriptorPool(device->GetVkDevice(), &createInfo,
nullptr, &*descriptorPool),
"CreateDescriptorPool"));
std::vector<VkDescriptorSetLayout> layouts(mMaxSets, mLayout->GetHandle());
VkDescriptorSetAllocateInfo allocateInfo;
allocateInfo.sType = VK_STRUCTURE_TYPE_DESCRIPTOR_SET_ALLOCATE_INFO;
allocateInfo.pNext = nullptr;
allocateInfo.descriptorPool = descriptorPool;
allocateInfo.descriptorSetCount = mMaxSets;
allocateInfo.pSetLayouts = AsVkArray(layouts.data());
std::vector<VkDescriptorSet> sets(mMaxSets);
MaybeError result =
CheckVkSuccess(device->fn.AllocateDescriptorSets(device->GetVkDevice(), &allocateInfo,
AsVkArray(sets.data())),
"AllocateDescriptorSets");
if (result.IsError()) {
// On an error we can destroy the pool immediately because no command references it.
device->fn.DestroyDescriptorPool(device->GetVkDevice(), descriptorPool, nullptr);
DAWN_TRY(std::move(result));
}
std::vector<SetIndex> freeSetIndices;
freeSetIndices.reserve(mMaxSets);
for (SetIndex i = 0; i < mMaxSets; ++i) {
freeSetIndices.push_back(i);
}
mAvailableDescriptorPoolIndices.push_back(mDescriptorPools.size());
mDescriptorPools.emplace_back(
DescriptorPool{descriptorPool, std::move(sets), std::move(freeSetIndices)});
return {};
}
} // namespace dawn::native::vulkan