blob: 1a2b96924cafa42975786976b1aa58712bb07539 [file] [log] [blame]
// Copyright 2018 The Dawn & Tint Authors
//
// Redistribution and use in source and binary forms, with or without
// modification, are permitted provided that the following conditions are met:
//
// 1. Redistributions of source code must retain the above copyright notice, this
// list of conditions and the following disclaimer.
//
// 2. Redistributions in binary form must reproduce the above copyright notice,
// this list of conditions and the following disclaimer in the documentation
// and/or other materials provided with the distribution.
//
// 3. Neither the name of the copyright holder nor the names of its
// contributors may be used to endorse or promote products derived from
// this software without specific prior written permission.
//
// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
// AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
// IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
// DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE
// FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
// DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
// SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER
// CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
// OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
#include "dawn/native/vulkan/ShaderModuleVk.h"
#include <cstdint>
#include <map>
#include <string>
#include <unordered_set>
#include <vector>
#include "absl/container/flat_hash_map.h"
#include "dawn/common/HashUtils.h"
#include "dawn/common/MatchVariant.h"
#include "dawn/native/Adapter.h"
#include "dawn/native/CacheRequest.h"
#include "dawn/native/PhysicalDevice.h"
#include "dawn/native/Serializable.h"
#include "dawn/native/TintUtils.h"
#include "dawn/native/vulkan/BindGroupLayoutVk.h"
#include "dawn/native/vulkan/DeviceVk.h"
#include "dawn/native/vulkan/FencedDeleter.h"
#include "dawn/native/vulkan/PhysicalDeviceVk.h"
#include "dawn/native/vulkan/PipelineLayoutVk.h"
#include "dawn/native/vulkan/UtilsVulkan.h"
#include "dawn/native/vulkan/VulkanError.h"
#include "dawn/platform/DawnPlatform.h"
#include "dawn/platform/metrics/HistogramMacros.h"
#include "dawn/platform/tracing/TraceEvent.h"
#include "partition_alloc/pointers/raw_ptr.h"
#include "tint/tint.h"
#ifdef DAWN_ENABLE_SPIRV_VALIDATION
#include "dawn/native/SpirvValidation.h"
#endif
namespace dawn::native::vulkan {
#define COMPILED_SPIRV_MEMBERS(X) X(std::vector<uint32_t>, spirv)
// Represents the result and metadata for a SPIR-V compilation.
DAWN_SERIALIZABLE(struct, CompiledSpirv, COMPILED_SPIRV_MEMBERS){};
#undef COMPILED_SPIRV_MEMBERS
bool TransformedShaderModuleCacheKey::operator==(
const TransformedShaderModuleCacheKey& other) const {
if (layoutPtr != other.layoutPtr || entryPoint != other.entryPoint ||
constants.size() != other.constants.size()) {
return false;
}
if (!std::equal(constants.begin(), constants.end(), other.constants.begin())) {
return false;
}
if (emitPointSize != other.emitPointSize) {
return false;
}
return true;
}
size_t TransformedShaderModuleCacheKeyHashFunc::operator()(
const TransformedShaderModuleCacheKey& key) const {
size_t hash = 0;
HashCombine(&hash, key.layoutPtr, key.entryPoint, key.emitPointSize);
for (const auto& entry : key.constants) {
HashCombine(&hash, entry.first, entry.second);
}
return hash;
}
class ShaderModule::ConcurrentTransformedShaderModuleCache {
public:
explicit ConcurrentTransformedShaderModuleCache(Device* device) : mDevice(device) {}
~ConcurrentTransformedShaderModuleCache() {
std::lock_guard<std::mutex> lock(mMutex);
for (const auto& [_, moduleAndSpirv] : mTransformedShaderModuleCache) {
mDevice->GetFencedDeleter()->DeleteWhenUnused(moduleAndSpirv.vkModule);
}
}
std::optional<ModuleAndSpirv> Find(const TransformedShaderModuleCacheKey& key) {
std::lock_guard<std::mutex> lock(mMutex);
auto iter = mTransformedShaderModuleCache.find(key);
if (iter != mTransformedShaderModuleCache.end()) {
return iter->second.AsRefs();
}
return {};
}
ModuleAndSpirv AddOrGet(const TransformedShaderModuleCacheKey& key,
VkShaderModule module,
CompiledSpirv compilation,
bool hasInputAttachment) {
DAWN_ASSERT(module != VK_NULL_HANDLE);
std::lock_guard<std::mutex> lock(mMutex);
auto iter = mTransformedShaderModuleCache.find(key);
if (iter == mTransformedShaderModuleCache.end()) {
bool added = false;
std::tie(iter, added) = mTransformedShaderModuleCache.emplace(
key, Entry{module, std::move(compilation.spirv), hasInputAttachment});
DAWN_ASSERT(added);
} else {
// No need to use FencedDeleter since this shader module was just created and does
// not need to wait for queue operations to complete.
// Also, use of fenced deleter here is not thread safe.
mDevice->fn.DestroyShaderModule(mDevice->GetVkDevice(), module, nullptr);
}
return iter->second.AsRefs();
}
private:
struct Entry {
VkShaderModule vkModule;
std::vector<uint32_t> spirv;
bool hasInputAttachment;
ModuleAndSpirv AsRefs() const {
return {
vkModule,
spirv.data(),
spirv.size(),
hasInputAttachment,
};
}
};
raw_ptr<Device> mDevice;
std::mutex mMutex;
absl::flat_hash_map<TransformedShaderModuleCacheKey,
Entry,
TransformedShaderModuleCacheKeyHashFunc>
mTransformedShaderModuleCache;
};
// static
ResultOrError<Ref<ShaderModule>> ShaderModule::Create(
Device* device,
const UnpackedPtr<ShaderModuleDescriptor>& descriptor,
const std::vector<tint::wgsl::Extension>& internalExtensions,
ShaderModuleParseResult* parseResult,
OwnedCompilationMessages* compilationMessages) {
Ref<ShaderModule> module = AcquireRef(new ShaderModule(device, descriptor, internalExtensions));
DAWN_TRY(module->Initialize(parseResult, compilationMessages));
return module;
}
ShaderModule::ShaderModule(Device* device,
const UnpackedPtr<ShaderModuleDescriptor>& descriptor,
std::vector<tint::wgsl::Extension> internalExtensions)
: ShaderModuleBase(device, descriptor, std::move(internalExtensions)),
mTransformedShaderModuleCache(
std::make_unique<ConcurrentTransformedShaderModuleCache>(device)) {}
MaybeError ShaderModule::Initialize(ShaderModuleParseResult* parseResult,
OwnedCompilationMessages* compilationMessages) {
ScopedTintICEHandler scopedICEHandler(GetDevice());
return InitializeBase(parseResult, compilationMessages);
}
void ShaderModule::DestroyImpl() {
ShaderModuleBase::DestroyImpl();
// Remove reference to internal cache to trigger cleanup.
mTransformedShaderModuleCache = nullptr;
}
ShaderModule::~ShaderModule() = default;
#if TINT_BUILD_SPV_WRITER
#define SPIRV_COMPILATION_REQUEST_MEMBERS(X) \
X(SingleShaderStage, stage) \
X(const tint::Program*, inputProgram) \
X(std::optional<tint::ast::transform::SubstituteOverride::Config>, substituteOverrideConfig) \
X(LimitsForCompilationRequest, limits) \
X(CacheKey::UnsafeUnkeyedValue<const AdapterBase*>, adapter) \
X(std::string_view, entryPointName) \
X(tint::spirv::writer::Options, tintOptions) \
X(CacheKey::UnsafeUnkeyedValue<dawn::platform::Platform*>, platform)
DAWN_MAKE_CACHE_REQUEST(SpirvCompilationRequest, SPIRV_COMPILATION_REQUEST_MEMBERS);
#undef SPIRV_COMPILATION_REQUEST_MEMBERS
#endif // TINT_BUILD_SPV_WRITER
ResultOrError<ShaderModule::ModuleAndSpirv> ShaderModule::GetHandleAndSpirv(
SingleShaderStage stage,
const ProgrammableStage& programmableStage,
const PipelineLayout* layout,
bool clampFragDepth,
bool emitPointSize) {
TRACE_EVENT0(GetDevice()->GetPlatform(), General, "ShaderModuleVk::GetHandleAndSpirv");
ScopedTintICEHandler scopedICEHandler(GetDevice());
// Check to see if we have the handle and spirv cached already
// TODO(chromium:345359083): Improve the computation of the cache key. For example, it isn't
// ideal to use `reinterpret_cast<uintptr_t>(layout)` as the layout may be freed and
// reallocated during the runtime.
auto cacheKey = TransformedShaderModuleCacheKey{reinterpret_cast<uintptr_t>(layout),
programmableStage.entryPoint.c_str(),
programmableStage.constants, emitPointSize};
auto handleAndSpirv = mTransformedShaderModuleCache->Find(cacheKey);
if (handleAndSpirv.has_value()) {
return std::move(*handleAndSpirv);
}
#if TINT_BUILD_SPV_WRITER
// Creation of module and spirv is deferred to this point when using tint generator
tint::spirv::writer::Bindings bindings;
std::unordered_set<tint::BindingPoint> statically_paired_texture_binding_points;
const BindingInfoArray& moduleBindingInfo =
GetEntryPoint(programmableStage.entryPoint.c_str()).bindings;
for (BindGroupIndex group : IterateBitSet(layout->GetBindGroupLayoutsMask())) {
const BindGroupLayout* bgl = ToBackend(layout->GetBindGroupLayout(group));
for (const auto& currentModuleBindingInfo : moduleBindingInfo[group]) {
// We cannot use structured binding here because lambda expressions can only capture
// variables, while structured binding doesn't introduce variables.
const auto& binding = currentModuleBindingInfo.first;
const auto& shaderBindingInfo = currentModuleBindingInfo.second;
tint::BindingPoint srcBindingPoint{static_cast<uint32_t>(group),
static_cast<uint32_t>(binding)};
tint::BindingPoint dstBindingPoint{
static_cast<uint32_t>(group), static_cast<uint32_t>(bgl->GetBindingIndex(binding))};
MatchVariant(
shaderBindingInfo.bindingInfo,
[&](const BufferBindingInfo& bindingInfo) {
switch (bindingInfo.type) {
case wgpu::BufferBindingType::Uniform:
bindings.uniform.emplace(
srcBindingPoint,
tint::spirv::writer::binding::Uniform{dstBindingPoint.group,
dstBindingPoint.binding});
break;
case kInternalStorageBufferBinding:
case wgpu::BufferBindingType::Storage:
case wgpu::BufferBindingType::ReadOnlyStorage:
bindings.storage.emplace(
srcBindingPoint,
tint::spirv::writer::binding::Storage{dstBindingPoint.group,
dstBindingPoint.binding});
break;
case wgpu::BufferBindingType::BindingNotUsed:
case wgpu::BufferBindingType::Undefined:
DAWN_UNREACHABLE();
break;
}
},
[&](const SamplerBindingInfo& bindingInfo) {
bindings.sampler.emplace(srcBindingPoint,
tint::spirv::writer::binding::Sampler{
dstBindingPoint.group, dstBindingPoint.binding});
},
[&](const TextureBindingInfo& bindingInfo) {
if (auto samplerIndex = bgl->GetStaticSamplerIndexForTexture(
BindingIndex{dstBindingPoint.binding})) {
dstBindingPoint.binding = static_cast<uint32_t>(samplerIndex.value());
statically_paired_texture_binding_points.insert(srcBindingPoint);
}
bindings.texture.emplace(srcBindingPoint,
tint::spirv::writer::binding::Texture{
dstBindingPoint.group, dstBindingPoint.binding});
},
[&](const StorageTextureBindingInfo& bindingInfo) {
bindings.storage_texture.emplace(
srcBindingPoint, tint::spirv::writer::binding::StorageTexture{
dstBindingPoint.group, dstBindingPoint.binding});
},
[&](const ExternalTextureBindingInfo& bindingInfo) {
const auto& bindingMap = bgl->GetExternalTextureBindingExpansionMap();
const auto& expansion = bindingMap.find(binding);
DAWN_ASSERT(expansion != bindingMap.end());
const auto& bindingExpansion = expansion->second;
tint::spirv::writer::binding::BindingInfo plane0{
static_cast<uint32_t>(group),
static_cast<uint32_t>(bgl->GetBindingIndex(bindingExpansion.plane0))};
tint::spirv::writer::binding::BindingInfo plane1{
static_cast<uint32_t>(group),
static_cast<uint32_t>(bgl->GetBindingIndex(bindingExpansion.plane1))};
tint::spirv::writer::binding::BindingInfo metadata{
static_cast<uint32_t>(group),
static_cast<uint32_t>(bgl->GetBindingIndex(bindingExpansion.params))};
bindings.external_texture.emplace(
srcBindingPoint,
tint::spirv::writer::binding::ExternalTexture{metadata, plane0, plane1});
},
[&](const InputAttachmentBindingInfo& bindingInfo) {
bindings.input_attachment.emplace(
srcBindingPoint, tint::spirv::writer::binding::InputAttachment{
dstBindingPoint.group, dstBindingPoint.binding});
});
}
}
const bool hasInputAttachment = !bindings.input_attachment.empty();
std::optional<tint::ast::transform::SubstituteOverride::Config> substituteOverrideConfig;
if (!programmableStage.metadata->overrides.empty()) {
substituteOverrideConfig = BuildSubstituteOverridesTransformConfig(programmableStage);
}
SpirvCompilationRequest req = {};
req.stage = stage;
auto tintProgram = GetTintProgram();
req.inputProgram = &(tintProgram->program);
req.entryPointName = programmableStage.entryPoint;
req.platform = UnsafeUnkeyedValue(GetDevice()->GetPlatform());
req.substituteOverrideConfig = std::move(substituteOverrideConfig);
req.tintOptions.remapped_entry_point_name = kRemappedEntryPointName;
req.tintOptions.strip_all_names = !GetDevice()->IsToggleEnabled(Toggle::DisableSymbolRenaming);
req.tintOptions.statically_paired_texture_binding_points =
std::move(statically_paired_texture_binding_points);
req.tintOptions.clamp_frag_depth = clampFragDepth;
req.tintOptions.disable_robustness = !GetDevice()->IsRobustnessEnabled();
req.tintOptions.emit_vertex_point_size = emitPointSize;
req.tintOptions.disable_workgroup_init =
GetDevice()->IsToggleEnabled(Toggle::DisableWorkgroupInit);
// The only possible alternative for the vulkan demote to helper extension is
// "OpTerminateInvocation" which remains unimplemented in dawn/tint.
req.tintOptions.use_demote_to_helper_invocation_extensions =
GetDevice()->IsToggleEnabled(Toggle::VulkanUseDemoteToHelperInvocationExtension);
req.tintOptions.use_zero_initialize_workgroup_memory_extension =
GetDevice()->IsToggleEnabled(Toggle::VulkanUseZeroInitializeWorkgroupMemoryExtension);
req.tintOptions.use_storage_input_output_16 =
GetDevice()->IsToggleEnabled(Toggle::VulkanUseStorageInputOutput16);
req.tintOptions.bindings = std::move(bindings);
req.tintOptions.disable_image_robustness =
GetDevice()->IsToggleEnabled(Toggle::VulkanUseImageRobustAccess2);
// Currently we can disable index clamping on all runtime-sized arrays in Tint robustness
// transform as unsized arrays can only be declared on storage address space.
req.tintOptions.disable_runtime_sized_array_index_clamping =
GetDevice()->IsToggleEnabled(Toggle::VulkanUseBufferRobustAccess2);
req.tintOptions.polyfill_dot_4x8_packed =
GetDevice()->IsToggleEnabled(Toggle::PolyFillPacked4x8DotProduct);
req.tintOptions.disable_polyfill_integer_div_mod =
GetDevice()->IsToggleEnabled(Toggle::DisablePolyfillsOnIntegerDivisonAndModulo);
// Pass matrices to user functions by pointer on Qualcomm devices to workaround a known bug.
// See crbug.com/tint/2045.
if (ToBackend(GetDevice()->GetPhysicalDevice())->IsAndroidQualcomm()) {
req.tintOptions.pass_matrix_by_pointer = true;
}
const CombinedLimits& limits = GetDevice()->GetLimits();
req.limits = LimitsForCompilationRequest::Create(limits.v1);
req.adapter = UnsafeUnkeyedValue(static_cast<const AdapterBase*>(GetDevice()->GetAdapter()));
CacheResult<CompiledSpirv> compilation;
DAWN_TRY_LOAD_OR_RUN(
compilation, GetDevice(), std::move(req), CompiledSpirv::FromBlob,
[](SpirvCompilationRequest r) -> ResultOrError<CompiledSpirv> {
tint::ast::transform::Manager transformManager;
tint::ast::transform::DataMap transformInputs;
// Many Vulkan drivers can't handle multi-entrypoint shader modules.
transformManager.append(std::make_unique<tint::ast::transform::SingleEntryPoint>());
transformInputs.Add<tint::ast::transform::SingleEntryPoint::Config>(
std::string(r.entryPointName));
if (r.substituteOverrideConfig) {
// This needs to run after SingleEntryPoint transform which removes unused overrides
// for current entry point.
transformManager.Add<tint::ast::transform::SubstituteOverride>();
transformInputs.Add<tint::ast::transform::SubstituteOverride::Config>(
std::move(r.substituteOverrideConfig).value());
}
tint::Program program;
tint::ast::transform::DataMap transformOutputs;
{
TRACE_EVENT0(r.platform.UnsafeGetValue(), General, "RunTransforms");
DAWN_TRY_ASSIGN(program,
RunTransforms(&transformManager, r.inputProgram, transformInputs,
&transformOutputs, nullptr));
}
TRACE_EVENT0(r.platform.UnsafeGetValue(), General, "tint::spirv::writer::Generate()");
// Convert the AST program to an IR module.
auto ir = tint::wgsl::reader::ProgramToLoweredIR(program);
DAWN_INVALID_IF(ir != tint::Success, "An error occurred while generating Tint IR\n%s",
ir.Failure().reason.Str());
// Generate SPIR-V from Tint IR.
auto tintResult = tint::spirv::writer::Generate(ir.Get(), r.tintOptions);
DAWN_INVALID_IF(tintResult != tint::Success,
"An error occurred while generating SPIR-V\n%s",
tintResult.Failure().reason.Str());
// Workgroup validation has to come after `Generate` because it may require overrides to
// have been substituted.
if (r.stage == SingleShaderStage::Compute) {
Extent3D _;
DAWN_TRY_ASSIGN(
_, ValidateComputeStageWorkgroupSize(
tintResult->workgroup_info.x, tintResult->workgroup_info.y,
tintResult->workgroup_info.z, tintResult->workgroup_info.storage_size,
r.limits, r.adapter.UnsafeGetValue()));
}
CompiledSpirv result;
result.spirv = std::move(tintResult.Get().spirv);
return result;
},
"Vulkan.CompileShaderToSPIRV");
#ifdef DAWN_ENABLE_SPIRV_VALIDATION
DAWN_TRY(ValidateSpirv(GetDevice(), compilation->spirv.data(), compilation->spirv.size(),
GetDevice()->IsToggleEnabled(Toggle::DumpShaders)));
#endif
VkShaderModuleCreateInfo createInfo;
createInfo.sType = VK_STRUCTURE_TYPE_SHADER_MODULE_CREATE_INFO;
createInfo.pNext = nullptr;
createInfo.flags = 0;
createInfo.codeSize = compilation->spirv.size() * sizeof(uint32_t);
createInfo.pCode = compilation->spirv.data();
Device* device = ToBackend(GetDevice());
VkShaderModule newHandle = VK_NULL_HANDLE;
{
TRACE_EVENT0(GetDevice()->GetPlatform(), General, "vkCreateShaderModule");
DAWN_TRY(CheckVkSuccess(
device->fn.CreateShaderModule(device->GetVkDevice(), &createInfo, nullptr, &*newHandle),
"CreateShaderModule"));
}
ModuleAndSpirv moduleAndSpirv;
if (newHandle != VK_NULL_HANDLE) {
device->GetBlobCache()->EnsureStored(compilation);
// Set the label on `newHandle` now, and not on `moduleAndSpirv.module` later
// since `moduleAndSpirv.module` may be in use by multiple threads.
SetDebugName(ToBackend(GetDevice()), newHandle, "Dawn_ShaderModule", GetLabel());
moduleAndSpirv = mTransformedShaderModuleCache->AddOrGet(
cacheKey, newHandle, compilation.Acquire(), hasInputAttachment);
}
return std::move(moduleAndSpirv);
#else
return DAWN_INTERNAL_ERROR("TINT_BUILD_SPV_WRITER is not defined.");
#endif
}
} // namespace dawn::native::vulkan