| // Copyright 2023 The Dawn & Tint Authors |
| // |
| // Redistribution and use in source and binary forms, with or without |
| // modification, are permitted provided that the following conditions are met: |
| // |
| // 1. Redistributions of source code must retain the above copyright notice, this |
| // list of conditions and the following disclaimer. |
| // |
| // 2. Redistributions in binary form must reproduce the above copyright notice, |
| // this list of conditions and the following disclaimer in the documentation |
| // and/or other materials provided with the distribution. |
| // |
| // 3. Neither the name of the copyright holder nor the names of its |
| // contributors may be used to endorse or promote products derived from |
| // this software without specific prior written permission. |
| // |
| // THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" |
| // AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE |
| // IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE |
| // DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE |
| // FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL |
| // DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR |
| // SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER |
| // CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, |
| // OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE |
| // OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
| |
| #include "src/tint/lang/spirv/writer/raise/builtin_polyfill.h" |
| |
| #include <utility> |
| |
| #include "spirv/unified1/spirv.h" |
| #include "src/tint/lang/core/fluent_types.h" |
| #include "src/tint/lang/core/ir/builder.h" |
| #include "src/tint/lang/core/ir/module.h" |
| #include "src/tint/lang/core/ir/validator.h" |
| #include "src/tint/lang/core/type/builtin_structs.h" |
| #include "src/tint/lang/core/type/depth_multisampled_texture.h" |
| #include "src/tint/lang/core/type/depth_texture.h" |
| #include "src/tint/lang/core/type/multisampled_texture.h" |
| #include "src/tint/lang/core/type/sampled_texture.h" |
| #include "src/tint/lang/core/type/storage_texture.h" |
| #include "src/tint/lang/core/type/texture.h" |
| #include "src/tint/lang/spirv/ir/builtin_call.h" |
| #include "src/tint/lang/spirv/ir/literal_operand.h" |
| #include "src/tint/lang/spirv/type/sampled_image.h" |
| #include "src/tint/utils/ice/ice.h" |
| |
| using namespace tint::core::number_suffixes; // NOLINT |
| using namespace tint::core::fluent_types; // NOLINT |
| |
| namespace tint::spirv::writer::raise { |
| |
| namespace { |
| |
| /// PIMPL state for the transform. |
| struct State { |
| /// The IR module. |
| core::ir::Module& ir; |
| |
| /// The IR builder. |
| core::ir::Builder b{ir}; |
| |
| /// The type manager. |
| core::type::Manager& ty{ir.Types()}; |
| |
| /// Process the module. |
| void Process() { |
| // Find the builtins that need replacing. |
| Vector<core::ir::CoreBuiltinCall*, 4> worklist; |
| for (auto* inst : ir.Instructions()) { |
| if (auto* builtin = inst->As<core::ir::CoreBuiltinCall>()) { |
| switch (builtin->Func()) { |
| case core::BuiltinFn::kArrayLength: |
| case core::BuiltinFn::kAtomicAdd: |
| case core::BuiltinFn::kAtomicAnd: |
| case core::BuiltinFn::kAtomicCompareExchangeWeak: |
| case core::BuiltinFn::kAtomicExchange: |
| case core::BuiltinFn::kAtomicLoad: |
| case core::BuiltinFn::kAtomicMax: |
| case core::BuiltinFn::kAtomicMin: |
| case core::BuiltinFn::kAtomicOr: |
| case core::BuiltinFn::kAtomicStore: |
| case core::BuiltinFn::kAtomicSub: |
| case core::BuiltinFn::kAtomicXor: |
| case core::BuiltinFn::kDot: |
| case core::BuiltinFn::kDot4I8Packed: |
| case core::BuiltinFn::kDot4U8Packed: |
| case core::BuiltinFn::kSelect: |
| case core::BuiltinFn::kTextureDimensions: |
| case core::BuiltinFn::kTextureGather: |
| case core::BuiltinFn::kTextureGatherCompare: |
| case core::BuiltinFn::kTextureLoad: |
| case core::BuiltinFn::kTextureNumLayers: |
| case core::BuiltinFn::kTextureSample: |
| case core::BuiltinFn::kTextureSampleBias: |
| case core::BuiltinFn::kTextureSampleCompare: |
| case core::BuiltinFn::kTextureSampleCompareLevel: |
| case core::BuiltinFn::kTextureSampleGrad: |
| case core::BuiltinFn::kTextureSampleLevel: |
| case core::BuiltinFn::kTextureStore: |
| worklist.Push(builtin); |
| break; |
| case core::BuiltinFn::kQuantizeToF16: |
| if (builtin->Result(0)->Type()->Is<core::type::Vector>()) { |
| worklist.Push(builtin); |
| } |
| break; |
| default: |
| break; |
| } |
| } |
| } |
| |
| // Replace the builtins that we found. |
| for (auto* builtin : worklist) { |
| core::ir::Value* replacement = nullptr; |
| switch (builtin->Func()) { |
| case core::BuiltinFn::kArrayLength: |
| replacement = ArrayLength(builtin); |
| break; |
| case core::BuiltinFn::kAtomicAdd: |
| case core::BuiltinFn::kAtomicAnd: |
| case core::BuiltinFn::kAtomicCompareExchangeWeak: |
| case core::BuiltinFn::kAtomicExchange: |
| case core::BuiltinFn::kAtomicLoad: |
| case core::BuiltinFn::kAtomicMax: |
| case core::BuiltinFn::kAtomicMin: |
| case core::BuiltinFn::kAtomicOr: |
| case core::BuiltinFn::kAtomicStore: |
| case core::BuiltinFn::kAtomicSub: |
| case core::BuiltinFn::kAtomicXor: |
| replacement = Atomic(builtin); |
| break; |
| case core::BuiltinFn::kDot: |
| replacement = Dot(builtin); |
| break; |
| case core::BuiltinFn::kDot4I8Packed: |
| case core::BuiltinFn::kDot4U8Packed: |
| replacement = DotPacked4x8(builtin); |
| break; |
| case core::BuiltinFn::kSelect: |
| replacement = Select(builtin); |
| break; |
| case core::BuiltinFn::kTextureDimensions: |
| replacement = TextureDimensions(builtin); |
| break; |
| case core::BuiltinFn::kTextureGather: |
| case core::BuiltinFn::kTextureGatherCompare: |
| replacement = TextureGather(builtin); |
| break; |
| case core::BuiltinFn::kTextureLoad: |
| replacement = TextureLoad(builtin); |
| break; |
| case core::BuiltinFn::kTextureNumLayers: |
| replacement = TextureNumLayers(builtin); |
| break; |
| case core::BuiltinFn::kTextureSample: |
| case core::BuiltinFn::kTextureSampleBias: |
| case core::BuiltinFn::kTextureSampleCompare: |
| case core::BuiltinFn::kTextureSampleCompareLevel: |
| case core::BuiltinFn::kTextureSampleGrad: |
| case core::BuiltinFn::kTextureSampleLevel: |
| replacement = TextureSample(builtin); |
| break; |
| case core::BuiltinFn::kTextureStore: |
| replacement = TextureStore(builtin); |
| break; |
| case core::BuiltinFn::kQuantizeToF16: |
| replacement = QuantizeToF16Vec(builtin); |
| break; |
| default: |
| break; |
| } |
| TINT_ASSERT_OR_RETURN(replacement); |
| |
| // Replace the old builtin result with the new value. |
| if (auto name = ir.NameOf(builtin->Result(0))) { |
| ir.SetName(replacement, name); |
| } |
| builtin->Result(0)->ReplaceAllUsesWith(replacement); |
| builtin->Destroy(); |
| } |
| } |
| |
| /// Create a literal operand. |
| /// @param value the literal value |
| /// @returns the literal operand |
| spirv::ir::LiteralOperand* Literal(u32 value) { |
| return ir.allocators.values.Create<spirv::ir::LiteralOperand>(b.ConstantValue(value)); |
| } |
| |
| /// Handle an `arrayLength()` builtin. |
| /// @param builtin the builtin call instruction |
| /// @returns the replacement value |
| core::ir::Value* ArrayLength(core::ir::CoreBuiltinCall* builtin) { |
| // Strip away any let instructions to get to the original struct member access instruction. |
| auto* ptr = builtin->Args()[0]->As<core::ir::InstructionResult>(); |
| while (auto* let = tint::As<core::ir::Let>(ptr->Instruction())) { |
| ptr = let->Value()->As<core::ir::InstructionResult>(); |
| } |
| TINT_ASSERT_OR_RETURN_VALUE(ptr, nullptr); |
| |
| auto* access = ptr->Instruction()->As<core::ir::Access>(); |
| TINT_ASSERT_OR_RETURN_VALUE(access, nullptr); |
| TINT_ASSERT_OR_RETURN_VALUE(access->Indices().Length() == 1u, nullptr); |
| TINT_ASSERT_OR_RETURN_VALUE(access->Object()->Type()->UnwrapPtr()->Is<core::type::Struct>(), |
| nullptr); |
| auto* const_idx = access->Indices()[0]->As<core::ir::Constant>(); |
| |
| // Replace the builtin call with a call to the spirv.array_length intrinsic. |
| auto* call = b.Call<spirv::ir::BuiltinCall>( |
| builtin->Result(0)->Type(), spirv::BuiltinFn::kArrayLength, |
| Vector{access->Object(), Literal(u32(const_idx->Value()->ValueAs<uint32_t>()))}); |
| call->InsertBefore(builtin); |
| return call->Result(0); |
| } |
| |
| /// Handle an atomic*() builtin. |
| /// @param builtin the builtin call instruction |
| /// @returns the replacement value |
| core::ir::Value* Atomic(core::ir::CoreBuiltinCall* builtin) { |
| auto* result_ty = builtin->Result(0)->Type(); |
| |
| auto* pointer = builtin->Args()[0]; |
| auto* memory = [&]() -> core::ir::Value* { |
| switch (pointer->Type()->As<core::type::Pointer>()->AddressSpace()) { |
| case core::AddressSpace::kWorkgroup: |
| return b.Constant(u32(SpvScopeWorkgroup)); |
| case core::AddressSpace::kStorage: |
| return b.Constant(u32(SpvScopeDevice)); |
| default: |
| TINT_UNREACHABLE() << "unhandled atomic address space"; |
| return nullptr; |
| } |
| }(); |
| auto* memory_semantics = b.Constant(u32(SpvMemorySemanticsMaskNone)); |
| |
| // Helper to build the builtin call with the common operands. |
| auto build = [&](const core::type::Type* type, enum spirv::BuiltinFn builtin_fn) { |
| return b.Call<spirv::ir::BuiltinCall>(type, builtin_fn, pointer, memory, |
| memory_semantics); |
| }; |
| |
| // Create the replacement call instruction. |
| core::ir::Call* call = nullptr; |
| switch (builtin->Func()) { |
| case core::BuiltinFn::kAtomicAdd: |
| call = build(result_ty, spirv::BuiltinFn::kAtomicIadd); |
| call->AppendArg(builtin->Args()[1]); |
| break; |
| case core::BuiltinFn::kAtomicAnd: |
| call = build(result_ty, spirv::BuiltinFn::kAtomicAnd); |
| call->AppendArg(builtin->Args()[1]); |
| break; |
| case core::BuiltinFn::kAtomicCompareExchangeWeak: { |
| auto* cmp = builtin->Args()[1]; |
| auto* value = builtin->Args()[2]; |
| auto* int_ty = value->Type(); |
| call = build(int_ty, spirv::BuiltinFn::kAtomicCompareExchange); |
| call->AppendArg(memory_semantics); |
| call->AppendArg(value); |
| call->AppendArg(cmp); |
| call->InsertBefore(builtin); |
| |
| // Compare the original value to the comparator to see if an exchange happened. |
| auto* original = call->Result(0); |
| auto* compare = b.Equal(ty.bool_(), original, cmp); |
| compare->InsertBefore(builtin); |
| |
| // Construct the atomicCompareExchange result structure. |
| call = b.Construct( |
| core::type::CreateAtomicCompareExchangeResult(ty, ir.symbols, int_ty), |
| Vector{original, compare->Result(0)}); |
| break; |
| } |
| case core::BuiltinFn::kAtomicExchange: |
| call = build(result_ty, spirv::BuiltinFn::kAtomicExchange); |
| call->AppendArg(builtin->Args()[1]); |
| break; |
| case core::BuiltinFn::kAtomicLoad: |
| call = build(result_ty, spirv::BuiltinFn::kAtomicLoad); |
| break; |
| case core::BuiltinFn::kAtomicOr: |
| call = build(result_ty, spirv::BuiltinFn::kAtomicOr); |
| call->AppendArg(builtin->Args()[1]); |
| break; |
| case core::BuiltinFn::kAtomicMax: |
| if (result_ty->is_signed_integer_scalar()) { |
| call = build(result_ty, spirv::BuiltinFn::kAtomicSmax); |
| } else { |
| call = build(result_ty, spirv::BuiltinFn::kAtomicUmax); |
| } |
| call->AppendArg(builtin->Args()[1]); |
| break; |
| case core::BuiltinFn::kAtomicMin: |
| if (result_ty->is_signed_integer_scalar()) { |
| call = build(result_ty, spirv::BuiltinFn::kAtomicSmin); |
| } else { |
| call = build(result_ty, spirv::BuiltinFn::kAtomicUmin); |
| } |
| call->AppendArg(builtin->Args()[1]); |
| break; |
| case core::BuiltinFn::kAtomicStore: |
| call = build(result_ty, spirv::BuiltinFn::kAtomicStore); |
| call->AppendArg(builtin->Args()[1]); |
| break; |
| case core::BuiltinFn::kAtomicSub: |
| call = build(result_ty, spirv::BuiltinFn::kAtomicIsub); |
| call->AppendArg(builtin->Args()[1]); |
| break; |
| case core::BuiltinFn::kAtomicXor: |
| call = build(result_ty, spirv::BuiltinFn::kAtomicXor); |
| call->AppendArg(builtin->Args()[1]); |
| break; |
| default: |
| return nullptr; |
| } |
| |
| call->InsertBefore(builtin); |
| return call->Result(0); |
| } |
| |
| /// Handle a `dot()` builtin. |
| /// @param builtin the builtin call instruction |
| /// @returns the replacement value |
| core::ir::Value* Dot(core::ir::CoreBuiltinCall* builtin) { |
| // OpDot only supports floating point operands, so we need to polyfill the integer case. |
| // TODO(crbug.com/tint/1267): If SPV_KHR_integer_dot_product is supported, use that instead. |
| if (builtin->Result(0)->Type()->is_integer_scalar()) { |
| core::ir::Instruction* sum = nullptr; |
| |
| auto* v1 = builtin->Args()[0]; |
| auto* v2 = builtin->Args()[1]; |
| auto* vec = v1->Type()->As<core::type::Vector>(); |
| auto* elty = vec->type(); |
| for (uint32_t i = 0; i < vec->Width(); i++) { |
| b.InsertBefore(builtin, [&] { |
| auto* e1 = b.Access(elty, v1, u32(i)); |
| auto* e2 = b.Access(elty, v2, u32(i)); |
| auto* mul = b.Multiply(elty, e1, e2); |
| if (sum) { |
| sum = b.Add(elty, sum, mul); |
| } else { |
| sum = mul; |
| } |
| }); |
| } |
| return sum->Result(0); |
| } |
| |
| // Replace the builtin call with a call to the spirv.dot intrinsic. |
| auto args = Vector<core::ir::Value*, 4>(builtin->Args()); |
| auto* call = b.Call<spirv::ir::BuiltinCall>(builtin->Result(0)->Type(), |
| spirv::BuiltinFn::kDot, std::move(args)); |
| call->InsertBefore(builtin); |
| return call->Result(0); |
| } |
| |
| /// Handle a `dot4{I,U}8Packed()` builtin. |
| /// @param builtin the builtin call instruction |
| /// @returns the replacement value |
| core::ir::Value* DotPacked4x8(core::ir::CoreBuiltinCall* builtin) { |
| // Replace the builtin call with a call to the spirv.{s,u}dot intrinsic. |
| auto* type = builtin->Result(0)->Type(); |
| auto is_signed = builtin->Func() == core::BuiltinFn::kDot4I8Packed; |
| auto inst = is_signed ? spirv::BuiltinFn::kSdot : spirv::BuiltinFn::kUdot; |
| |
| auto args = Vector<core::ir::Value*, 3>(builtin->Args()); |
| args.Push(Literal(u32(SpvPackedVectorFormatPackedVectorFormat4x8Bit))); |
| |
| auto* call = b.Call<spirv::ir::BuiltinCall>(type, inst, std::move(args)); |
| call->InsertBefore(builtin); |
| return call->Result(0); |
| } |
| |
| /// Handle a `select()` builtin. |
| /// @param builtin the builtin call instruction |
| /// @returns the replacement value |
| core::ir::Value* Select(core::ir::CoreBuiltinCall* builtin) { |
| // Argument order is different in SPIR-V: (condition, true_operand, false_operand). |
| Vector<core::ir::Value*, 4> args = { |
| builtin->Args()[2], |
| builtin->Args()[1], |
| builtin->Args()[0], |
| }; |
| |
| // If the condition is scalar and the objects are vectors, we need to splat the condition |
| // into a vector of the same size. |
| // TODO(jrprice): We don't need to do this if we're targeting SPIR-V 1.4 or newer. |
| auto* vec = builtin->Result(0)->Type()->As<core::type::Vector>(); |
| if (vec && args[0]->Type()->Is<core::type::Scalar>()) { |
| Vector<core::ir::Value*, 4> elements; |
| elements.Resize(vec->Width(), args[0]); |
| |
| auto* construct = b.Construct(ty.vec(ty.bool_(), vec->Width()), std::move(elements)); |
| construct->InsertBefore(builtin); |
| args[0] = construct->Result(0); |
| } |
| |
| // Replace the builtin call with a call to the spirv.select intrinsic. |
| auto* call = b.Call<spirv::ir::BuiltinCall>(builtin->Result(0)->Type(), |
| spirv::BuiltinFn::kSelect, std::move(args)); |
| call->InsertBefore(builtin); |
| return call->Result(0); |
| } |
| |
| /// ImageOperands represents the optional image operands for an image instruction. |
| struct ImageOperands { |
| /// Bias |
| core::ir::Value* bias = nullptr; |
| /// Lod |
| core::ir::Value* lod = nullptr; |
| /// Grad (dx) |
| core::ir::Value* ddx = nullptr; |
| /// Grad (dy) |
| core::ir::Value* ddy = nullptr; |
| /// ConstOffset |
| core::ir::Value* offset = nullptr; |
| /// Sample |
| core::ir::Value* sample = nullptr; |
| }; |
| |
| /// Append optional image operands to an image intrinsic argument list. |
| /// @param operands the operands |
| /// @param args the argument list |
| /// @param insertion_point the insertion point for new instructions |
| /// @param requires_float_lod true if the lod needs to be a floating point value |
| void AppendImageOperands(ImageOperands& operands, |
| Vector<core::ir::Value*, 8>& args, |
| core::ir::Instruction* insertion_point, |
| bool requires_float_lod) { |
| // Add a placeholder argument for the image operand mask, which we will fill in when we have |
| // processed the image operands. |
| uint32_t image_operand_mask = 0u; |
| size_t mask_idx = args.Length(); |
| args.Push(nullptr); |
| |
| // Add each of the optional image operands if used, updating the image operand mask. |
| if (operands.bias) { |
| image_operand_mask |= SpvImageOperandsBiasMask; |
| args.Push(operands.bias); |
| } |
| if (operands.lod) { |
| image_operand_mask |= SpvImageOperandsLodMask; |
| if (requires_float_lod && operands.lod->Type()->is_integer_scalar()) { |
| auto* convert = b.Convert(ty.f32(), operands.lod); |
| convert->InsertBefore(insertion_point); |
| operands.lod = convert->Result(0); |
| } |
| args.Push(operands.lod); |
| } |
| if (operands.ddx) { |
| image_operand_mask |= SpvImageOperandsGradMask; |
| args.Push(operands.ddx); |
| args.Push(operands.ddy); |
| } |
| if (operands.offset) { |
| image_operand_mask |= SpvImageOperandsConstOffsetMask; |
| args.Push(operands.offset); |
| } |
| if (operands.sample) { |
| image_operand_mask |= SpvImageOperandsSampleMask; |
| args.Push(operands.sample); |
| } |
| |
| // Replace the image operand mask with the final mask value, as a literal operand. |
| args[mask_idx] = Literal(u32(image_operand_mask)); |
| } |
| |
| /// Append an array index to a coordinate vector. |
| /// @param coords the coordinate vector |
| /// @param array_idx the array index |
| /// @param insertion_point the insertion point for new instructions |
| /// @returns the modified coordinate vector |
| core::ir::Value* AppendArrayIndex(core::ir::Value* coords, |
| core::ir::Value* array_idx, |
| core::ir::Instruction* insertion_point) { |
| auto* vec = coords->Type()->As<core::type::Vector>(); |
| auto* element_ty = vec->type(); |
| |
| // Convert the index to match the coordinate type if needed. |
| if (array_idx->Type() != element_ty) { |
| auto* array_idx_converted = b.Convert(element_ty, array_idx); |
| array_idx_converted->InsertBefore(insertion_point); |
| array_idx = array_idx_converted->Result(0); |
| } |
| |
| // Construct a new coordinate vector. |
| auto num_coords = vec->Width(); |
| auto* coord_ty = ty.vec(element_ty, num_coords + 1); |
| auto* construct = b.Construct(coord_ty, Vector{coords, array_idx}); |
| construct->InsertBefore(insertion_point); |
| return construct->Result(0); |
| } |
| |
| /// Handle a textureSample*() builtin. |
| /// @param builtin the builtin call instruction |
| /// @returns the replacement value |
| core::ir::Value* TextureSample(core::ir::CoreBuiltinCall* builtin) { |
| // Helper to get the next argument from the call, or nullptr if there are no more arguments. |
| uint32_t arg_idx = 0; |
| auto next_arg = [&]() { |
| return arg_idx < builtin->Args().Length() ? builtin->Args()[arg_idx++] : nullptr; |
| }; |
| |
| auto* texture = next_arg(); |
| auto* sampler = next_arg(); |
| auto* coords = next_arg(); |
| auto* texture_ty = texture->Type()->As<core::type::Texture>(); |
| |
| // Use OpSampledImage to create an OpTypeSampledImage object. |
| auto* sampled_image = b.Call<spirv::ir::BuiltinCall>(ty.Get<type::SampledImage>(texture_ty), |
| spirv::BuiltinFn::kSampledImage, |
| Vector{texture, sampler}); |
| sampled_image->InsertBefore(builtin); |
| |
| // Append the array index to the coordinates if provided. |
| auto* array_idx = IsTextureArray(texture_ty->dim()) ? next_arg() : nullptr; |
| if (array_idx) { |
| coords = AppendArrayIndex(coords, array_idx, builtin); |
| } |
| |
| // Determine which SPIR-V function to use and which optional image operands are needed. |
| enum spirv::BuiltinFn function; |
| core::ir::Value* depth = nullptr; |
| ImageOperands operands; |
| switch (builtin->Func()) { |
| case core::BuiltinFn::kTextureSample: |
| function = spirv::BuiltinFn::kImageSampleImplicitLod; |
| operands.offset = next_arg(); |
| break; |
| case core::BuiltinFn::kTextureSampleBias: |
| function = spirv::BuiltinFn::kImageSampleImplicitLod; |
| operands.bias = next_arg(); |
| operands.offset = next_arg(); |
| break; |
| case core::BuiltinFn::kTextureSampleCompare: |
| function = spirv::BuiltinFn::kImageSampleDrefImplicitLod; |
| depth = next_arg(); |
| operands.offset = next_arg(); |
| break; |
| case core::BuiltinFn::kTextureSampleCompareLevel: |
| function = spirv::BuiltinFn::kImageSampleDrefExplicitLod; |
| depth = next_arg(); |
| operands.lod = b.Constant(0_f); |
| operands.offset = next_arg(); |
| break; |
| case core::BuiltinFn::kTextureSampleGrad: |
| function = spirv::BuiltinFn::kImageSampleExplicitLod; |
| operands.ddx = next_arg(); |
| operands.ddy = next_arg(); |
| operands.offset = next_arg(); |
| break; |
| case core::BuiltinFn::kTextureSampleLevel: |
| function = spirv::BuiltinFn::kImageSampleExplicitLod; |
| operands.lod = next_arg(); |
| operands.offset = next_arg(); |
| break; |
| default: |
| return nullptr; |
| } |
| |
| // Start building the argument list for the function. |
| // The first two operands are always the sampled image and then the coordinates, followed by |
| // the depth reference if used. |
| Vector<core::ir::Value*, 8> function_args; |
| function_args.Push(sampled_image->Result(0)); |
| function_args.Push(coords); |
| if (depth) { |
| function_args.Push(depth); |
| } |
| |
| // Add the optional image operands, if any. |
| AppendImageOperands(operands, function_args, builtin, /* requires_float_lod */ true); |
| |
| // Call the function. |
| // If this is a depth comparison, the result is always f32, otherwise vec4f. |
| auto* result_ty = depth ? static_cast<const core::type::Type*>(ty.f32()) : ty.vec4<f32>(); |
| auto* texture_call = |
| b.Call<spirv::ir::BuiltinCall>(result_ty, function, std::move(function_args)); |
| texture_call->InsertBefore(builtin); |
| |
| auto* result = texture_call->Result(0); |
| |
| // If this is not a depth comparison but we are sampling a depth texture, extract the first |
| // component to get the scalar f32 that SPIR-V expects. |
| if (!depth && |
| texture_ty->IsAnyOf<core::type::DepthTexture, core::type::DepthMultisampledTexture>()) { |
| auto* extract = b.Access(ty.f32(), result, 0_u); |
| extract->InsertBefore(builtin); |
| result = extract->Result(0); |
| } |
| |
| return result; |
| } |
| |
| /// Handle a textureGather*() builtin. |
| /// @param builtin the builtin call instruction |
| /// @returns the replacement value |
| core::ir::Value* TextureGather(core::ir::CoreBuiltinCall* builtin) { |
| // Helper to get the next argument from the call, or nullptr if there are no more arguments. |
| uint32_t arg_idx = 0; |
| auto next_arg = [&]() { |
| return arg_idx < builtin->Args().Length() ? builtin->Args()[arg_idx++] : nullptr; |
| }; |
| |
| auto* component = next_arg(); |
| if (!component->Type()->is_integer_scalar()) { |
| // The first argument wasn't the component, so it must be the texture instead. |
| // Use constant zero for the component. |
| component = b.Constant(0_u); |
| arg_idx--; |
| } |
| auto* texture = next_arg(); |
| auto* sampler = next_arg(); |
| auto* coords = next_arg(); |
| auto* texture_ty = texture->Type()->As<core::type::Texture>(); |
| |
| // Use OpSampledImage to create an OpTypeSampledImage object. |
| auto* sampled_image = b.Call<spirv::ir::BuiltinCall>(ty.Get<type::SampledImage>(texture_ty), |
| spirv::BuiltinFn::kSampledImage, |
| Vector{texture, sampler}); |
| sampled_image->InsertBefore(builtin); |
| |
| // Append the array index to the coordinates if provided. |
| auto* array_idx = IsTextureArray(texture_ty->dim()) ? next_arg() : nullptr; |
| if (array_idx) { |
| coords = AppendArrayIndex(coords, array_idx, builtin); |
| } |
| |
| // Determine which SPIR-V function to use and which optional image operands are needed. |
| enum spirv::BuiltinFn function; |
| core::ir::Value* depth = nullptr; |
| ImageOperands operands; |
| switch (builtin->Func()) { |
| case core::BuiltinFn::kTextureGather: |
| function = spirv::BuiltinFn::kImageGather; |
| operands.offset = next_arg(); |
| break; |
| case core::BuiltinFn::kTextureGatherCompare: |
| function = spirv::BuiltinFn::kImageDrefGather; |
| depth = next_arg(); |
| operands.offset = next_arg(); |
| break; |
| default: |
| return nullptr; |
| } |
| |
| // Start building the argument list for the function. |
| // The first two operands are always the sampled image and then the coordinates, followed by |
| // either the depth reference or the component. |
| Vector<core::ir::Value*, 8> function_args; |
| function_args.Push(sampled_image->Result(0)); |
| function_args.Push(coords); |
| if (depth) { |
| function_args.Push(depth); |
| } else { |
| function_args.Push(component); |
| } |
| |
| // Add the optional image operands, if any. |
| AppendImageOperands(operands, function_args, builtin, /* requires_float_lod */ true); |
| |
| // Call the function. |
| auto* result_ty = builtin->Result(0)->Type(); |
| auto* texture_call = |
| b.Call<spirv::ir::BuiltinCall>(result_ty, function, std::move(function_args)); |
| texture_call->InsertBefore(builtin); |
| return texture_call->Result(0); |
| } |
| |
| /// Handle a textureLoad() builtin. |
| /// @param builtin the builtin call instruction |
| /// @returns the replacement value |
| core::ir::Value* TextureLoad(core::ir::CoreBuiltinCall* builtin) { |
| // Helper to get the next argument from the call, or nullptr if there are no more arguments. |
| uint32_t arg_idx = 0; |
| auto next_arg = [&]() { |
| return arg_idx < builtin->Args().Length() ? builtin->Args()[arg_idx++] : nullptr; |
| }; |
| |
| auto* texture = next_arg(); |
| auto* coords = next_arg(); |
| auto* texture_ty = texture->Type()->As<core::type::Texture>(); |
| |
| // Append the array index to the coordinates if provided. |
| auto* array_idx = IsTextureArray(texture_ty->dim()) ? next_arg() : nullptr; |
| if (array_idx) { |
| coords = AppendArrayIndex(coords, array_idx, builtin); |
| } |
| |
| // Start building the argument list for the builtin. |
| // The first two operands are always the texture and then the coordinates. |
| Vector<core::ir::Value*, 8> builtin_args; |
| builtin_args.Push(texture); |
| builtin_args.Push(coords); |
| |
| // Add the optional image operands, if any. |
| ImageOperands operands; |
| if (texture_ty->IsAnyOf<core::type::MultisampledTexture, |
| core::type::DepthMultisampledTexture>()) { |
| operands.sample = next_arg(); |
| } else { |
| operands.lod = next_arg(); |
| } |
| AppendImageOperands(operands, builtin_args, builtin, /* requires_float_lod */ false); |
| |
| // Call the builtin. |
| // The result is always a vec4 in SPIR-V. |
| auto* result_ty = builtin->Result(0)->Type(); |
| bool expects_scalar_result = result_ty->Is<core::type::Scalar>(); |
| if (expects_scalar_result) { |
| result_ty = ty.vec4(result_ty); |
| } |
| auto kind = texture_ty->Is<core::type::StorageTexture>() ? spirv::BuiltinFn::kImageRead |
| : spirv::BuiltinFn::kImageFetch; |
| auto* texture_call = |
| b.Call<spirv::ir::BuiltinCall>(result_ty, kind, std::move(builtin_args)); |
| texture_call->InsertBefore(builtin); |
| auto* result = texture_call->Result(0); |
| |
| // If we are expecting a scalar result, extract the first component. |
| if (expects_scalar_result) { |
| auto* extract = b.Access(ty.f32(), result, 0_u); |
| extract->InsertBefore(builtin); |
| result = extract->Result(0); |
| } |
| |
| return result; |
| } |
| |
| /// Handle a textureStore() builtin. |
| /// @param builtin the builtin call instruction |
| /// @returns the replacement value |
| core::ir::Value* TextureStore(core::ir::CoreBuiltinCall* builtin) { |
| // Helper to get the next argument from the call, or nullptr if there are no more arguments. |
| uint32_t arg_idx = 0; |
| auto next_arg = [&]() { |
| return arg_idx < builtin->Args().Length() ? builtin->Args()[arg_idx++] : nullptr; |
| }; |
| |
| auto* texture = next_arg(); |
| auto* coords = next_arg(); |
| auto* texture_ty = texture->Type()->As<core::type::Texture>(); |
| |
| // Append the array index to the coordinates if provided. |
| auto* array_idx = IsTextureArray(texture_ty->dim()) ? next_arg() : nullptr; |
| if (array_idx) { |
| coords = AppendArrayIndex(coords, array_idx, builtin); |
| } |
| |
| auto* texel = next_arg(); |
| |
| // Start building the argument list for the function. |
| // The first two operands are always the texture and then the coordinates. |
| Vector<core::ir::Value*, 8> function_args; |
| function_args.Push(texture); |
| function_args.Push(coords); |
| function_args.Push(texel); |
| |
| ImageOperands operands; |
| AppendImageOperands(operands, function_args, builtin, /* requires_float_lod */ false); |
| |
| // Call the function. |
| auto* texture_call = b.Call<spirv::ir::BuiltinCall>( |
| ty.void_(), spirv::BuiltinFn::kImageWrite, std::move(function_args)); |
| texture_call->InsertBefore(builtin); |
| return texture_call->Result(0); |
| } |
| |
| /// Handle a textureDimensions() builtin. |
| /// @param builtin the builtin call instruction |
| /// @returns the replacement value |
| core::ir::Value* TextureDimensions(core::ir::CoreBuiltinCall* builtin) { |
| // Helper to get the next argument from the call, or nullptr if there are no more arguments. |
| uint32_t arg_idx = 0; |
| auto next_arg = [&]() { |
| return arg_idx < builtin->Args().Length() ? builtin->Args()[arg_idx++] : nullptr; |
| }; |
| |
| auto* texture = next_arg(); |
| auto* texture_ty = texture->Type()->As<core::type::Texture>(); |
| |
| Vector<core::ir::Value*, 8> function_args; |
| function_args.Push(texture); |
| |
| // Determine which SPIR-V function to use, and add the Lod argument if needed. |
| enum spirv::BuiltinFn function; |
| if (texture_ty |
| ->IsAnyOf<core::type::MultisampledTexture, core::type::DepthMultisampledTexture, |
| core::type::StorageTexture>()) { |
| function = spirv::BuiltinFn::kImageQuerySize; |
| } else { |
| function = spirv::BuiltinFn::kImageQuerySizeLod; |
| if (auto* lod = next_arg()) { |
| function_args.Push(lod); |
| } else { |
| // Lod wasn't explicit, so assume 0. |
| function_args.Push(b.Constant(0_u)); |
| } |
| } |
| |
| // Add an extra component to the result vector for arrayed textures. |
| auto* result_ty = builtin->Result(0)->Type(); |
| if (core::type::IsTextureArray(texture_ty->dim())) { |
| auto* vec = result_ty->As<core::type::Vector>(); |
| result_ty = ty.vec(vec->type(), vec->Width() + 1); |
| } |
| |
| // Call the function. |
| auto* texture_call = |
| b.Call<spirv::ir::BuiltinCall>(result_ty, function, std::move(function_args)); |
| texture_call->InsertBefore(builtin); |
| |
| auto* result = texture_call->Result(0); |
| |
| // Swizzle the first two components from the result for arrayed textures. |
| if (core::type::IsTextureArray(texture_ty->dim())) { |
| auto* swizzle = b.Swizzle(builtin->Result(0)->Type(), result, {0, 1}); |
| swizzle->InsertBefore(builtin); |
| result = swizzle->Result(0); |
| } |
| |
| return result; |
| } |
| |
| /// Handle a textureNumLayers() builtin. |
| /// @param builtin the builtin call instruction |
| /// @returns the replacement value |
| core::ir::Value* TextureNumLayers(core::ir::CoreBuiltinCall* builtin) { |
| auto* texture = builtin->Args()[0]; |
| auto* texture_ty = texture->Type()->As<core::type::Texture>(); |
| |
| Vector<core::ir::Value*, 2> function_args; |
| function_args.Push(texture); |
| |
| // Determine which SPIR-V function to use, and add the Lod argument if needed. |
| enum spirv::BuiltinFn function; |
| if (texture_ty |
| ->IsAnyOf<core::type::MultisampledTexture, core::type::DepthMultisampledTexture, |
| core::type::StorageTexture>()) { |
| function = spirv::BuiltinFn::kImageQuerySize; |
| } else { |
| function = spirv::BuiltinFn::kImageQuerySizeLod; |
| function_args.Push(b.Constant(0_u)); |
| } |
| |
| // Call the function. |
| auto* texture_call = |
| b.Call<spirv::ir::BuiltinCall>(ty.vec3<u32>(), function, std::move(function_args)); |
| texture_call->InsertBefore(builtin); |
| |
| // Extract the third component to get the number of array layers. |
| auto* extract = b.Access(ty.u32(), texture_call->Result(0), 2_u); |
| extract->InsertBefore(builtin); |
| return extract->Result(0); |
| } |
| |
| /// Scalarize the vector form of a `quantizeToF16()` builtin. |
| /// See crbug.com/tint/1741. |
| /// @param builtin the builtin call instruction |
| /// @returns the replacement value |
| core::ir::Value* QuantizeToF16Vec(core::ir::CoreBuiltinCall* builtin) { |
| auto* arg = builtin->Args()[0]; |
| auto* vec = arg->Type()->As<core::type::Vector>(); |
| TINT_ASSERT(vec); |
| |
| // Replace the builtin call with a call to the spirv.dot intrinsic. |
| Vector<core::ir::Value*, 4> args; |
| for (uint32_t i = 0; i < vec->Width(); i++) { |
| auto* el = b.Access(ty.f32(), arg, u32(i)); |
| auto* scalar_call = b.Call(ty.f32(), core::BuiltinFn::kQuantizeToF16, el); |
| args.Push(scalar_call->Result(0)); |
| el->InsertBefore(builtin); |
| scalar_call->InsertBefore(builtin); |
| } |
| auto* construct = b.Construct(vec, std::move(args)); |
| construct->InsertBefore(builtin); |
| return construct->Result(0); |
| } |
| }; |
| |
| } // namespace |
| |
| Result<SuccessType> BuiltinPolyfill(core::ir::Module& ir) { |
| auto result = ValidateAndDumpIfNeeded(ir, "BuiltinPolyfill transform"); |
| if (result != Success) { |
| return result.Failure(); |
| } |
| |
| State{ir}.Process(); |
| |
| return Success; |
| } |
| |
| } // namespace tint::spirv::writer::raise |