| // Copyright 2019 The Dawn Authors |
| // |
| // Licensed under the Apache License, Version 2.0 (the "License"); |
| // you may not use this file except in compliance with the License. |
| // You may obtain a copy of the License at |
| // |
| // http://www.apache.org/licenses/LICENSE-2.0 |
| // |
| // Unless required by applicable law or agreed to in writing, software |
| // distributed under the License is distributed on an "AS IS" BASIS, |
| // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| // See the License for the specific language governing permissions and |
| // limitations under the License. |
| |
| #include "dawn_native/CommandEncoder.h" |
| |
| #include "common/BitSetIterator.h" |
| #include "common/Math.h" |
| #include "dawn_native/BindGroup.h" |
| #include "dawn_native/Buffer.h" |
| #include "dawn_native/CommandBuffer.h" |
| #include "dawn_native/CommandBufferStateTracker.h" |
| #include "dawn_native/CommandValidation.h" |
| #include "dawn_native/Commands.h" |
| #include "dawn_native/ComputePassEncoder.h" |
| #include "dawn_native/Device.h" |
| #include "dawn_native/ErrorData.h" |
| #include "dawn_native/QueryHelper.h" |
| #include "dawn_native/QuerySet.h" |
| #include "dawn_native/Queue.h" |
| #include "dawn_native/RenderPassEncoder.h" |
| #include "dawn_native/RenderPipeline.h" |
| #include "dawn_native/ValidationUtils_autogen.h" |
| #include "dawn_platform/DawnPlatform.h" |
| #include "dawn_platform/tracing/TraceEvent.h" |
| |
| #include <cmath> |
| #include <map> |
| |
| namespace dawn_native { |
| |
| namespace { |
| |
| MaybeError ValidateB2BCopyAlignment(uint64_t dataSize, |
| uint64_t srcOffset, |
| uint64_t dstOffset) { |
| // Copy size must be a multiple of 4 bytes on macOS. |
| if (dataSize % 4 != 0) { |
| return DAWN_VALIDATION_ERROR("Copy size must be a multiple of 4 bytes"); |
| } |
| |
| // SourceOffset and destinationOffset must be multiples of 4 bytes on macOS. |
| if (srcOffset % 4 != 0 || dstOffset % 4 != 0) { |
| return DAWN_VALIDATION_ERROR( |
| "Source offset and destination offset must be multiples of 4 bytes"); |
| } |
| |
| return {}; |
| } |
| |
| MaybeError ValidateTextureSampleCountInBufferCopyCommands(const TextureBase* texture) { |
| if (texture->GetSampleCount() > 1) { |
| return DAWN_VALIDATION_ERROR( |
| "The sample count of textures must be 1 when copying between buffers and " |
| "textures"); |
| } |
| |
| return {}; |
| } |
| |
| MaybeError ValidateLinearTextureCopyOffset(const TextureDataLayout& layout, |
| const TexelBlockInfo& blockInfo, |
| const bool hasDepthOrStencil) { |
| if (hasDepthOrStencil) { |
| // For depth-stencil texture, buffer offset must be a multiple of 4. |
| if (layout.offset % 4 != 0) { |
| return DAWN_VALIDATION_ERROR( |
| "offset must be a multiple of 4 for depth/stencil texture."); |
| } |
| } else { |
| if (layout.offset % blockInfo.byteSize != 0) { |
| return DAWN_VALIDATION_ERROR( |
| "offset must be a multiple of the texel block byte size."); |
| } |
| } |
| return {}; |
| } |
| |
| MaybeError ValidateTextureDepthStencilToBufferCopyRestrictions( |
| const ImageCopyTexture& src) { |
| Aspect aspectUsed; |
| DAWN_TRY_ASSIGN(aspectUsed, SingleAspectUsedByImageCopyTexture(src)); |
| if (aspectUsed == Aspect::Depth) { |
| switch (src.texture->GetFormat().format) { |
| case wgpu::TextureFormat::Depth24Plus: |
| case wgpu::TextureFormat::Depth24PlusStencil8: |
| return DAWN_VALIDATION_ERROR( |
| "The depth aspect of depth24plus texture cannot be selected in a " |
| "texture to buffer copy"); |
| case wgpu::TextureFormat::Depth32Float: |
| break; |
| |
| default: |
| UNREACHABLE(); |
| } |
| } |
| |
| return {}; |
| } |
| |
| MaybeError ValidateAttachmentArrayLayersAndLevelCount(const TextureViewBase* attachment) { |
| // Currently we do not support layered rendering. |
| if (attachment->GetLayerCount() > 1) { |
| return DAWN_VALIDATION_ERROR( |
| "The layer count of the texture view used as attachment cannot be greater than " |
| "1"); |
| } |
| |
| if (attachment->GetLevelCount() > 1) { |
| return DAWN_VALIDATION_ERROR( |
| "The mipmap level count of the texture view used as attachment cannot be " |
| "greater than 1"); |
| } |
| |
| return {}; |
| } |
| |
| MaybeError ValidateOrSetAttachmentSize(const TextureViewBase* attachment, |
| uint32_t* width, |
| uint32_t* height) { |
| const Extent3D& attachmentSize = |
| attachment->GetTexture()->GetMipLevelVirtualSize(attachment->GetBaseMipLevel()); |
| |
| if (*width == 0) { |
| DAWN_ASSERT(*height == 0); |
| *width = attachmentSize.width; |
| *height = attachmentSize.height; |
| DAWN_ASSERT(*width != 0 && *height != 0); |
| } else if (*width != attachmentSize.width || *height != attachmentSize.height) { |
| return DAWN_VALIDATION_ERROR("Attachment size mismatch"); |
| } |
| |
| return {}; |
| } |
| |
| MaybeError ValidateOrSetColorAttachmentSampleCount(const TextureViewBase* colorAttachment, |
| uint32_t* sampleCount) { |
| if (*sampleCount == 0) { |
| *sampleCount = colorAttachment->GetTexture()->GetSampleCount(); |
| DAWN_ASSERT(*sampleCount != 0); |
| } else if (*sampleCount != colorAttachment->GetTexture()->GetSampleCount()) { |
| return DAWN_VALIDATION_ERROR("Color attachment sample counts mismatch"); |
| } |
| |
| return {}; |
| } |
| |
| MaybeError ValidateResolveTarget(const DeviceBase* device, |
| const RenderPassColorAttachment& colorAttachment) { |
| if (colorAttachment.resolveTarget == nullptr) { |
| return {}; |
| } |
| |
| const TextureViewBase* resolveTarget = colorAttachment.resolveTarget; |
| const TextureViewBase* attachment = colorAttachment.view; |
| DAWN_TRY(device->ValidateObject(colorAttachment.resolveTarget)); |
| DAWN_TRY(ValidateCanUseAs(colorAttachment.resolveTarget->GetTexture(), |
| wgpu::TextureUsage::RenderAttachment)); |
| |
| if (!attachment->GetTexture()->IsMultisampledTexture()) { |
| return DAWN_VALIDATION_ERROR( |
| "Cannot set resolve target when the sample count of the color attachment is 1"); |
| } |
| |
| if (resolveTarget->GetTexture()->IsMultisampledTexture()) { |
| return DAWN_VALIDATION_ERROR("Cannot use multisampled texture as resolve target"); |
| } |
| |
| if (resolveTarget->GetLayerCount() > 1) { |
| return DAWN_VALIDATION_ERROR( |
| "The array layer count of the resolve target must be 1"); |
| } |
| |
| if (resolveTarget->GetLevelCount() > 1) { |
| return DAWN_VALIDATION_ERROR("The mip level count of the resolve target must be 1"); |
| } |
| |
| const Extent3D& colorTextureSize = |
| attachment->GetTexture()->GetMipLevelVirtualSize(attachment->GetBaseMipLevel()); |
| const Extent3D& resolveTextureSize = |
| resolveTarget->GetTexture()->GetMipLevelVirtualSize( |
| resolveTarget->GetBaseMipLevel()); |
| if (colorTextureSize.width != resolveTextureSize.width || |
| colorTextureSize.height != resolveTextureSize.height) { |
| return DAWN_VALIDATION_ERROR( |
| "The size of the resolve target must be the same as the color attachment"); |
| } |
| |
| wgpu::TextureFormat resolveTargetFormat = resolveTarget->GetFormat().format; |
| if (resolveTargetFormat != attachment->GetFormat().format) { |
| return DAWN_VALIDATION_ERROR( |
| "The format of the resolve target must be the same as the color attachment"); |
| } |
| |
| return {}; |
| } |
| |
| MaybeError ValidateRenderPassColorAttachment( |
| DeviceBase* device, |
| const RenderPassColorAttachment& colorAttachment, |
| uint32_t* width, |
| uint32_t* height, |
| uint32_t* sampleCount) { |
| TextureViewBase* attachment = colorAttachment.view; |
| DAWN_TRY(device->ValidateObject(attachment)); |
| DAWN_TRY( |
| ValidateCanUseAs(attachment->GetTexture(), wgpu::TextureUsage::RenderAttachment)); |
| |
| if (!(attachment->GetAspects() & Aspect::Color) || |
| !attachment->GetFormat().isRenderable) { |
| return DAWN_VALIDATION_ERROR( |
| "The format of the texture view used as color attachment is not color " |
| "renderable"); |
| } |
| |
| DAWN_TRY(ValidateLoadOp(colorAttachment.loadOp)); |
| DAWN_TRY(ValidateStoreOp(colorAttachment.storeOp)); |
| |
| if (colorAttachment.loadOp == wgpu::LoadOp::Clear) { |
| if (std::isnan(colorAttachment.clearColor.r) || |
| std::isnan(colorAttachment.clearColor.g) || |
| std::isnan(colorAttachment.clearColor.b) || |
| std::isnan(colorAttachment.clearColor.a)) { |
| return DAWN_VALIDATION_ERROR("Color clear value cannot contain NaN"); |
| } |
| } |
| |
| DAWN_TRY(ValidateOrSetColorAttachmentSampleCount(attachment, sampleCount)); |
| |
| DAWN_TRY(ValidateResolveTarget(device, colorAttachment)); |
| |
| DAWN_TRY(ValidateAttachmentArrayLayersAndLevelCount(attachment)); |
| DAWN_TRY(ValidateOrSetAttachmentSize(attachment, width, height)); |
| |
| return {}; |
| } |
| |
| MaybeError ValidateRenderPassDepthStencilAttachment( |
| DeviceBase* device, |
| const RenderPassDepthStencilAttachment* depthStencilAttachment, |
| uint32_t* width, |
| uint32_t* height, |
| uint32_t* sampleCount) { |
| DAWN_ASSERT(depthStencilAttachment != nullptr); |
| |
| TextureViewBase* attachment = depthStencilAttachment->view; |
| DAWN_TRY(device->ValidateObject(attachment)); |
| DAWN_TRY( |
| ValidateCanUseAs(attachment->GetTexture(), wgpu::TextureUsage::RenderAttachment)); |
| |
| const Format& format = attachment->GetFormat(); |
| if (!format.HasDepthOrStencil()) { |
| return DAWN_VALIDATION_ERROR( |
| "The format of the texture view used as depth stencil attachment is not a " |
| "depth stencil format"); |
| } |
| if (!format.isRenderable) { |
| return DAWN_VALIDATION_ERROR( |
| "The format of the texture view used as depth stencil attachment is not " |
| "renderable"); |
| } |
| if (attachment->GetAspects() != format.aspects) { |
| // TODO(https://crbug.com/dawn/812): Investigate if this limitation should be added |
| // to the WebGPU spec of lifted from Dawn. |
| return DAWN_VALIDATION_ERROR( |
| "The texture view used as depth stencil view must encompass all aspects"); |
| } |
| |
| DAWN_TRY(ValidateLoadOp(depthStencilAttachment->depthLoadOp)); |
| DAWN_TRY(ValidateLoadOp(depthStencilAttachment->stencilLoadOp)); |
| DAWN_TRY(ValidateStoreOp(depthStencilAttachment->depthStoreOp)); |
| DAWN_TRY(ValidateStoreOp(depthStencilAttachment->stencilStoreOp)); |
| |
| if (attachment->GetAspects() == (Aspect::Depth | Aspect::Stencil) && |
| depthStencilAttachment->depthReadOnly != depthStencilAttachment->stencilReadOnly) { |
| return DAWN_VALIDATION_ERROR( |
| "depthReadOnly and stencilReadOnly must be the same when texture aspect is " |
| "'all'"); |
| } |
| |
| if (depthStencilAttachment->depthReadOnly && |
| (depthStencilAttachment->depthLoadOp != wgpu::LoadOp::Load || |
| depthStencilAttachment->depthStoreOp != wgpu::StoreOp::Store)) { |
| return DAWN_VALIDATION_ERROR( |
| "depthLoadOp must be load and depthStoreOp must be store when depthReadOnly " |
| "is true."); |
| } |
| |
| if (depthStencilAttachment->stencilReadOnly && |
| (depthStencilAttachment->stencilLoadOp != wgpu::LoadOp::Load || |
| depthStencilAttachment->stencilStoreOp != wgpu::StoreOp::Store)) { |
| return DAWN_VALIDATION_ERROR( |
| "stencilLoadOp must be load and stencilStoreOp must be store when " |
| "stencilReadOnly " |
| "is true."); |
| } |
| |
| if (depthStencilAttachment->depthLoadOp == wgpu::LoadOp::Clear && |
| std::isnan(depthStencilAttachment->clearDepth)) { |
| return DAWN_VALIDATION_ERROR("Depth clear value cannot be NaN"); |
| } |
| |
| // *sampleCount == 0 must only happen when there is no color attachment. In that case we |
| // do not need to validate the sample count of the depth stencil attachment. |
| const uint32_t depthStencilSampleCount = attachment->GetTexture()->GetSampleCount(); |
| if (*sampleCount != 0) { |
| if (depthStencilSampleCount != *sampleCount) { |
| return DAWN_VALIDATION_ERROR("Depth stencil attachment sample counts mismatch"); |
| } |
| } else { |
| *sampleCount = depthStencilSampleCount; |
| } |
| |
| DAWN_TRY(ValidateAttachmentArrayLayersAndLevelCount(attachment)); |
| DAWN_TRY(ValidateOrSetAttachmentSize(attachment, width, height)); |
| |
| return {}; |
| } |
| |
| MaybeError ValidateRenderPassDescriptor(DeviceBase* device, |
| const RenderPassDescriptor* descriptor, |
| uint32_t* width, |
| uint32_t* height, |
| uint32_t* sampleCount) { |
| if (descriptor->colorAttachmentCount > kMaxColorAttachments) { |
| return DAWN_VALIDATION_ERROR("Setting color attachments out of bounds"); |
| } |
| |
| for (uint32_t i = 0; i < descriptor->colorAttachmentCount; ++i) { |
| DAWN_TRY(ValidateRenderPassColorAttachment(device, descriptor->colorAttachments[i], |
| width, height, sampleCount)); |
| } |
| |
| if (descriptor->depthStencilAttachment != nullptr) { |
| DAWN_TRY(ValidateRenderPassDepthStencilAttachment( |
| device, descriptor->depthStencilAttachment, width, height, sampleCount)); |
| } |
| |
| if (descriptor->occlusionQuerySet != nullptr) { |
| DAWN_TRY(device->ValidateObject(descriptor->occlusionQuerySet)); |
| |
| if (descriptor->occlusionQuerySet->GetQueryType() != wgpu::QueryType::Occlusion) { |
| return DAWN_VALIDATION_ERROR("The type of query set must be Occlusion"); |
| } |
| } |
| |
| if (descriptor->colorAttachmentCount == 0 && |
| descriptor->depthStencilAttachment == nullptr) { |
| return DAWN_VALIDATION_ERROR("Cannot use render pass with no attachments."); |
| } |
| |
| return {}; |
| } |
| |
| MaybeError ValidateComputePassDescriptor(const DeviceBase* device, |
| const ComputePassDescriptor* descriptor) { |
| return {}; |
| } |
| |
| MaybeError ValidateQuerySetResolve(const QuerySetBase* querySet, |
| uint32_t firstQuery, |
| uint32_t queryCount, |
| const BufferBase* destination, |
| uint64_t destinationOffset) { |
| if (firstQuery >= querySet->GetQueryCount()) { |
| return DAWN_VALIDATION_ERROR("Query index out of bounds"); |
| } |
| |
| if (queryCount > querySet->GetQueryCount() - firstQuery) { |
| return DAWN_VALIDATION_ERROR( |
| "The sum of firstQuery and queryCount exceeds the number of queries in query " |
| "set"); |
| } |
| |
| if (destinationOffset % 256 != 0) { |
| return DAWN_VALIDATION_ERROR( |
| "The alignment offset into the destination buffer must be a multiple of 256 " |
| "bytes"); |
| } |
| |
| uint64_t bufferSize = destination->GetSize(); |
| // The destination buffer must have enough storage, from destination offset, to contain |
| // the result of resolved queries |
| bool fitsInBuffer = destinationOffset <= bufferSize && |
| (static_cast<uint64_t>(queryCount) * sizeof(uint64_t) <= |
| (bufferSize - destinationOffset)); |
| if (!fitsInBuffer) { |
| return DAWN_VALIDATION_ERROR("The resolved query data would overflow the buffer"); |
| } |
| |
| return {}; |
| } |
| |
| MaybeError EncodeTimestampsToNanosecondsConversion(CommandEncoder* encoder, |
| QuerySetBase* querySet, |
| uint32_t firstQuery, |
| uint32_t queryCount, |
| BufferBase* destination, |
| uint64_t destinationOffset) { |
| DeviceBase* device = encoder->GetDevice(); |
| |
| // The availability got from query set is a reference to vector<bool>, need to covert |
| // bool to uint32_t due to a user input in pipeline must not contain a bool type in |
| // WGSL. |
| std::vector<uint32_t> availability{querySet->GetQueryAvailability().begin(), |
| querySet->GetQueryAvailability().end()}; |
| |
| // Timestamp availability storage buffer |
| BufferDescriptor availabilityDesc = {}; |
| availabilityDesc.usage = wgpu::BufferUsage::Storage | wgpu::BufferUsage::CopyDst; |
| availabilityDesc.size = querySet->GetQueryCount() * sizeof(uint32_t); |
| Ref<BufferBase> availabilityBuffer; |
| DAWN_TRY_ASSIGN(availabilityBuffer, device->CreateBuffer(&availabilityDesc)); |
| |
| DAWN_TRY(device->GetQueue()->WriteBuffer(availabilityBuffer.Get(), 0, |
| availability.data(), |
| availability.size() * sizeof(uint32_t))); |
| |
| // Timestamp params uniform buffer |
| TimestampParams params = {firstQuery, queryCount, |
| static_cast<uint32_t>(destinationOffset), |
| device->GetTimestampPeriodInNS()}; |
| |
| BufferDescriptor parmsDesc = {}; |
| parmsDesc.usage = wgpu::BufferUsage::Uniform | wgpu::BufferUsage::CopyDst; |
| parmsDesc.size = sizeof(params); |
| Ref<BufferBase> paramsBuffer; |
| DAWN_TRY_ASSIGN(paramsBuffer, device->CreateBuffer(&parmsDesc)); |
| |
| DAWN_TRY( |
| device->GetQueue()->WriteBuffer(paramsBuffer.Get(), 0, ¶ms, sizeof(params))); |
| |
| return EncodeConvertTimestampsToNanoseconds( |
| encoder, destination, availabilityBuffer.Get(), paramsBuffer.Get()); |
| } |
| |
| } // namespace |
| |
| CommandEncoder::CommandEncoder(DeviceBase* device, const CommandEncoderDescriptor*) |
| : ObjectBase(device, kLabelNotImplemented), mEncodingContext(device, this) { |
| } |
| |
| CommandBufferResourceUsage CommandEncoder::AcquireResourceUsages() { |
| return CommandBufferResourceUsage{ |
| mEncodingContext.AcquireRenderPassUsages(), mEncodingContext.AcquireComputePassUsages(), |
| std::move(mTopLevelBuffers), std::move(mTopLevelTextures), std::move(mUsedQuerySets)}; |
| } |
| |
| CommandIterator CommandEncoder::AcquireCommands() { |
| return mEncodingContext.AcquireCommands(); |
| } |
| |
| void CommandEncoder::TrackUsedQuerySet(QuerySetBase* querySet) { |
| mUsedQuerySets.insert(querySet); |
| } |
| |
| void CommandEncoder::TrackQueryAvailability(QuerySetBase* querySet, uint32_t queryIndex) { |
| DAWN_ASSERT(querySet != nullptr); |
| |
| if (GetDevice()->IsValidationEnabled()) { |
| TrackUsedQuerySet(querySet); |
| } |
| |
| // Set the query at queryIndex to available for resolving in query set. |
| querySet->SetQueryAvailability(queryIndex, true); |
| } |
| |
| // Implementation of the API's command recording methods |
| |
| ComputePassEncoder* CommandEncoder::APIBeginComputePass( |
| const ComputePassDescriptor* descriptor) { |
| DeviceBase* device = GetDevice(); |
| |
| bool success = |
| mEncodingContext.TryEncode(this, [&](CommandAllocator* allocator) -> MaybeError { |
| DAWN_TRY(ValidateComputePassDescriptor(device, descriptor)); |
| |
| allocator->Allocate<BeginComputePassCmd>(Command::BeginComputePass); |
| |
| return {}; |
| }); |
| |
| if (success) { |
| ComputePassEncoder* passEncoder = |
| new ComputePassEncoder(device, this, &mEncodingContext); |
| mEncodingContext.EnterPass(passEncoder); |
| return passEncoder; |
| } |
| |
| return ComputePassEncoder::MakeError(device, this, &mEncodingContext); |
| } |
| |
| RenderPassEncoder* CommandEncoder::APIBeginRenderPass(const RenderPassDescriptor* descriptor) { |
| DeviceBase* device = GetDevice(); |
| |
| RenderPassResourceUsageTracker usageTracker; |
| |
| uint32_t width = 0; |
| uint32_t height = 0; |
| Ref<AttachmentState> attachmentState; |
| mEncodingContext.WillBeginRenderPass(); |
| bool success = |
| mEncodingContext.TryEncode(this, [&](CommandAllocator* allocator) -> MaybeError { |
| uint32_t sampleCount = 0; |
| |
| DAWN_TRY(ValidateRenderPassDescriptor(device, descriptor, &width, &height, |
| &sampleCount)); |
| |
| ASSERT(width > 0 && height > 0 && sampleCount > 0); |
| |
| BeginRenderPassCmd* cmd = |
| allocator->Allocate<BeginRenderPassCmd>(Command::BeginRenderPass); |
| |
| cmd->attachmentState = device->GetOrCreateAttachmentState(descriptor); |
| attachmentState = cmd->attachmentState; |
| |
| for (ColorAttachmentIndex index : |
| IterateBitSet(cmd->attachmentState->GetColorAttachmentsMask())) { |
| uint8_t i = static_cast<uint8_t>(index); |
| TextureViewBase* view = descriptor->colorAttachments[i].view; |
| TextureViewBase* resolveTarget = descriptor->colorAttachments[i].resolveTarget; |
| |
| cmd->colorAttachments[index].view = view; |
| cmd->colorAttachments[index].resolveTarget = resolveTarget; |
| cmd->colorAttachments[index].loadOp = descriptor->colorAttachments[i].loadOp; |
| cmd->colorAttachments[index].storeOp = descriptor->colorAttachments[i].storeOp; |
| cmd->colorAttachments[index].clearColor = |
| descriptor->colorAttachments[i].clearColor; |
| |
| usageTracker.TextureViewUsedAs(view, wgpu::TextureUsage::RenderAttachment); |
| |
| if (resolveTarget != nullptr) { |
| usageTracker.TextureViewUsedAs(resolveTarget, |
| wgpu::TextureUsage::RenderAttachment); |
| } |
| } |
| |
| if (cmd->attachmentState->HasDepthStencilAttachment()) { |
| TextureViewBase* view = descriptor->depthStencilAttachment->view; |
| |
| cmd->depthStencilAttachment.view = view; |
| cmd->depthStencilAttachment.clearDepth = |
| descriptor->depthStencilAttachment->clearDepth; |
| cmd->depthStencilAttachment.clearStencil = |
| descriptor->depthStencilAttachment->clearStencil; |
| cmd->depthStencilAttachment.depthLoadOp = |
| descriptor->depthStencilAttachment->depthLoadOp; |
| cmd->depthStencilAttachment.depthStoreOp = |
| descriptor->depthStencilAttachment->depthStoreOp; |
| cmd->depthStencilAttachment.stencilLoadOp = |
| descriptor->depthStencilAttachment->stencilLoadOp; |
| cmd->depthStencilAttachment.stencilStoreOp = |
| descriptor->depthStencilAttachment->stencilStoreOp; |
| |
| usageTracker.TextureViewUsedAs(view, wgpu::TextureUsage::RenderAttachment); |
| } |
| |
| cmd->width = width; |
| cmd->height = height; |
| |
| cmd->occlusionQuerySet = descriptor->occlusionQuerySet; |
| |
| return {}; |
| }); |
| |
| if (success) { |
| RenderPassEncoder* passEncoder = new RenderPassEncoder( |
| device, this, &mEncodingContext, std::move(usageTracker), |
| std::move(attachmentState), descriptor->occlusionQuerySet, width, height); |
| mEncodingContext.EnterPass(passEncoder); |
| return passEncoder; |
| } |
| |
| return RenderPassEncoder::MakeError(device, this, &mEncodingContext); |
| } |
| |
| void CommandEncoder::APICopyBufferToBuffer(BufferBase* source, |
| uint64_t sourceOffset, |
| BufferBase* destination, |
| uint64_t destinationOffset, |
| uint64_t size) { |
| mEncodingContext.TryEncode(this, [&](CommandAllocator* allocator) -> MaybeError { |
| if (GetDevice()->IsValidationEnabled()) { |
| DAWN_TRY(GetDevice()->ValidateObject(source)); |
| DAWN_TRY(GetDevice()->ValidateObject(destination)); |
| |
| if (source == destination) { |
| return DAWN_VALIDATION_ERROR( |
| "Source and destination cannot be the same buffer."); |
| } |
| |
| DAWN_TRY(ValidateCopySizeFitsInBuffer(source, sourceOffset, size)); |
| DAWN_TRY(ValidateCopySizeFitsInBuffer(destination, destinationOffset, size)); |
| DAWN_TRY(ValidateB2BCopyAlignment(size, sourceOffset, destinationOffset)); |
| |
| DAWN_TRY(ValidateCanUseAs(source, wgpu::BufferUsage::CopySrc)); |
| DAWN_TRY(ValidateCanUseAs(destination, wgpu::BufferUsage::CopyDst)); |
| |
| mTopLevelBuffers.insert(source); |
| mTopLevelBuffers.insert(destination); |
| } |
| |
| CopyBufferToBufferCmd* copy = |
| allocator->Allocate<CopyBufferToBufferCmd>(Command::CopyBufferToBuffer); |
| copy->source = source; |
| copy->sourceOffset = sourceOffset; |
| copy->destination = destination; |
| copy->destinationOffset = destinationOffset; |
| copy->size = size; |
| |
| return {}; |
| }); |
| } |
| |
| void CommandEncoder::APICopyBufferToTexture(const ImageCopyBuffer* source, |
| const ImageCopyTexture* destination, |
| const Extent3D* copySize) { |
| mEncodingContext.TryEncode(this, [&](CommandAllocator* allocator) -> MaybeError { |
| if (GetDevice()->IsValidationEnabled()) { |
| DAWN_TRY(ValidateImageCopyBuffer(GetDevice(), *source)); |
| DAWN_TRY(ValidateCanUseAs(source->buffer, wgpu::BufferUsage::CopySrc)); |
| |
| DAWN_TRY(ValidateImageCopyTexture(GetDevice(), *destination, *copySize)); |
| DAWN_TRY(ValidateCanUseAs(destination->texture, wgpu::TextureUsage::CopyDst)); |
| DAWN_TRY(ValidateTextureSampleCountInBufferCopyCommands(destination->texture)); |
| |
| DAWN_TRY(ValidateLinearToDepthStencilCopyRestrictions(*destination)); |
| // We validate texture copy range before validating linear texture data, |
| // because in the latter we divide copyExtent.width by blockWidth and |
| // copyExtent.height by blockHeight while the divisibility conditions are |
| // checked in validating texture copy range. |
| DAWN_TRY(ValidateTextureCopyRange(GetDevice(), *destination, *copySize)); |
| } |
| const TexelBlockInfo& blockInfo = |
| destination->texture->GetFormat().GetAspectInfo(destination->aspect).block; |
| if (GetDevice()->IsValidationEnabled()) { |
| DAWN_TRY(ValidateLinearTextureCopyOffset( |
| source->layout, blockInfo, |
| destination->texture->GetFormat().HasDepthOrStencil())); |
| DAWN_TRY(ValidateLinearTextureData(source->layout, source->buffer->GetSize(), |
| blockInfo, *copySize)); |
| |
| mTopLevelBuffers.insert(source->buffer); |
| mTopLevelTextures.insert(destination->texture); |
| } |
| |
| TextureDataLayout srcLayout = source->layout; |
| ApplyDefaultTextureDataLayoutOptions(&srcLayout, blockInfo, *copySize); |
| |
| CopyBufferToTextureCmd* copy = |
| allocator->Allocate<CopyBufferToTextureCmd>(Command::CopyBufferToTexture); |
| copy->source.buffer = source->buffer; |
| copy->source.offset = srcLayout.offset; |
| copy->source.bytesPerRow = srcLayout.bytesPerRow; |
| copy->source.rowsPerImage = srcLayout.rowsPerImage; |
| copy->destination.texture = destination->texture; |
| copy->destination.origin = destination->origin; |
| copy->destination.mipLevel = destination->mipLevel; |
| copy->destination.aspect = |
| ConvertAspect(destination->texture->GetFormat(), destination->aspect); |
| copy->copySize = *copySize; |
| |
| return {}; |
| }); |
| } |
| |
| void CommandEncoder::APICopyTextureToBuffer(const ImageCopyTexture* source, |
| const ImageCopyBuffer* destination, |
| const Extent3D* copySize) { |
| mEncodingContext.TryEncode(this, [&](CommandAllocator* allocator) -> MaybeError { |
| if (GetDevice()->IsValidationEnabled()) { |
| DAWN_TRY(ValidateImageCopyTexture(GetDevice(), *source, *copySize)); |
| DAWN_TRY(ValidateCanUseAs(source->texture, wgpu::TextureUsage::CopySrc)); |
| DAWN_TRY(ValidateTextureSampleCountInBufferCopyCommands(source->texture)); |
| DAWN_TRY(ValidateTextureDepthStencilToBufferCopyRestrictions(*source)); |
| |
| DAWN_TRY(ValidateImageCopyBuffer(GetDevice(), *destination)); |
| DAWN_TRY(ValidateCanUseAs(destination->buffer, wgpu::BufferUsage::CopyDst)); |
| |
| // We validate texture copy range before validating linear texture data, |
| // because in the latter we divide copyExtent.width by blockWidth and |
| // copyExtent.height by blockHeight while the divisibility conditions are |
| // checked in validating texture copy range. |
| DAWN_TRY(ValidateTextureCopyRange(GetDevice(), *source, *copySize)); |
| } |
| const TexelBlockInfo& blockInfo = |
| source->texture->GetFormat().GetAspectInfo(source->aspect).block; |
| if (GetDevice()->IsValidationEnabled()) { |
| DAWN_TRY(ValidateLinearTextureCopyOffset( |
| destination->layout, blockInfo, |
| source->texture->GetFormat().HasDepthOrStencil())); |
| DAWN_TRY(ValidateLinearTextureData( |
| destination->layout, destination->buffer->GetSize(), blockInfo, *copySize)); |
| |
| mTopLevelTextures.insert(source->texture); |
| mTopLevelBuffers.insert(destination->buffer); |
| } |
| |
| TextureDataLayout dstLayout = destination->layout; |
| ApplyDefaultTextureDataLayoutOptions(&dstLayout, blockInfo, *copySize); |
| |
| CopyTextureToBufferCmd* copy = |
| allocator->Allocate<CopyTextureToBufferCmd>(Command::CopyTextureToBuffer); |
| copy->source.texture = source->texture; |
| copy->source.origin = source->origin; |
| copy->source.mipLevel = source->mipLevel; |
| copy->source.aspect = ConvertAspect(source->texture->GetFormat(), source->aspect); |
| copy->destination.buffer = destination->buffer; |
| copy->destination.offset = dstLayout.offset; |
| copy->destination.bytesPerRow = dstLayout.bytesPerRow; |
| copy->destination.rowsPerImage = dstLayout.rowsPerImage; |
| copy->copySize = *copySize; |
| |
| return {}; |
| }); |
| } |
| |
| void CommandEncoder::APICopyTextureToTexture(const ImageCopyTexture* source, |
| const ImageCopyTexture* destination, |
| const Extent3D* copySize) { |
| APICopyTextureToTextureHelper<false>(source, destination, copySize); |
| } |
| |
| void CommandEncoder::APICopyTextureToTextureInternal(const ImageCopyTexture* source, |
| const ImageCopyTexture* destination, |
| const Extent3D* copySize) { |
| APICopyTextureToTextureHelper<true>(source, destination, copySize); |
| } |
| |
| template <bool Internal> |
| void CommandEncoder::APICopyTextureToTextureHelper(const ImageCopyTexture* source, |
| const ImageCopyTexture* destination, |
| const Extent3D* copySize) { |
| mEncodingContext.TryEncode(this, [&](CommandAllocator* allocator) -> MaybeError { |
| if (GetDevice()->IsValidationEnabled()) { |
| DAWN_TRY(GetDevice()->ValidateObject(source->texture)); |
| DAWN_TRY(GetDevice()->ValidateObject(destination->texture)); |
| |
| DAWN_TRY(ValidateImageCopyTexture(GetDevice(), *source, *copySize)); |
| DAWN_TRY(ValidateImageCopyTexture(GetDevice(), *destination, *copySize)); |
| |
| DAWN_TRY( |
| ValidateTextureToTextureCopyRestrictions(*source, *destination, *copySize)); |
| |
| DAWN_TRY(ValidateTextureCopyRange(GetDevice(), *source, *copySize)); |
| DAWN_TRY(ValidateTextureCopyRange(GetDevice(), *destination, *copySize)); |
| |
| // For internal usages (CopyToCopyInternal) we don't care if the user has added |
| // CopySrc as a usage for this texture, but we will always add it internally. |
| if (Internal) { |
| DAWN_TRY( |
| ValidateInternalCanUseAs(source->texture, wgpu::TextureUsage::CopySrc)); |
| DAWN_TRY(ValidateInternalCanUseAs(destination->texture, |
| wgpu::TextureUsage::CopyDst)); |
| } else { |
| DAWN_TRY(ValidateCanUseAs(source->texture, wgpu::TextureUsage::CopySrc)); |
| DAWN_TRY(ValidateCanUseAs(destination->texture, wgpu::TextureUsage::CopyDst)); |
| } |
| |
| mTopLevelTextures.insert(source->texture); |
| mTopLevelTextures.insert(destination->texture); |
| } |
| |
| CopyTextureToTextureCmd* copy = |
| allocator->Allocate<CopyTextureToTextureCmd>(Command::CopyTextureToTexture); |
| copy->source.texture = source->texture; |
| copy->source.origin = source->origin; |
| copy->source.mipLevel = source->mipLevel; |
| copy->source.aspect = ConvertAspect(source->texture->GetFormat(), source->aspect); |
| copy->destination.texture = destination->texture; |
| copy->destination.origin = destination->origin; |
| copy->destination.mipLevel = destination->mipLevel; |
| copy->destination.aspect = |
| ConvertAspect(destination->texture->GetFormat(), destination->aspect); |
| copy->copySize = *copySize; |
| |
| return {}; |
| }); |
| } |
| |
| void CommandEncoder::APIInjectValidationError(const char* message) { |
| if (mEncodingContext.CheckCurrentEncoder(this)) { |
| mEncodingContext.HandleError(DAWN_VALIDATION_ERROR(message)); |
| } |
| } |
| |
| void CommandEncoder::APIInsertDebugMarker(const char* groupLabel) { |
| mEncodingContext.TryEncode(this, [&](CommandAllocator* allocator) -> MaybeError { |
| InsertDebugMarkerCmd* cmd = |
| allocator->Allocate<InsertDebugMarkerCmd>(Command::InsertDebugMarker); |
| cmd->length = strlen(groupLabel); |
| |
| char* label = allocator->AllocateData<char>(cmd->length + 1); |
| memcpy(label, groupLabel, cmd->length + 1); |
| |
| return {}; |
| }); |
| } |
| |
| void CommandEncoder::APIPopDebugGroup() { |
| mEncodingContext.TryEncode(this, [&](CommandAllocator* allocator) -> MaybeError { |
| if (GetDevice()->IsValidationEnabled()) { |
| if (mDebugGroupStackSize == 0) { |
| return DAWN_VALIDATION_ERROR("Pop must be balanced by a corresponding Push."); |
| } |
| } |
| allocator->Allocate<PopDebugGroupCmd>(Command::PopDebugGroup); |
| mDebugGroupStackSize--; |
| |
| return {}; |
| }); |
| } |
| |
| void CommandEncoder::APIPushDebugGroup(const char* groupLabel) { |
| mEncodingContext.TryEncode(this, [&](CommandAllocator* allocator) -> MaybeError { |
| PushDebugGroupCmd* cmd = |
| allocator->Allocate<PushDebugGroupCmd>(Command::PushDebugGroup); |
| cmd->length = strlen(groupLabel); |
| |
| char* label = allocator->AllocateData<char>(cmd->length + 1); |
| memcpy(label, groupLabel, cmd->length + 1); |
| |
| mDebugGroupStackSize++; |
| |
| return {}; |
| }); |
| } |
| |
| void CommandEncoder::APIResolveQuerySet(QuerySetBase* querySet, |
| uint32_t firstQuery, |
| uint32_t queryCount, |
| BufferBase* destination, |
| uint64_t destinationOffset) { |
| mEncodingContext.TryEncode(this, [&](CommandAllocator* allocator) -> MaybeError { |
| if (GetDevice()->IsValidationEnabled()) { |
| DAWN_TRY(GetDevice()->ValidateObject(querySet)); |
| DAWN_TRY(GetDevice()->ValidateObject(destination)); |
| |
| DAWN_TRY(ValidateQuerySetResolve(querySet, firstQuery, queryCount, destination, |
| destinationOffset)); |
| |
| DAWN_TRY(ValidateCanUseAs(destination, wgpu::BufferUsage::QueryResolve)); |
| |
| TrackUsedQuerySet(querySet); |
| mTopLevelBuffers.insert(destination); |
| } |
| |
| ResolveQuerySetCmd* cmd = |
| allocator->Allocate<ResolveQuerySetCmd>(Command::ResolveQuerySet); |
| cmd->querySet = querySet; |
| cmd->firstQuery = firstQuery; |
| cmd->queryCount = queryCount; |
| cmd->destination = destination; |
| cmd->destinationOffset = destinationOffset; |
| |
| // Encode internal compute pipeline for timestamp query |
| if (querySet->GetQueryType() == wgpu::QueryType::Timestamp) { |
| DAWN_TRY(EncodeTimestampsToNanosecondsConversion( |
| this, querySet, firstQuery, queryCount, destination, destinationOffset)); |
| } |
| |
| return {}; |
| }); |
| } |
| |
| void CommandEncoder::APIWriteBuffer(BufferBase* buffer, |
| uint64_t bufferOffset, |
| const uint8_t* data, |
| uint64_t size) { |
| mEncodingContext.TryEncode(this, [&](CommandAllocator* allocator) -> MaybeError { |
| if (GetDevice()->IsValidationEnabled()) { |
| DAWN_TRY(ValidateWriteBuffer(GetDevice(), buffer, bufferOffset, size)); |
| } |
| |
| WriteBufferCmd* cmd = allocator->Allocate<WriteBufferCmd>(Command::WriteBuffer); |
| cmd->buffer = buffer; |
| cmd->offset = bufferOffset; |
| cmd->size = size; |
| |
| uint8_t* inlinedData = allocator->AllocateData<uint8_t>(size); |
| memcpy(inlinedData, data, size); |
| |
| mTopLevelBuffers.insert(buffer); |
| |
| return {}; |
| }); |
| } |
| |
| void CommandEncoder::APIWriteTimestamp(QuerySetBase* querySet, uint32_t queryIndex) { |
| mEncodingContext.TryEncode(this, [&](CommandAllocator* allocator) -> MaybeError { |
| if (GetDevice()->IsValidationEnabled()) { |
| DAWN_TRY(GetDevice()->ValidateObject(querySet)); |
| DAWN_TRY(ValidateTimestampQuery(querySet, queryIndex)); |
| } |
| |
| TrackQueryAvailability(querySet, queryIndex); |
| |
| WriteTimestampCmd* cmd = |
| allocator->Allocate<WriteTimestampCmd>(Command::WriteTimestamp); |
| cmd->querySet = querySet; |
| cmd->queryIndex = queryIndex; |
| |
| return {}; |
| }); |
| } |
| |
| CommandBufferBase* CommandEncoder::APIFinish(const CommandBufferDescriptor* descriptor) { |
| Ref<CommandBufferBase> commandBuffer; |
| if (GetDevice()->ConsumedError(FinishInternal(descriptor), &commandBuffer)) { |
| return CommandBufferBase::MakeError(GetDevice()); |
| } |
| ASSERT(!IsError()); |
| return commandBuffer.Detach(); |
| } |
| |
| void CommandEncoder::EncodeSetValidatedBufferLocationsInternal( |
| std::vector<DeferredBufferLocationUpdate> updates) { |
| ASSERT(GetDevice()->IsValidationEnabled()); |
| mEncodingContext.TryEncode(this, [&](CommandAllocator* allocator) -> MaybeError { |
| SetValidatedBufferLocationsInternalCmd* cmd = |
| allocator->Allocate<SetValidatedBufferLocationsInternalCmd>( |
| Command::SetValidatedBufferLocationsInternal); |
| cmd->updates = std::move(updates); |
| return {}; |
| }); |
| } |
| |
| ResultOrError<Ref<CommandBufferBase>> CommandEncoder::FinishInternal( |
| const CommandBufferDescriptor* descriptor) { |
| DeviceBase* device = GetDevice(); |
| |
| // Even if mEncodingContext.Finish() validation fails, calling it will mutate the internal |
| // state of the encoding context. The internal state is set to finished, and subsequent |
| // calls to encode commands will generate errors. |
| DAWN_TRY(mEncodingContext.Finish()); |
| DAWN_TRY(device->ValidateIsAlive()); |
| |
| if (device->IsValidationEnabled()) { |
| DAWN_TRY(ValidateFinish()); |
| } |
| return device->CreateCommandBuffer(this, descriptor); |
| } |
| |
| // Implementation of the command buffer validation that can be precomputed before submit |
| MaybeError CommandEncoder::ValidateFinish() const { |
| TRACE_EVENT0(GetDevice()->GetPlatform(), Validation, "CommandEncoder::ValidateFinish"); |
| DAWN_TRY(GetDevice()->ValidateObject(this)); |
| |
| for (const RenderPassResourceUsage& passUsage : mEncodingContext.GetRenderPassUsages()) { |
| DAWN_TRY(ValidateSyncScopeResourceUsage(passUsage)); |
| } |
| |
| for (const ComputePassResourceUsage& passUsage : mEncodingContext.GetComputePassUsages()) { |
| for (const SyncScopeResourceUsage& scope : passUsage.dispatchUsages) { |
| DAWN_TRY(ValidateSyncScopeResourceUsage(scope)); |
| } |
| } |
| |
| if (mDebugGroupStackSize != 0) { |
| return DAWN_VALIDATION_ERROR("Each Push must be balanced by a corresponding Pop."); |
| } |
| |
| return {}; |
| } |
| |
| } // namespace dawn_native |