| // Copyright 2019 The Dawn Authors |
| // |
| // Licensed under the Apache License, Version 2.0 (the "License"); |
| // you may not use this file except in compliance with the License. |
| // You may obtain a copy of the License at |
| // |
| // http://www.apache.org/licenses/LICENSE-2.0 |
| // |
| // Unless required by applicable law or agreed to in writing, software |
| // distributed under the License is distributed on an "AS IS" BASIS, |
| // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| // See the License for the specific language governing permissions and |
| // limitations under the License. |
| |
| #include "common/Assert.h" |
| #include "common/Constants.h" |
| #include "common/Math.h" |
| #include "tests/DawnTest.h" |
| #include "utils/ComboRenderPipelineDescriptor.h" |
| #include "utils/WGPUHelpers.h" |
| |
| class GpuMemorySyncTests : public DawnTest { |
| protected: |
| wgpu::Buffer CreateBuffer() { |
| wgpu::BufferDescriptor srcDesc; |
| srcDesc.size = 4; |
| srcDesc.usage = |
| wgpu::BufferUsage::CopySrc | wgpu::BufferUsage::CopyDst | wgpu::BufferUsage::Storage; |
| wgpu::Buffer buffer = device.CreateBuffer(&srcDesc); |
| |
| int myData = 0; |
| buffer.SetSubData(0, sizeof(myData), &myData); |
| return buffer; |
| } |
| |
| std::tuple<wgpu::ComputePipeline, wgpu::BindGroup> CreatePipelineAndBindGroupForCompute( |
| const wgpu::Buffer& buffer) { |
| wgpu::ShaderModule csModule = |
| utils::CreateShaderModule(device, utils::SingleShaderStage::Compute, R"( |
| #version 450 |
| layout(std140, set = 0, binding = 0) buffer Data { |
| int a; |
| } data; |
| void main() { |
| data.a += 1; |
| })"); |
| |
| wgpu::BindGroupLayout bgl = utils::MakeBindGroupLayout( |
| device, { |
| {0, wgpu::ShaderStage::Compute, wgpu::BindingType::StorageBuffer}, |
| }); |
| wgpu::PipelineLayout pipelineLayout0 = utils::MakeBasicPipelineLayout(device, &bgl); |
| |
| wgpu::ComputePipelineDescriptor cpDesc; |
| cpDesc.layout = pipelineLayout0; |
| cpDesc.computeStage.module = csModule; |
| cpDesc.computeStage.entryPoint = "main"; |
| wgpu::ComputePipeline pipeline = device.CreateComputePipeline(&cpDesc); |
| |
| wgpu::BindGroup bindGroup = |
| utils::MakeBindGroup(device, bgl, {{0, buffer, 0, sizeof(int)}}); |
| return std::make_tuple(pipeline, bindGroup); |
| } |
| |
| std::tuple<wgpu::RenderPipeline, wgpu::BindGroup> CreatePipelineAndBindGroupForRender( |
| const wgpu::Buffer& buffer, |
| wgpu::TextureFormat colorFormat) { |
| wgpu::ShaderModule vsModule = |
| utils::CreateShaderModule(device, utils::SingleShaderStage::Vertex, R"( |
| #version 450 |
| void main() { |
| gl_Position = vec4(0.f, 0.f, 0.f, 1.f); |
| gl_PointSize = 1.0; |
| })"); |
| |
| wgpu::ShaderModule fsModule = |
| utils::CreateShaderModule(device, utils::SingleShaderStage::Fragment, R"( |
| #version 450 |
| layout (set = 0, binding = 0) buffer Data { |
| int i; |
| } data; |
| layout(location = 0) out vec4 fragColor; |
| void main() { |
| data.i += 1; |
| fragColor = vec4(data.i / 255.f, 0.f, 0.f, 1.f); |
| })"); |
| |
| wgpu::BindGroupLayout bgl = utils::MakeBindGroupLayout( |
| device, { |
| {0, wgpu::ShaderStage::Fragment, wgpu::BindingType::StorageBuffer}, |
| }); |
| wgpu::PipelineLayout pipelineLayout = utils::MakeBasicPipelineLayout(device, &bgl); |
| |
| utils::ComboRenderPipelineDescriptor rpDesc(device); |
| rpDesc.layout = pipelineLayout; |
| rpDesc.vertexStage.module = vsModule; |
| rpDesc.cFragmentStage.module = fsModule; |
| rpDesc.primitiveTopology = wgpu::PrimitiveTopology::PointList; |
| rpDesc.cColorStates[0].format = colorFormat; |
| |
| wgpu::RenderPipeline pipeline = device.CreateRenderPipeline(&rpDesc); |
| |
| wgpu::BindGroup bindGroup = |
| utils::MakeBindGroup(device, bgl, {{0, buffer, 0, sizeof(int)}}); |
| return std::make_tuple(pipeline, bindGroup); |
| } |
| }; |
| |
| // Clear storage buffer with zero. Then read data, add one, and write the result to storage buffer |
| // in compute pass. Iterate this read-add-write steps per compute pass a few time. The successive |
| // iteration reads the result in buffer from last iteration, which makes the iterations a data |
| // dependency chain. The test verifies that data in buffer among iterations in compute passes is |
| // correctly synchronized. |
| TEST_P(GpuMemorySyncTests, ComputePass) { |
| // Create pipeline, bind group, and buffer for compute pass. |
| wgpu::Buffer buffer = CreateBuffer(); |
| wgpu::ComputePipeline compute; |
| wgpu::BindGroup bindGroup; |
| std::tie(compute, bindGroup) = CreatePipelineAndBindGroupForCompute(buffer); |
| wgpu::CommandEncoder encoder = device.CreateCommandEncoder(); |
| |
| // Iterate the read-add-write operations in compute pass a few times. |
| int iteration = 3; |
| for (int i = 0; i < iteration; ++i) { |
| wgpu::ComputePassEncoder pass = encoder.BeginComputePass(); |
| pass.SetPipeline(compute); |
| pass.SetBindGroup(0, bindGroup); |
| pass.Dispatch(1, 1, 1); |
| pass.EndPass(); |
| } |
| |
| wgpu::CommandBuffer commands = encoder.Finish(); |
| queue.Submit(1, &commands); |
| |
| // Verify the result. |
| EXPECT_BUFFER_U32_EQ(iteration, buffer, 0); |
| } |
| |
| // Clear storage buffer with zero. Then read data, add one, and write the result to storage buffer |
| // in render pass. Iterate this read-add-write steps per render pass a few time. The successive |
| // iteration reads the result in buffer from last iteration, which makes the iterations a data |
| // dependency chain. In addition, color output by fragment shader depends on the data in storage |
| // buffer, so we can check color in render target to verify that data in buffer among iterations in |
| // render passes is correctly synchronized. |
| TEST_P(GpuMemorySyncTests, RenderPass) { |
| // Create pipeline, bind group, and buffer for render pass. |
| wgpu::Buffer buffer = CreateBuffer(); |
| utils::BasicRenderPass renderPass = utils::CreateBasicRenderPass(device, 1, 1); |
| wgpu::RenderPipeline render; |
| wgpu::BindGroup bindGroup; |
| std::tie(render, bindGroup) = |
| CreatePipelineAndBindGroupForRender(buffer, renderPass.colorFormat); |
| wgpu::CommandEncoder encoder = device.CreateCommandEncoder(); |
| |
| // Iterate the read-add-write operations in render pass a few times. |
| int iteration = 3; |
| for (int i = 0; i < iteration; ++i) { |
| wgpu::RenderPassEncoder pass = encoder.BeginRenderPass(&renderPass.renderPassInfo); |
| pass.SetPipeline(render); |
| pass.SetBindGroup(0, bindGroup); |
| pass.Draw(1, 1, 0, 0); |
| pass.EndPass(); |
| } |
| |
| wgpu::CommandBuffer commands = encoder.Finish(); |
| queue.Submit(1, &commands); |
| |
| // Verify the result. |
| EXPECT_PIXEL_RGBA8_EQ(RGBA8(iteration, 0, 0, 255), renderPass.color, 0, 0); |
| } |
| |
| // Write into a storage buffer in a render pass. Then read that data in a compute |
| // pass. And verify the data flow is correctly synchronized. |
| TEST_P(GpuMemorySyncTests, RenderPassToComputePass) { |
| // Create pipeline, bind group, and buffer for render pass and compute pass. |
| wgpu::Buffer buffer = CreateBuffer(); |
| utils::BasicRenderPass renderPass = utils::CreateBasicRenderPass(device, 1, 1); |
| wgpu::RenderPipeline render; |
| wgpu::BindGroup bindGroup0; |
| std::tie(render, bindGroup0) = |
| CreatePipelineAndBindGroupForRender(buffer, renderPass.colorFormat); |
| |
| wgpu::ComputePipeline compute; |
| wgpu::BindGroup bindGroup1; |
| std::tie(compute, bindGroup1) = CreatePipelineAndBindGroupForCompute(buffer); |
| |
| wgpu::CommandEncoder encoder = device.CreateCommandEncoder(); |
| |
| // Write data into a storage buffer in render pass. |
| wgpu::RenderPassEncoder pass0 = encoder.BeginRenderPass(&renderPass.renderPassInfo); |
| pass0.SetPipeline(render); |
| pass0.SetBindGroup(0, bindGroup0); |
| pass0.Draw(1, 1, 0, 0); |
| pass0.EndPass(); |
| |
| // Read that data in render pass. |
| wgpu::ComputePassEncoder pass1 = encoder.BeginComputePass(); |
| pass1.SetPipeline(compute); |
| pass1.SetBindGroup(0, bindGroup1); |
| pass1.Dispatch(1, 1, 1); |
| pass1.EndPass(); |
| |
| wgpu::CommandBuffer commands = encoder.Finish(); |
| queue.Submit(1, &commands); |
| |
| // Verify the result. |
| EXPECT_BUFFER_U32_EQ(2, buffer, 0); |
| } |
| |
| // Write into a storage buffer in a compute pass. Then read that data in a render |
| // pass. And verify the data flow is correctly synchronized. |
| TEST_P(GpuMemorySyncTests, ComputePassToRenderPass) { |
| // Create pipeline, bind group, and buffer for compute pass and render pass. |
| wgpu::Buffer buffer = CreateBuffer(); |
| wgpu::ComputePipeline compute; |
| wgpu::BindGroup bindGroup1; |
| std::tie(compute, bindGroup1) = CreatePipelineAndBindGroupForCompute(buffer); |
| |
| utils::BasicRenderPass renderPass = utils::CreateBasicRenderPass(device, 1, 1); |
| wgpu::RenderPipeline render; |
| wgpu::BindGroup bindGroup0; |
| std::tie(render, bindGroup0) = |
| CreatePipelineAndBindGroupForRender(buffer, renderPass.colorFormat); |
| |
| wgpu::CommandEncoder encoder = device.CreateCommandEncoder(); |
| |
| // Write data into a storage buffer in render pass. |
| wgpu::ComputePassEncoder pass0 = encoder.BeginComputePass(); |
| pass0.SetPipeline(compute); |
| pass0.SetBindGroup(0, bindGroup1); |
| pass0.Dispatch(1, 1, 1); |
| pass0.EndPass(); |
| |
| // Read that data in render pass. |
| wgpu::RenderPassEncoder pass1 = encoder.BeginRenderPass(&renderPass.renderPassInfo); |
| pass1.SetPipeline(render); |
| pass1.SetBindGroup(0, bindGroup0); |
| pass1.Draw(1, 1, 0, 0); |
| pass1.EndPass(); |
| |
| wgpu::CommandBuffer commands = encoder.Finish(); |
| queue.Submit(1, &commands); |
| |
| // Verify the result. |
| EXPECT_PIXEL_RGBA8_EQ(RGBA8(2, 0, 0, 255), renderPass.color, 0, 0); |
| } |
| |
| class StorageToUniformSyncTests : public DawnTest { |
| protected: |
| void CreateBuffer() { |
| wgpu::BufferDescriptor bufferDesc; |
| bufferDesc.size = sizeof(float); |
| bufferDesc.usage = wgpu::BufferUsage::Storage | wgpu::BufferUsage::Uniform; |
| mBuffer = device.CreateBuffer(&bufferDesc); |
| } |
| |
| std::tuple<wgpu::ComputePipeline, wgpu::BindGroup> CreatePipelineAndBindGroupForCompute() { |
| wgpu::ShaderModule csModule = |
| utils::CreateShaderModule(device, utils::SingleShaderStage::Compute, R"( |
| #version 450 |
| layout(std140, set = 0, binding = 0) buffer Data { |
| float a; |
| } data; |
| void main() { |
| data.a = 1.0; |
| })"); |
| |
| wgpu::BindGroupLayout bgl = utils::MakeBindGroupLayout( |
| device, { |
| {0, wgpu::ShaderStage::Compute, wgpu::BindingType::StorageBuffer}, |
| }); |
| wgpu::PipelineLayout pipelineLayout0 = utils::MakeBasicPipelineLayout(device, &bgl); |
| |
| wgpu::ComputePipelineDescriptor cpDesc; |
| cpDesc.layout = pipelineLayout0; |
| cpDesc.computeStage.module = csModule; |
| cpDesc.computeStage.entryPoint = "main"; |
| wgpu::ComputePipeline pipeline = device.CreateComputePipeline(&cpDesc); |
| |
| wgpu::BindGroup bindGroup = |
| utils::MakeBindGroup(device, bgl, {{0, mBuffer, 0, sizeof(float)}}); |
| return std::make_tuple(pipeline, bindGroup); |
| } |
| |
| std::tuple<wgpu::RenderPipeline, wgpu::BindGroup> CreatePipelineAndBindGroupForRender( |
| wgpu::TextureFormat colorFormat) { |
| wgpu::ShaderModule vsModule = |
| utils::CreateShaderModule(device, utils::SingleShaderStage::Vertex, R"( |
| #version 450 |
| void main() { |
| gl_Position = vec4(0.f, 0.f, 0.f, 1.f); |
| gl_PointSize = 1.0; |
| })"); |
| |
| wgpu::ShaderModule fsModule = |
| utils::CreateShaderModule(device, utils::SingleShaderStage::Fragment, R"( |
| #version 450 |
| layout (set = 0, binding = 0) uniform Contents{ |
| float color; |
| }; |
| layout(location = 0) out vec4 fragColor; |
| void main() { |
| fragColor = vec4(color, 0.f, 0.f, 1.f); |
| })"); |
| |
| wgpu::BindGroupLayout bgl = utils::MakeBindGroupLayout( |
| device, { |
| {0, wgpu::ShaderStage::Fragment, wgpu::BindingType::UniformBuffer}, |
| }); |
| wgpu::PipelineLayout pipelineLayout = utils::MakeBasicPipelineLayout(device, &bgl); |
| |
| utils::ComboRenderPipelineDescriptor rpDesc(device); |
| rpDesc.layout = pipelineLayout; |
| rpDesc.vertexStage.module = vsModule; |
| rpDesc.cFragmentStage.module = fsModule; |
| rpDesc.primitiveTopology = wgpu::PrimitiveTopology::PointList; |
| rpDesc.cColorStates[0].format = colorFormat; |
| |
| wgpu::RenderPipeline pipeline = device.CreateRenderPipeline(&rpDesc); |
| |
| wgpu::BindGroup bindGroup = |
| utils::MakeBindGroup(device, bgl, {{0, mBuffer, 0, sizeof(float)}}); |
| return std::make_tuple(pipeline, bindGroup); |
| } |
| |
| wgpu::Buffer mBuffer; |
| }; |
| |
| // Write into a storage buffer in compute pass in a command buffer. Then read that data in a render |
| // pass. The two passes use the same command buffer. |
| TEST_P(StorageToUniformSyncTests, ReadAfterWriteWithSameCommandBuffer) { |
| // Create pipeline, bind group, and buffer for compute pass and render pass. |
| CreateBuffer(); |
| utils::BasicRenderPass renderPass = utils::CreateBasicRenderPass(device, 1, 1); |
| wgpu::ComputePipeline compute; |
| wgpu::BindGroup computeBindGroup; |
| std::tie(compute, computeBindGroup) = CreatePipelineAndBindGroupForCompute(); |
| wgpu::RenderPipeline render; |
| wgpu::BindGroup renderBindGroup; |
| std::tie(render, renderBindGroup) = CreatePipelineAndBindGroupForRender(renderPass.colorFormat); |
| |
| // Write data into a storage buffer in compute pass. |
| wgpu::CommandEncoder encoder0 = device.CreateCommandEncoder(); |
| wgpu::ComputePassEncoder pass0 = encoder0.BeginComputePass(); |
| pass0.SetPipeline(compute); |
| pass0.SetBindGroup(0, computeBindGroup); |
| pass0.Dispatch(1, 1, 1); |
| pass0.EndPass(); |
| |
| // Read that data in render pass. |
| wgpu::RenderPassEncoder pass1 = encoder0.BeginRenderPass(&renderPass.renderPassInfo); |
| pass1.SetPipeline(render); |
| pass1.SetBindGroup(0, renderBindGroup); |
| pass1.Draw(1, 1, 0, 0); |
| pass1.EndPass(); |
| |
| wgpu::CommandBuffer commands = encoder0.Finish(); |
| queue.Submit(1, &commands); |
| |
| // Verify the rendering result. |
| EXPECT_PIXEL_RGBA8_EQ(RGBA8::kRed, renderPass.color, 0, 0); |
| } |
| |
| // Write into a storage buffer in compute pass in a command buffer. Then read that data in a render |
| // pass. The two passes use the different command buffers. The command buffers are submitted to the |
| // queue in one shot. |
| TEST_P(StorageToUniformSyncTests, ReadAfterWriteWithDifferentCommandBuffers) { |
| // Create pipeline, bind group, and buffer for compute pass and render pass. |
| CreateBuffer(); |
| utils::BasicRenderPass renderPass = utils::CreateBasicRenderPass(device, 1, 1); |
| wgpu::ComputePipeline compute; |
| wgpu::BindGroup computeBindGroup; |
| std::tie(compute, computeBindGroup) = CreatePipelineAndBindGroupForCompute(); |
| wgpu::RenderPipeline render; |
| wgpu::BindGroup renderBindGroup; |
| std::tie(render, renderBindGroup) = CreatePipelineAndBindGroupForRender(renderPass.colorFormat); |
| |
| // Write data into a storage buffer in compute pass. |
| wgpu::CommandBuffer cb[2]; |
| wgpu::CommandEncoder encoder0 = device.CreateCommandEncoder(); |
| wgpu::ComputePassEncoder pass0 = encoder0.BeginComputePass(); |
| pass0.SetPipeline(compute); |
| pass0.SetBindGroup(0, computeBindGroup); |
| pass0.Dispatch(1, 1, 1); |
| pass0.EndPass(); |
| cb[0] = encoder0.Finish(); |
| |
| // Read that data in render pass. |
| wgpu::CommandEncoder encoder1 = device.CreateCommandEncoder(); |
| wgpu::RenderPassEncoder pass1 = encoder1.BeginRenderPass(&renderPass.renderPassInfo); |
| pass1.SetPipeline(render); |
| pass1.SetBindGroup(0, renderBindGroup); |
| pass1.Draw(1, 1, 0, 0); |
| pass1.EndPass(); |
| |
| cb[1] = encoder1.Finish(); |
| queue.Submit(2, cb); |
| |
| // Verify the rendering result. |
| EXPECT_PIXEL_RGBA8_EQ(RGBA8::kRed, renderPass.color, 0, 0); |
| } |
| |
| // Write into a storage buffer in compute pass in a command buffer. Then read that data in a render |
| // pass. The two passes use the different command buffers. The command buffers are submitted to the |
| // queue separately. |
| TEST_P(StorageToUniformSyncTests, ReadAfterWriteWithDifferentQueueSubmits) { |
| // Create pipeline, bind group, and buffer for compute pass and render pass. |
| CreateBuffer(); |
| utils::BasicRenderPass renderPass = utils::CreateBasicRenderPass(device, 1, 1); |
| wgpu::ComputePipeline compute; |
| wgpu::BindGroup computeBindGroup; |
| std::tie(compute, computeBindGroup) = CreatePipelineAndBindGroupForCompute(); |
| wgpu::RenderPipeline render; |
| wgpu::BindGroup renderBindGroup; |
| std::tie(render, renderBindGroup) = CreatePipelineAndBindGroupForRender(renderPass.colorFormat); |
| |
| // Write data into a storage buffer in compute pass. |
| wgpu::CommandBuffer cb[2]; |
| wgpu::CommandEncoder encoder0 = device.CreateCommandEncoder(); |
| wgpu::ComputePassEncoder pass0 = encoder0.BeginComputePass(); |
| pass0.SetPipeline(compute); |
| pass0.SetBindGroup(0, computeBindGroup); |
| pass0.Dispatch(1, 1, 1); |
| pass0.EndPass(); |
| cb[0] = encoder0.Finish(); |
| queue.Submit(1, &cb[0]); |
| |
| // Read that data in render pass. |
| wgpu::CommandEncoder encoder1 = device.CreateCommandEncoder(); |
| wgpu::RenderPassEncoder pass1 = encoder1.BeginRenderPass(&renderPass.renderPassInfo); |
| pass1.SetPipeline(render); |
| pass1.SetBindGroup(0, renderBindGroup); |
| pass1.Draw(1, 1, 0, 0); |
| pass1.EndPass(); |
| |
| cb[1] = encoder1.Finish(); |
| queue.Submit(1, &cb[1]); |
| |
| // Verify the rendering result. |
| EXPECT_PIXEL_RGBA8_EQ(RGBA8::kRed, renderPass.color, 0, 0); |
| } |
| |
| DAWN_INSTANTIATE_TEST(GpuMemorySyncTests, D3D12Backend, MetalBackend, OpenGLBackend, VulkanBackend); |
| |
| DAWN_INSTANTIATE_TEST(StorageToUniformSyncTests, |
| D3D12Backend, |
| MetalBackend, |
| OpenGLBackend, |
| VulkanBackend); |