Implement Buffer::MapAsync
MapAsync in dawn_native is fully implemented and only missing
a couple cleanups that can be done once MapRead/WriteAsync are
removed.
MapAsync in dawn_wire is left as a pure shim on top of
MapRead/WriteAsync and will be transitioned to its own commands
in follow-ups.
All MapRead/WriteAsync end2end and validation tests are duplicated
for MapAsync.
Bug: dawn:445
Change-Id: Ib1430b9257149917be19a84f13e0ddd2a8eccc32
Reviewed-on: https://dawn-review.googlesource.com/c/dawn/+/24260
Commit-Queue: Corentin Wallez <cwallez@chromium.org>
Reviewed-by: Austin Eng <enga@chromium.org>
Reviewed-by: Stephen White <senorblanco@chromium.org>
diff --git a/dawn.json b/dawn.json
index 09b76bd..d4ad9f8 100644
--- a/dawn.json
+++ b/dawn.json
@@ -198,6 +198,16 @@
]
},
{
+ "name": "map async",
+ "args": [
+ {"name": "flags", "type": "map mode"},
+ {"name": "offset", "type": "size_t"},
+ {"name": "size", "type": "size_t"},
+ {"name": "callback", "type": "buffer map callback"},
+ {"name": "userdata", "type": "void", "annotation": "*"}
+ ]
+ },
+ {
"name": "get mapped range",
"returns": "void *"
},
@@ -234,6 +244,13 @@
{"name": "mapped at creation", "type": "bool", "default": "false"}
]
},
+ "buffer map callback": {
+ "category": "callback",
+ "args": [
+ {"name": "status", "type": "buffer map async status"},
+ {"name": "userdata", "type": "void", "annotation": "*"}
+ ]
+ },
"buffer map read callback": {
"category": "callback",
"args": [
@@ -864,6 +881,14 @@
{"value": 1, "name": "load"}
]
},
+ "map mode": {
+ "category": "bitmask",
+ "values": [
+ {"value": 0, "name": "none"},
+ {"value": 1, "name": "read"},
+ {"value": 2, "name": "write"}
+ ]
+ },
"store op": {
"category": "enum",
"values": [
diff --git a/dawn_wire.json b/dawn_wire.json
index 6279f05..8bbb473 100644
--- a/dawn_wire.json
+++ b/dawn_wire.json
@@ -102,6 +102,7 @@
"SurfaceDescriptorFromXlib"
],
"client_side_commands": [
+ "BufferMapAsync",
"BufferMapReadAsync",
"BufferMapWriteAsync",
"BufferSetSubData",
diff --git a/generator/templates/mock_webgpu.cpp b/generator/templates/mock_webgpu.cpp
index 6b8c5a6..0abfe9a 100644
--- a/generator/templates/mock_webgpu.cpp
+++ b/generator/templates/mock_webgpu.cpp
@@ -96,6 +96,19 @@
OnBufferMapWriteAsyncCallback(self, callback, userdata);
}
+void ProcTableAsClass::BufferMapAsync(WGPUBuffer self,
+ WGPUMapModeFlags mode,
+ size_t offset,
+ size_t size,
+ WGPUBufferMapCallback callback,
+ void* userdata) {
+ auto object = reinterpret_cast<ProcTableAsClass::Object*>(self);
+ object->mapAsyncCallback = callback;
+ object->userdata = userdata;
+
+ OnBufferMapAsyncCallback(self, callback, userdata);
+}
+
void ProcTableAsClass::FenceOnCompletion(WGPUFence self,
uint64_t value,
WGPUFenceOnCompletionCallback callback,
@@ -135,6 +148,11 @@
object->mapWriteCallback(status, data, dataLength, object->userdata);
}
+void ProcTableAsClass::CallMapAsyncCallback(WGPUBuffer buffer, WGPUBufferMapAsyncStatus status) {
+ auto object = reinterpret_cast<ProcTableAsClass::Object*>(buffer);
+ object->mapAsyncCallback(status, object->userdata);
+}
+
void ProcTableAsClass::CallFenceOnCompletionCallback(WGPUFence fence,
WGPUFenceCompletionStatus status) {
auto object = reinterpret_cast<ProcTableAsClass::Object*>(fence);
diff --git a/generator/templates/mock_webgpu.h b/generator/templates/mock_webgpu.h
index 78b9b78..0a44248 100644
--- a/generator/templates/mock_webgpu.h
+++ b/generator/templates/mock_webgpu.h
@@ -65,6 +65,12 @@
void BufferMapWriteAsync(WGPUBuffer self,
WGPUBufferMapWriteCallback callback,
void* userdata);
+ void BufferMapAsync(WGPUBuffer self,
+ WGPUMapModeFlags mode,
+ size_t offset,
+ size_t size,
+ WGPUBufferMapCallback callback,
+ void* userdata);
void FenceOnCompletion(WGPUFence self,
uint64_t value,
WGPUFenceOnCompletionCallback callback,
@@ -86,6 +92,9 @@
virtual void OnBufferMapWriteAsyncCallback(WGPUBuffer buffer,
WGPUBufferMapWriteCallback callback,
void* userdata) = 0;
+ virtual void OnBufferMapAsyncCallback(WGPUBuffer buffer,
+ WGPUBufferMapCallback callback,
+ void* userdata) = 0;
virtual void OnFenceOnCompletionCallback(WGPUFence fence,
uint64_t value,
WGPUFenceOnCompletionCallback callback,
@@ -96,6 +105,7 @@
void CallDeviceLostCallback(WGPUDevice device, const char* message);
void CallMapReadCallback(WGPUBuffer buffer, WGPUBufferMapAsyncStatus status, const void* data, uint64_t dataLength);
void CallMapWriteCallback(WGPUBuffer buffer, WGPUBufferMapAsyncStatus status, void* data, uint64_t dataLength);
+ void CallMapAsyncCallback(WGPUBuffer buffer, WGPUBufferMapAsyncStatus status);
void CallFenceOnCompletionCallback(WGPUFence fence, WGPUFenceCompletionStatus status);
struct Object {
@@ -104,6 +114,7 @@
WGPUDeviceLostCallback deviceLostCallback = nullptr;
WGPUBufferMapReadCallback mapReadCallback = nullptr;
WGPUBufferMapWriteCallback mapWriteCallback = nullptr;
+ WGPUBufferMapCallback mapAsyncCallback = nullptr;
WGPUFenceOnCompletionCallback fenceOnCompletionCallback = nullptr;
void* userdata = 0;
};
@@ -140,6 +151,10 @@
MOCK_METHOD(bool, OnDevicePopErrorScopeCallback, (WGPUDevice device, WGPUErrorCallback callback, void* userdata), (override));
MOCK_METHOD(void, OnBufferMapReadAsyncCallback, (WGPUBuffer buffer, WGPUBufferMapReadCallback callback, void* userdata), (override));
MOCK_METHOD(void, OnBufferMapWriteAsyncCallback, (WGPUBuffer buffer, WGPUBufferMapWriteCallback callback, void* userdata), (override));
+ MOCK_METHOD(void,
+ OnBufferMapAsyncCallback,
+ (WGPUBuffer buffer, WGPUBufferMapCallback callback, void* userdata),
+ (override));
MOCK_METHOD(void, OnFenceOnCompletionCallback, (WGPUFence fence, uint64_t value, WGPUFenceOnCompletionCallback callback, void* userdata), (override));
};
diff --git a/src/dawn_native/Buffer.cpp b/src/dawn_native/Buffer.cpp
index 1db64a7..f83b590 100644
--- a/src/dawn_native/Buffer.cpp
+++ b/src/dawn_native/Buffer.cpp
@@ -72,6 +72,10 @@
UNREACHABLE();
return {};
}
+ MaybeError MapAsyncImpl(wgpu::MapMode mode, size_t offset, size_t size) override {
+ UNREACHABLE();
+ return {};
+ }
void* GetMappedPointerImpl() override {
return mFakeMappedData.get();
}
@@ -245,6 +249,22 @@
}
}
+ void BufferBase::CallMapCallback(uint32_t serial, WGPUBufferMapAsyncStatus status) {
+ ASSERT(!IsError());
+ if (mMapCallback != nullptr && serial == mMapSerial) {
+ // Tag the callback as fired before firing it, otherwise it could fire a second time if
+ // for example buffer.Unmap() is called inside the application-provided callback.
+ WGPUBufferMapCallback callback = mMapCallback;
+ mMapCallback = nullptr;
+
+ if (GetDevice()->IsLost()) {
+ callback(WGPUBufferMapAsyncStatus_DeviceLost, mMapUserdata);
+ } else {
+ callback(status, mMapUserdata);
+ }
+ }
+ }
+
void BufferBase::SetSubData(uint64_t start, uint64_t count, const void* data) {
if (count > uint64_t(std::numeric_limits<size_t>::max())) {
GetDevice()->HandleError(InternalErrorType::Validation, "count too big");
@@ -252,11 +272,14 @@
Ref<QueueBase> queue = AcquireRef(GetDevice()->GetDefaultQueue());
GetDevice()->EmitDeprecationWarning(
- "Buffer::SetSubData is deprecated, use Queue::WriteBuffer instead");
+ "Buffer::SetSubData is deprecate. Use Queue::WriteBuffer instead");
queue->WriteBuffer(this, start, data, static_cast<size_t>(count));
}
void BufferBase::MapReadAsync(WGPUBufferMapReadCallback callback, void* userdata) {
+ GetDevice()->EmitDeprecationWarning(
+ "Buffer::MapReadAsync is deprecated. Use Buffer::MapAsync instead");
+
WGPUBufferMapAsyncStatus status;
if (GetDevice()->ConsumedError(ValidateMap(wgpu::BufferUsage::MapRead, &status))) {
callback(status, nullptr, 0, userdata);
@@ -270,6 +293,7 @@
mMapSerial++;
mMapReadCallback = callback;
mMapUserdata = userdata;
+ mMapOffset = 0;
mState = BufferState::Mapped;
if (GetDevice()->ConsumedError(MapReadAsyncImpl())) {
@@ -278,10 +302,13 @@
}
MapRequestTracker* tracker = GetDevice()->GetMapRequestTracker();
- tracker->Track(this, mMapSerial, false);
+ tracker->Track(this, mMapSerial, MapType::Read);
}
void BufferBase::MapWriteAsync(WGPUBufferMapWriteCallback callback, void* userdata) {
+ GetDevice()->EmitDeprecationWarning(
+ "Buffer::MapReadAsync is deprecated. Use Buffer::MapAsync instead");
+
WGPUBufferMapAsyncStatus status;
if (GetDevice()->ConsumedError(ValidateMap(wgpu::BufferUsage::MapWrite, &status))) {
callback(status, nullptr, 0, userdata);
@@ -295,6 +322,7 @@
mMapSerial++;
mMapWriteCallback = callback;
mMapUserdata = userdata;
+ mMapOffset = 0;
mState = BufferState::Mapped;
if (GetDevice()->ConsumedError(MapWriteAsyncImpl())) {
@@ -303,7 +331,45 @@
}
MapRequestTracker* tracker = GetDevice()->GetMapRequestTracker();
- tracker->Track(this, mMapSerial, true);
+ tracker->Track(this, mMapSerial, MapType::Write);
+ }
+
+ void BufferBase::MapAsync(wgpu::MapMode mode,
+ size_t offset,
+ size_t size,
+ WGPUBufferMapCallback callback,
+ void* userdata) {
+ // Handle the defaulting of size required by WebGPU, even if in webgpu_cpp.h it is not
+ // possible to default the function argument (because there is the callback later in the
+ // argument list)
+ if (size == 0 && offset < mSize) {
+ size = mSize - offset;
+ }
+
+ WGPUBufferMapAsyncStatus status;
+ if (GetDevice()->ConsumedError(ValidateMapAsync(mode, offset, size, &status))) {
+ if (callback) {
+ callback(status, userdata);
+ }
+ return;
+ }
+ ASSERT(!IsError());
+
+ // TODO(cwallez@chromium.org): what to do on wraparound? Could cause crashes.
+ mMapSerial++;
+ mMapMode = mode;
+ mMapOffset = offset;
+ mMapCallback = callback;
+ mMapUserdata = userdata;
+ mState = BufferState::Mapped;
+
+ if (GetDevice()->ConsumedError(MapAsyncImpl(mode, offset, size))) {
+ CallMapCallback(mMapSerial, WGPUBufferMapAsyncStatus_DeviceLost);
+ return;
+ }
+
+ MapRequestTracker* tracker = GetDevice()->GetMapRequestTracker();
+ tracker->Track(this, mMapSerial, MapType::Async);
}
void* BufferBase::GetMappedRange() {
@@ -322,12 +388,12 @@
}
if (mStagingBuffer != nullptr) {
- return mStagingBuffer->GetMappedPointer();
+ return static_cast<uint8_t*>(mStagingBuffer->GetMappedPointer()) + mMapOffset;
}
if (mSize == 0) {
return reinterpret_cast<uint8_t*>(intptr_t(0xCAFED00D));
}
- return GetMappedPointerImpl();
+ return static_cast<uint8_t*>(GetMappedPointerImpl()) + mMapOffset;
}
void BufferBase::Destroy() {
@@ -389,6 +455,7 @@
// CreateBufferMapped.
CallMapReadCallback(mMapSerial, WGPUBufferMapAsyncStatus_Unknown, nullptr, 0u);
CallMapWriteCallback(mMapSerial, WGPUBufferMapAsyncStatus_Unknown, nullptr, 0u);
+ CallMapCallback(mMapSerial, WGPUBufferMapAsyncStatus_Unknown);
UnmapImpl();
mMapReadCallback = nullptr;
@@ -418,7 +485,7 @@
switch (mState) {
case BufferState::Mapped:
case BufferState::MappedAtCreation:
- return DAWN_VALIDATION_ERROR("Buffer already mapped");
+ return DAWN_VALIDATION_ERROR("Buffer is already mapped");
case BufferState::Destroyed:
return DAWN_VALIDATION_ERROR("Buffer is destroyed");
case BufferState::Unmapped:
@@ -433,6 +500,60 @@
return {};
}
+ MaybeError BufferBase::ValidateMapAsync(wgpu::MapMode mode,
+ size_t offset,
+ size_t size,
+ WGPUBufferMapAsyncStatus* status) const {
+ *status = WGPUBufferMapAsyncStatus_DeviceLost;
+ DAWN_TRY(GetDevice()->ValidateIsAlive());
+
+ *status = WGPUBufferMapAsyncStatus_Error;
+ DAWN_TRY(GetDevice()->ValidateObject(this));
+
+ if (offset % 4 != 0) {
+ return DAWN_VALIDATION_ERROR("offset must be a multiple of 4");
+ }
+
+ if (size % 4 != 0) {
+ return DAWN_VALIDATION_ERROR("size must be a multiple of 4");
+ }
+
+ if (uint64_t(offset) > mSize || uint64_t(size) > mSize - uint64_t(offset)) {
+ return DAWN_VALIDATION_ERROR("size + offset must fit in the buffer");
+ }
+
+ switch (mState) {
+ case BufferState::Mapped:
+ case BufferState::MappedAtCreation:
+ return DAWN_VALIDATION_ERROR("Buffer is already mapped");
+ case BufferState::Destroyed:
+ return DAWN_VALIDATION_ERROR("Buffer is destroyed");
+ case BufferState::Unmapped:
+ break;
+ }
+
+ bool isReadMode = mode & wgpu::MapMode::Read;
+ bool isWriteMode = mode & wgpu::MapMode::Write;
+ if (!(isReadMode ^ isWriteMode)) {
+ return DAWN_VALIDATION_ERROR("Exactly one of Read or Write mode must be set");
+ }
+
+ if (mode & wgpu::MapMode::Read) {
+ if (!(mUsage & wgpu::BufferUsage::MapRead)) {
+ return DAWN_VALIDATION_ERROR("The buffer must have the MapRead usage");
+ }
+ } else {
+ ASSERT(mode & wgpu::MapMode::Write);
+
+ if (!(mUsage & wgpu::BufferUsage::MapWrite)) {
+ return DAWN_VALIDATION_ERROR("The buffer must have the MapWrite usage");
+ }
+ }
+
+ *status = WGPUBufferMapAsyncStatus_Success;
+ return {};
+ }
+
bool BufferBase::CanGetMappedRange(bool writable) const {
// Note that:
//
@@ -448,6 +569,8 @@
return true;
case BufferState::Mapped:
+ // TODO(dawn:445): When mapRead/WriteAsync is removed, check against mMapMode
+ // instead of mUsage
ASSERT(bool(mUsage & wgpu::BufferUsage::MapRead) ^
bool(mUsage & wgpu::BufferUsage::MapWrite));
return !writable || (mUsage & wgpu::BufferUsage::MapWrite);
@@ -495,16 +618,19 @@
mState = BufferState::Destroyed;
}
- bool BufferBase::IsMapped() const {
- return mState == BufferState::Mapped;
- }
-
- void BufferBase::OnMapCommandSerialFinished(uint32_t mapSerial, bool isWrite) {
- void* data = GetMappedRangeInternal(isWrite);
- if (isWrite) {
- CallMapWriteCallback(mapSerial, WGPUBufferMapAsyncStatus_Success, data, GetSize());
- } else {
- CallMapReadCallback(mapSerial, WGPUBufferMapAsyncStatus_Success, data, GetSize());
+ void BufferBase::OnMapCommandSerialFinished(uint32_t mapSerial, MapType type) {
+ switch (type) {
+ case MapType::Read:
+ CallMapReadCallback(mapSerial, WGPUBufferMapAsyncStatus_Success,
+ GetMappedRangeInternal(false), GetSize());
+ break;
+ case MapType::Write:
+ CallMapWriteCallback(mapSerial, WGPUBufferMapAsyncStatus_Success,
+ GetMappedRangeInternal(true), GetSize());
+ break;
+ case MapType::Async:
+ CallMapCallback(mapSerial, WGPUBufferMapAsyncStatus_Success);
+ break;
}
}
diff --git a/src/dawn_native/Buffer.h b/src/dawn_native/Buffer.h
index 457e415..9772bfc 100644
--- a/src/dawn_native/Buffer.h
+++ b/src/dawn_native/Buffer.h
@@ -25,6 +25,8 @@
namespace dawn_native {
+ enum class MapType : uint32_t;
+
MaybeError ValidateBufferDescriptor(DeviceBase* device, const BufferDescriptor* descriptor);
static constexpr wgpu::BufferUsage kReadOnlyBufferUsages =
@@ -48,7 +50,7 @@
wgpu::BufferUsage GetUsage() const;
MaybeError MapAtCreation();
- void OnMapCommandSerialFinished(uint32_t mapSerial, bool isWrite);
+ void OnMapCommandSerialFinished(uint32_t mapSerial, MapType type);
MaybeError ValidateCanUseOnQueueNow() const;
@@ -60,6 +62,11 @@
void SetSubData(uint64_t start, uint64_t count, const void* data);
void MapReadAsync(WGPUBufferMapReadCallback callback, void* userdata);
void MapWriteAsync(WGPUBufferMapWriteCallback callback, void* userdata);
+ void MapAsync(wgpu::MapMode mode,
+ size_t offset,
+ size_t size,
+ WGPUBufferMapCallback callback,
+ void* userdata);
void* GetMappedRange();
const void* GetConstMappedRange();
void Unmap();
@@ -79,6 +86,7 @@
virtual MaybeError MapAtCreationImpl() = 0;
virtual MaybeError MapReadAsyncImpl() = 0;
virtual MaybeError MapWriteAsyncImpl() = 0;
+ virtual MaybeError MapAsyncImpl(wgpu::MapMode mode, size_t offset, size_t size) = 0;
virtual void UnmapImpl() = 0;
virtual void DestroyImpl() = 0;
virtual void* GetMappedPointerImpl() = 0;
@@ -94,26 +102,32 @@
WGPUBufferMapAsyncStatus status,
void* pointer,
uint64_t dataLength);
+ void CallMapCallback(uint32_t serial, WGPUBufferMapAsyncStatus status);
MaybeError ValidateMap(wgpu::BufferUsage requiredUsage,
WGPUBufferMapAsyncStatus* status) const;
+ MaybeError ValidateMapAsync(wgpu::MapMode mode,
+ size_t offset,
+ size_t size,
+ WGPUBufferMapAsyncStatus* status) const;
MaybeError ValidateUnmap() const;
MaybeError ValidateDestroy() const;
bool CanGetMappedRange(bool writable) const;
uint64_t mSize = 0;
wgpu::BufferUsage mUsage = wgpu::BufferUsage::None;
-
- WGPUBufferMapReadCallback mMapReadCallback = nullptr;
- WGPUBufferMapWriteCallback mMapWriteCallback = nullptr;
- void* mMapUserdata = 0;
- uint32_t mMapSerial = 0;
+ BufferState mState;
+ bool mIsDataInitialized = false;
std::unique_ptr<StagingBufferBase> mStagingBuffer;
- BufferState mState;
-
- bool mIsDataInitialized = false;
+ WGPUBufferMapReadCallback mMapReadCallback = nullptr;
+ WGPUBufferMapWriteCallback mMapWriteCallback = nullptr;
+ WGPUBufferMapCallback mMapCallback = nullptr;
+ void* mMapUserdata = 0;
+ uint32_t mMapSerial = 0;
+ wgpu::MapMode mMapMode = wgpu::MapMode::None;
+ size_t mMapOffset = 0;
};
} // namespace dawn_native
diff --git a/src/dawn_native/MapRequestTracker.cpp b/src/dawn_native/MapRequestTracker.cpp
index 956dfa0..91af10b 100644
--- a/src/dawn_native/MapRequestTracker.cpp
+++ b/src/dawn_native/MapRequestTracker.cpp
@@ -27,11 +27,11 @@
ASSERT(mInflightRequests.Empty());
}
- void MapRequestTracker::Track(BufferBase* buffer, uint32_t mapSerial, bool isWrite) {
+ void MapRequestTracker::Track(BufferBase* buffer, uint32_t mapSerial, MapType type) {
Request request;
request.buffer = buffer;
request.mapSerial = mapSerial;
- request.isWrite = isWrite;
+ request.type = type;
mInflightRequests.Enqueue(std::move(request), mDevice->GetPendingCommandSerial());
mDevice->AddFutureCallbackSerial(mDevice->GetPendingCommandSerial());
@@ -39,7 +39,7 @@
void MapRequestTracker::Tick(Serial finishedSerial) {
for (auto& request : mInflightRequests.IterateUpTo(finishedSerial)) {
- request.buffer->OnMapCommandSerialFinished(request.mapSerial, request.isWrite);
+ request.buffer->OnMapCommandSerialFinished(request.mapSerial, request.type);
}
mInflightRequests.ClearUpTo(finishedSerial);
}
diff --git a/src/dawn_native/MapRequestTracker.h b/src/dawn_native/MapRequestTracker.h
index 68c0c86..1daf47d 100644
--- a/src/dawn_native/MapRequestTracker.h
+++ b/src/dawn_native/MapRequestTracker.h
@@ -20,12 +20,15 @@
namespace dawn_native {
+ // TODO(dawn:22) remove this enum once MapReadAsync/MapWriteAsync are removed.
+ enum class MapType : uint32_t { Read, Write, Async };
+
class MapRequestTracker {
public:
MapRequestTracker(DeviceBase* device);
~MapRequestTracker();
- void Track(BufferBase* buffer, uint32_t mapSerial, bool isWrite);
+ void Track(BufferBase* buffer, uint32_t mapSerial, MapType type);
void Tick(Serial finishedSerial);
private:
@@ -34,7 +37,7 @@
struct Request {
Ref<BufferBase> buffer;
uint32_t mapSerial;
- bool isWrite;
+ MapType type;
};
SerialQueue<Request> mInflightRequests;
};
diff --git a/src/dawn_native/d3d12/BufferD3D12.cpp b/src/dawn_native/d3d12/BufferD3D12.cpp
index c38545b..7075661 100644
--- a/src/dawn_native/d3d12/BufferD3D12.cpp
+++ b/src/dawn_native/d3d12/BufferD3D12.cpp
@@ -248,13 +248,23 @@
return (GetUsage() & (wgpu::BufferUsage::MapRead | wgpu::BufferUsage::MapWrite)) != 0;
}
- MaybeError Buffer::MapInternal(bool isWrite, const char* contextInfo) {
+ MaybeError Buffer::MapInternal(bool isWrite,
+ size_t offset,
+ size_t size,
+ const char* contextInfo) {
// The mapped buffer can be accessed at any time, so it must be locked to ensure it is never
// evicted. This buffer should already have been made resident when it was created.
Heap* heap = ToBackend(mResourceAllocation.GetResourceHeap());
DAWN_TRY(ToBackend(GetDevice())->GetResidencyManager()->LockAllocation(heap));
- D3D12_RANGE range = {0, size_t(GetSize())};
+ D3D12_RANGE range = {offset, offset + size};
+ // mMappedData is the pointer to the start of the resource, irrespective of offset.
+ // MSDN says (note the weird use of "never"):
+ //
+ // When ppData is not NULL, the pointer returned is never offset by any values in
+ // pReadRange.
+ //
+ // https://docs.microsoft.com/en-us/windows/win32/api/d3d12/nf-d3d12-id3d12resource-map
DAWN_TRY(CheckHRESULT(GetD3D12Resource()->Map(0, &range, &mMappedData), contextInfo));
if (isWrite) {
@@ -267,16 +277,20 @@
MaybeError Buffer::MapAtCreationImpl() {
// Setting isMapWrite to false on MapRead buffers to silence D3D12 debug layer warning.
bool isMapWrite = (GetUsage() & wgpu::BufferUsage::MapWrite) != 0;
- DAWN_TRY(MapInternal(isMapWrite, "D3D12 map at creation"));
+ DAWN_TRY(MapInternal(isMapWrite, 0, size_t(GetSize()), "D3D12 map at creation"));
return {};
}
MaybeError Buffer::MapReadAsyncImpl() {
- return MapInternal(false, "D3D12 map read async");
+ return MapInternal(false, 0, size_t(GetSize()), "D3D12 map read async");
}
MaybeError Buffer::MapWriteAsyncImpl() {
- return MapInternal(true, "D3D12 map write async");
+ return MapInternal(true, 0, size_t(GetSize()), "D3D12 map write async");
+ }
+
+ MaybeError Buffer::MapAsyncImpl(wgpu::MapMode mode, size_t offset, size_t size) {
+ return MapInternal(mode & wgpu::MapMode::Write, offset, size, "D3D12 map async");
}
void Buffer::UnmapImpl() {
@@ -291,6 +305,8 @@
}
void* Buffer::GetMappedPointerImpl() {
+ // The frontend asks that the pointer returned is from the start of the resource
+ // irrespective of the offset passed in MapAsyncImpl, which is what mMappedData is.
return mMappedData;
}
@@ -366,7 +382,7 @@
// The state of the buffers on UPLOAD heap must always be GENERIC_READ and cannot be
// changed away, so we can only clear such buffer with buffer mapping.
if (D3D12HeapType(GetUsage()) == D3D12_HEAP_TYPE_UPLOAD) {
- DAWN_TRY(MapInternal(true, "D3D12 map at clear buffer"));
+ DAWN_TRY(MapInternal(true, 0, size_t(GetSize()), "D3D12 map at clear buffer"));
memset(mMappedData, clearValue, GetSize());
UnmapImpl();
} else {
diff --git a/src/dawn_native/d3d12/BufferD3D12.h b/src/dawn_native/d3d12/BufferD3D12.h
index 33804d7..ea1cde1 100644
--- a/src/dawn_native/d3d12/BufferD3D12.h
+++ b/src/dawn_native/d3d12/BufferD3D12.h
@@ -54,13 +54,14 @@
// Dawn API
MaybeError MapReadAsyncImpl() override;
MaybeError MapWriteAsyncImpl() override;
+ MaybeError MapAsyncImpl(wgpu::MapMode mode, size_t offset, size_t size) override;
void UnmapImpl() override;
void DestroyImpl() override;
bool IsMappableAtCreation() const override;
virtual MaybeError MapAtCreationImpl() override;
void* GetMappedPointerImpl() override;
- MaybeError MapInternal(bool isWrite, const char* contextInfo);
+ MaybeError MapInternal(bool isWrite, size_t start, size_t end, const char* contextInfo);
bool TransitionUsageAndGetResourceBarrier(CommandRecordingContext* commandContext,
D3D12_RESOURCE_BARRIER* barrier,
diff --git a/src/dawn_native/metal/BufferMTL.h b/src/dawn_native/metal/BufferMTL.h
index f26c7e0..76c9e4c 100644
--- a/src/dawn_native/metal/BufferMTL.h
+++ b/src/dawn_native/metal/BufferMTL.h
@@ -43,6 +43,7 @@
// Dawn API
MaybeError MapReadAsyncImpl() override;
MaybeError MapWriteAsyncImpl() override;
+ MaybeError MapAsyncImpl(wgpu::MapMode mode, size_t offset, size_t size) override;
void UnmapImpl() override;
void DestroyImpl() override;
void* GetMappedPointerImpl() override;
diff --git a/src/dawn_native/metal/BufferMTL.mm b/src/dawn_native/metal/BufferMTL.mm
index d231e86..a356cd9 100644
--- a/src/dawn_native/metal/BufferMTL.mm
+++ b/src/dawn_native/metal/BufferMTL.mm
@@ -121,6 +121,10 @@
return {};
}
+ MaybeError Buffer::MapAsyncImpl(wgpu::MapMode mode, size_t offset, size_t size) {
+ return {};
+ }
+
void* Buffer::GetMappedPointerImpl() {
return [mMtlBuffer contents];
}
diff --git a/src/dawn_native/null/DeviceNull.cpp b/src/dawn_native/null/DeviceNull.cpp
index ab430c3..9b04aee 100644
--- a/src/dawn_native/null/DeviceNull.cpp
+++ b/src/dawn_native/null/DeviceNull.cpp
@@ -312,6 +312,10 @@
return {};
}
+ MaybeError Buffer::MapAsyncImpl(wgpu::MapMode mode, size_t offset, size_t size) {
+ return {};
+ }
+
void* Buffer::GetMappedPointerImpl() {
return mBackingData.get();
}
diff --git a/src/dawn_native/null/DeviceNull.h b/src/dawn_native/null/DeviceNull.h
index 94d3826..9f39cba 100644
--- a/src/dawn_native/null/DeviceNull.h
+++ b/src/dawn_native/null/DeviceNull.h
@@ -201,6 +201,7 @@
// Dawn API
MaybeError MapReadAsyncImpl() override;
MaybeError MapWriteAsyncImpl() override;
+ MaybeError MapAsyncImpl(wgpu::MapMode mode, size_t offset, size_t size) override;
void UnmapImpl() override;
void DestroyImpl() override;
diff --git a/src/dawn_native/opengl/BufferGL.cpp b/src/dawn_native/opengl/BufferGL.cpp
index e7398d6..56fe397 100644
--- a/src/dawn_native/opengl/BufferGL.cpp
+++ b/src/dawn_native/opengl/BufferGL.cpp
@@ -125,7 +125,37 @@
return {};
}
+ MaybeError Buffer::MapAsyncImpl(wgpu::MapMode mode, size_t offset, size_t size) {
+ const OpenGLFunctions& gl = ToBackend(GetDevice())->gl;
+
+ // It is an error to map an empty range in OpenGL. We always have at least a 4-byte buffer
+ // so we extend the range to be 4 bytes.
+ if (size == 0) {
+ if (offset != 0) {
+ offset -= 4;
+ }
+ size = 4;
+ }
+
+ // TODO(cwallez@chromium.org): this does GPU->CPU synchronization, we could require a high
+ // version of OpenGL that would let us map the buffer unsynchronized.
+ gl.BindBuffer(GL_ARRAY_BUFFER, mBuffer);
+ void* mappedData = nullptr;
+ if (mode & wgpu::MapMode::Read) {
+ mappedData = gl.MapBufferRange(GL_ARRAY_BUFFER, offset, size, GL_MAP_READ_BIT);
+ } else {
+ ASSERT(mode & wgpu::MapMode::Write);
+ mappedData = gl.MapBufferRange(GL_ARRAY_BUFFER, offset, size, GL_MAP_WRITE_BIT);
+ }
+
+ // The frontend asks that the pointer returned by GetMappedPointerImpl is from the start of
+ // the resource but OpenGL gives us the pointer at offset. Remove the offset.
+ mMappedData = static_cast<uint8_t*>(mappedData) - offset;
+ return {};
+ }
+
void* Buffer::GetMappedPointerImpl() {
+ // The mapping offset has already been removed.
return mMappedData;
}
diff --git a/src/dawn_native/opengl/BufferGL.h b/src/dawn_native/opengl/BufferGL.h
index 8dbe764..c1ba068 100644
--- a/src/dawn_native/opengl/BufferGL.h
+++ b/src/dawn_native/opengl/BufferGL.h
@@ -37,6 +37,7 @@
// Dawn API
MaybeError MapReadAsyncImpl() override;
MaybeError MapWriteAsyncImpl() override;
+ MaybeError MapAsyncImpl(wgpu::MapMode mode, size_t offset, size_t size) override;
void UnmapImpl() override;
void DestroyImpl() override;
diff --git a/src/dawn_native/vulkan/BufferVk.cpp b/src/dawn_native/vulkan/BufferVk.cpp
index a3dcf2e..613b589 100644
--- a/src/dawn_native/vulkan/BufferVk.cpp
+++ b/src/dawn_native/vulkan/BufferVk.cpp
@@ -259,6 +259,19 @@
return {};
}
+ MaybeError Buffer::MapAsyncImpl(wgpu::MapMode mode, size_t offset, size_t size) {
+ Device* device = ToBackend(GetDevice());
+
+ CommandRecordingContext* recordingContext = device->GetPendingRecordingContext();
+ if (mode & wgpu::MapMode::Read) {
+ TransitionUsageNow(recordingContext, wgpu::BufferUsage::MapRead);
+ } else {
+ ASSERT(mode & wgpu::MapMode::Write);
+ TransitionUsageNow(recordingContext, wgpu::BufferUsage::MapWrite);
+ }
+ return {};
+ }
+
void Buffer::UnmapImpl() {
// No need to do anything, we keep CPU-visible memory mapped at all time.
}
diff --git a/src/dawn_native/vulkan/BufferVk.h b/src/dawn_native/vulkan/BufferVk.h
index d0d3ea2..062ff25 100644
--- a/src/dawn_native/vulkan/BufferVk.h
+++ b/src/dawn_native/vulkan/BufferVk.h
@@ -58,6 +58,7 @@
// Dawn API
MaybeError MapReadAsyncImpl() override;
MaybeError MapWriteAsyncImpl() override;
+ MaybeError MapAsyncImpl(wgpu::MapMode mode, size_t offset, size_t size) override;
void UnmapImpl() override;
void DestroyImpl() override;
diff --git a/src/dawn_wire/client/Buffer.cpp b/src/dawn_wire/client/Buffer.cpp
index eeb5e49..7220181 100644
--- a/src/dawn_wire/client/Buffer.cpp
+++ b/src/dawn_wire/client/Buffer.cpp
@@ -239,6 +239,72 @@
SerializeBufferMapAsync(this, serial, writeHandle);
}
+ void Buffer::MapAsync(WGPUMapModeFlags mode,
+ size_t offset,
+ size_t size,
+ WGPUBufferMapCallback callback,
+ void* userdata) {
+ // Do early validation for mode because it needs to be correct for the proxying to
+ // MapReadAsync or MapWriteAsync to work.
+ bool isReadMode = mode & WGPUMapMode_Read;
+ bool isWriteMode = mode & WGPUMapMode_Write;
+ bool modeOk = isReadMode ^ isWriteMode;
+ // Do early validation of offset and size because it isn't checked by MapReadAsync /
+ // MapWriteAsync.
+ bool offsetOk = (uint64_t(offset) <= mSize) && offset % 4 == 0;
+ bool sizeOk = (uint64_t(size) <= mSize - uint64_t(offset)) && size % 4 == 0;
+
+ if (!(modeOk && offsetOk && sizeOk)) {
+ device->InjectError(WGPUErrorType_Validation, "MapAsync error (you figure out :P)");
+ if (callback != nullptr) {
+ callback(WGPUBufferMapAsyncStatus_Error, userdata);
+ }
+ return;
+ }
+
+ // The structure to keep arguments so we can forward the MapReadAsync and MapWriteAsync to
+ // `callback`
+ struct ProxyData {
+ WGPUBufferMapCallback callback;
+ void* userdata;
+ size_t mapOffset;
+ Buffer* self;
+ };
+ ProxyData* proxy = new ProxyData;
+ proxy->callback = callback;
+ proxy->userdata = userdata;
+ proxy->mapOffset = offset;
+ proxy->self = this;
+ // Note technically we should keep the buffer alive until the callback is fired but the
+ // client doesn't have good facilities to do that yet.
+
+ // Call MapReadAsync or MapWriteAsync and forward the callback.
+ if (mode & WGPUMapMode_Read) {
+ MapReadAsync(
+ [](WGPUBufferMapAsyncStatus status, const void*, uint64_t, void* userdata) {
+ ProxyData* proxy = static_cast<ProxyData*>(userdata);
+ if (proxy->callback) {
+ proxy->callback(status, proxy->userdata);
+ }
+ proxy->self->mMapOffset = proxy->mapOffset;
+ delete proxy;
+ },
+ proxy);
+ } else {
+ ASSERT(mode & WGPUMapMode_Write);
+ MapWriteAsync(
+ [](WGPUBufferMapAsyncStatus status, void*, uint64_t, void* userdata) {
+ ProxyData* proxy = static_cast<ProxyData*>(userdata);
+ if (proxy->callback) {
+ proxy->callback(status, proxy->userdata);
+ }
+ proxy->self->mMapOffset = proxy->mapOffset;
+ delete proxy;
+ },
+ proxy);
+ }
+ }
+
bool Buffer::OnMapReadAsyncCallback(uint32_t requestSerial,
uint32_t status,
uint64_t initialDataInfoLength,
@@ -369,14 +435,14 @@
if (!IsMappedForWriting()) {
return nullptr;
}
- return mMappedData;
+ return static_cast<uint8_t*>(mMappedData) + mMapOffset;
}
const void* Buffer::GetConstMappedRange() {
if (!IsMappedForWriting() && !IsMappedForReading()) {
return nullptr;
}
- return mMappedData;
+ return static_cast<uint8_t*>(mMappedData) + mMapOffset;
}
void Buffer::Unmap() {
@@ -413,6 +479,7 @@
mReadHandle = nullptr;
}
mMappedData = nullptr;
+ mMapOffset = 0;
ClearMapRequests(WGPUBufferMapAsyncStatus_Unknown);
BufferUnmapCmd cmd;
diff --git a/src/dawn_wire/client/Buffer.h b/src/dawn_wire/client/Buffer.h
index a69b054..b7ec545c 100644
--- a/src/dawn_wire/client/Buffer.h
+++ b/src/dawn_wire/client/Buffer.h
@@ -43,6 +43,11 @@
uint64_t initialDataInfoLength,
const uint8_t* initialDataInfo);
bool OnMapWriteAsyncCallback(uint32_t requestSerial, uint32_t status);
+ void MapAsync(WGPUMapModeFlags mode,
+ size_t offset,
+ size_t size,
+ WGPUBufferMapCallback callback,
+ void* userdata);
void* GetMappedRange();
const void* GetConstMappedRange();
void Unmap();
@@ -77,6 +82,7 @@
std::unique_ptr<MemoryTransferService::ReadHandle> mReadHandle = nullptr;
std::unique_ptr<MemoryTransferService::WriteHandle> mWriteHandle = nullptr;
void* mMappedData = nullptr;
+ size_t mMapOffset = 0;
};
}} // namespace dawn_wire::client
diff --git a/src/tests/end2end/BufferTests.cpp b/src/tests/end2end/BufferTests.cpp
index 4b626ee..eacd533 100644
--- a/src/tests/end2end/BufferTests.cpp
+++ b/src/tests/end2end/BufferTests.cpp
@@ -317,6 +317,216 @@
OpenGLBackend(),
VulkanBackend());
+class BufferMappingTests : public DawnTest {
+ protected:
+ void MapAsyncAndWait(const wgpu::Buffer& buffer,
+ wgpu::MapMode mode,
+ size_t offset,
+ size_t size) {
+ bool done = false;
+ buffer.MapAsync(
+ mode, offset, size,
+ [](WGPUBufferMapAsyncStatus status, void* userdata) {
+ ASSERT_EQ(WGPUBufferMapAsyncStatus_Success, status);
+ *static_cast<bool*>(userdata) = true;
+ },
+ &done);
+
+ while (!done) {
+ WaitABit();
+ }
+ }
+
+ wgpu::Buffer CreateMapReadBuffer(uint64_t size) {
+ wgpu::BufferDescriptor descriptor;
+ descriptor.size = size;
+ descriptor.usage = wgpu::BufferUsage::MapRead | wgpu::BufferUsage::CopyDst;
+ return device.CreateBuffer(&descriptor);
+ }
+
+ wgpu::Buffer CreateMapWriteBuffer(uint64_t size) {
+ wgpu::BufferDescriptor descriptor;
+ descriptor.size = size;
+ descriptor.usage = wgpu::BufferUsage::MapWrite | wgpu::BufferUsage::CopySrc;
+ return device.CreateBuffer(&descriptor);
+ }
+};
+
+// Test that the simplest map read works
+TEST_P(BufferMappingTests, MapRead_Basic) {
+ wgpu::Buffer buffer = CreateMapReadBuffer(4);
+
+ uint32_t myData = 0x01020304;
+ queue.WriteBuffer(buffer, 0, &myData, sizeof(myData));
+
+ MapAsyncAndWait(buffer, wgpu::MapMode::Read, 0, 4);
+ ASSERT_NE(nullptr, buffer.GetConstMappedRange());
+ ASSERT_EQ(myData, *static_cast<const uint32_t*>(buffer.GetConstMappedRange()));
+ buffer.Unmap();
+}
+
+// Test map-reading a zero-sized buffer.
+TEST_P(BufferMappingTests, MapRead_ZeroSized) {
+ wgpu::Buffer buffer = CreateMapReadBuffer(0);
+
+ MapAsyncAndWait(buffer, wgpu::MapMode::Read, 0, 0);
+ ASSERT_NE(buffer.GetConstMappedRange(), nullptr);
+ buffer.Unmap();
+}
+
+// Test map-reading with a non-zero offset
+TEST_P(BufferMappingTests, MapRead_NonZeroOffset) {
+ wgpu::Buffer buffer = CreateMapReadBuffer(8);
+
+ uint32_t myData[2] = {0x01020304, 0x05060708};
+ queue.WriteBuffer(buffer, 0, &myData, sizeof(myData));
+
+ MapAsyncAndWait(buffer, wgpu::MapMode::Read, 4, 4);
+ ASSERT_EQ(myData[1], *static_cast<const uint32_t*>(buffer.GetConstMappedRange()));
+ buffer.Unmap();
+}
+
+// Map read and unmap twice. Test that both of these two iterations work.
+TEST_P(BufferMappingTests, MapRead_Twice) {
+ wgpu::Buffer buffer = CreateMapReadBuffer(4);
+
+ uint32_t myData = 0x01020304;
+ queue.WriteBuffer(buffer, 0, &myData, sizeof(myData));
+
+ MapAsyncAndWait(buffer, wgpu::MapMode::Read, 0, 4);
+ ASSERT_EQ(myData, *static_cast<const uint32_t*>(buffer.GetConstMappedRange()));
+ buffer.Unmap();
+
+ myData = 0x05060708;
+ queue.WriteBuffer(buffer, 0, &myData, sizeof(myData));
+
+ MapAsyncAndWait(buffer, wgpu::MapMode::Read, 0, 4);
+ ASSERT_EQ(myData, *static_cast<const uint32_t*>(buffer.GetConstMappedRange()));
+ buffer.Unmap();
+}
+
+// Test map-reading a large buffer.
+TEST_P(BufferMappingTests, MapRead_Large) {
+ constexpr uint32_t kDataSize = 1000 * 1000;
+ wgpu::Buffer buffer = CreateMapReadBuffer(kDataSize * sizeof(uint32_t));
+
+ std::vector<uint32_t> myData;
+ for (uint32_t i = 0; i < kDataSize; ++i) {
+ myData.push_back(i);
+ }
+ queue.WriteBuffer(buffer, 0, myData.data(), kDataSize * sizeof(uint32_t));
+
+ MapAsyncAndWait(buffer, wgpu::MapMode::Read, 0, 4);
+ ASSERT_EQ(0, memcmp(buffer.GetConstMappedRange(), myData.data(), kDataSize * sizeof(uint32_t)));
+ buffer.Unmap();
+}
+
+// Test that the simplest map write works.
+TEST_P(BufferMappingTests, MapWrite_Basic) {
+ wgpu::Buffer buffer = CreateMapWriteBuffer(4);
+
+ uint32_t myData = 2934875;
+ MapAsyncAndWait(buffer, wgpu::MapMode::Write, 0, 4);
+ ASSERT_NE(nullptr, buffer.GetMappedRange());
+ ASSERT_NE(nullptr, buffer.GetConstMappedRange());
+ memcpy(buffer.GetMappedRange(), &myData, sizeof(myData));
+ buffer.Unmap();
+
+ EXPECT_BUFFER_U32_EQ(myData, buffer, 0);
+}
+
+// Test map-writing a zero-sized buffer.
+TEST_P(BufferMappingTests, MapWrite_ZeroSized) {
+ wgpu::Buffer buffer = CreateMapWriteBuffer(0);
+
+ MapAsyncAndWait(buffer, wgpu::MapMode::Write, 0, 0);
+ ASSERT_NE(buffer.GetConstMappedRange(), nullptr);
+ ASSERT_NE(buffer.GetMappedRange(), nullptr);
+ buffer.Unmap();
+}
+
+// Test map-writing with a non-zero offset.
+TEST_P(BufferMappingTests, MapWrite_NonZeroOffset) {
+ wgpu::Buffer buffer = CreateMapWriteBuffer(8);
+
+ uint32_t myData = 2934875;
+ MapAsyncAndWait(buffer, wgpu::MapMode::Write, 4, 4);
+ memcpy(buffer.GetMappedRange(), &myData, sizeof(myData));
+ buffer.Unmap();
+
+ EXPECT_BUFFER_U32_EQ(myData, buffer, 4);
+}
+
+// Map, write and unmap twice. Test that both of these two iterations work.
+TEST_P(BufferMappingTests, MapWrite_Twice) {
+ wgpu::Buffer buffer = CreateMapWriteBuffer(4);
+
+ uint32_t myData = 2934875;
+ MapAsyncAndWait(buffer, wgpu::MapMode::Write, 0, 4);
+ memcpy(buffer.GetMappedRange(), &myData, sizeof(myData));
+ buffer.Unmap();
+
+ EXPECT_BUFFER_U32_EQ(myData, buffer, 0);
+
+ myData = 9999999;
+ MapAsyncAndWait(buffer, wgpu::MapMode::Write, 0, 4);
+ memcpy(buffer.GetMappedRange(), &myData, sizeof(myData));
+ buffer.Unmap();
+
+ EXPECT_BUFFER_U32_EQ(myData, buffer, 0);
+}
+
+// Test mapping a large buffer.
+TEST_P(BufferMappingTests, MapWrite_Large) {
+ constexpr uint32_t kDataSize = 1000 * 1000;
+ wgpu::Buffer buffer = CreateMapWriteBuffer(kDataSize * sizeof(uint32_t));
+
+ std::vector<uint32_t> myData;
+ for (uint32_t i = 0; i < kDataSize; ++i) {
+ myData.push_back(i);
+ }
+
+ MapAsyncAndWait(buffer, wgpu::MapMode::Write, 0, 4);
+ memcpy(buffer.GetMappedRange(), myData.data(), kDataSize * sizeof(uint32_t));
+ buffer.Unmap();
+
+ EXPECT_BUFFER_U32_RANGE_EQ(myData.data(), buffer, 0, kDataSize);
+}
+
+// Test that the map offset isn't updated when the call is an error.
+TEST_P(BufferMappingTests, OffsetNotUpdatedOnError) {
+ uint32_t data[3] = {0xCA7, 0xB0A7, 0xBA7};
+ wgpu::Buffer buffer = CreateMapReadBuffer(sizeof(data));
+ queue.WriteBuffer(buffer, 0, data, sizeof(data));
+
+ // Map the buffer but do not wait on the result yet.
+ bool done = false;
+ buffer.MapAsync(
+ wgpu::MapMode::Read, 4, 4,
+ [](WGPUBufferMapAsyncStatus status, void* userdata) {
+ ASSERT_EQ(WGPUBufferMapAsyncStatus_Success, status);
+ *static_cast<bool*>(userdata) = true;
+ },
+ &done);
+
+ // Call MapAsync another time, it is an error because the buffer is already being mapped so
+ // mMapOffset is not updated.
+ ASSERT_DEVICE_ERROR(buffer.MapAsync(wgpu::MapMode::Read, 8, 4, nullptr, nullptr));
+
+ while (!done) {
+ WaitABit();
+ }
+
+ // mMapOffset has not been updated so it should still be 4, which is data[1]
+ ASSERT_EQ(0, memcmp(buffer.GetConstMappedRange(), &data[1], sizeof(uint32_t)));
+}
+
+DAWN_INSTANTIATE_TEST(BufferMappingTests,
+ D3D12Backend(),
+ MetalBackend(),
+ OpenGLBackend(),
+ VulkanBackend());
+
class CreateBufferMappedTests : public DawnTest {
protected:
static void MapReadCallback(WGPUBufferMapAsyncStatus status,
@@ -1058,6 +1268,42 @@
RunTest(descriptor);
}
+// Test that mapping an OOM buffer fails gracefully
+TEST_P(BufferTests, CreateBufferOOMMapAsync) {
+ // TODO(http://crbug.com/dawn/27): Missing support.
+ DAWN_SKIP_TEST_IF(IsOpenGL());
+ DAWN_SKIP_TEST_IF(IsAsan());
+
+ auto RunTest = [this](const wgpu::BufferDescriptor& descriptor) {
+ wgpu::Buffer buffer;
+ ASSERT_DEVICE_ERROR(buffer = device.CreateBuffer(&descriptor));
+
+ bool done = false;
+ ASSERT_DEVICE_ERROR(buffer.MapAsync(
+ wgpu::MapMode::Write, 0, 4,
+ [](WGPUBufferMapAsyncStatus status, void* userdata) {
+ EXPECT_EQ(status, WGPUBufferMapAsyncStatus_Error);
+ *static_cast<bool*>(userdata) = true;
+ },
+ &done));
+
+ while (!done) {
+ WaitABit();
+ }
+ };
+
+ wgpu::BufferDescriptor descriptor;
+ descriptor.usage = wgpu::BufferUsage::CopySrc | wgpu::BufferUsage::MapWrite;
+
+ // Test an enormous buffer
+ descriptor.size = std::numeric_limits<uint64_t>::max();
+ RunTest(descriptor);
+
+ // UINT64_MAX may be special cased. Test a smaller, but really large buffer also fails
+ descriptor.size = 1ull << 50;
+ RunTest(descriptor);
+}
+
DAWN_INSTANTIATE_TEST(BufferTests,
D3D12Backend(),
MetalBackend(),
diff --git a/src/tests/unittests/validation/BufferValidationTests.cpp b/src/tests/unittests/validation/BufferValidationTests.cpp
index a724057..c31412e 100644
--- a/src/tests/unittests/validation/BufferValidationTests.cpp
+++ b/src/tests/unittests/validation/BufferValidationTests.cpp
@@ -58,6 +58,16 @@
userdata);
}
+class MockBufferMapAsyncCallback {
+ public:
+ MOCK_METHOD(void, Call, (WGPUBufferMapAsyncStatus status, void* userdata));
+};
+
+static std::unique_ptr<MockBufferMapAsyncCallback> mockBufferMapAsyncCallback;
+static void ToMockBufferMapAsyncCallback(WGPUBufferMapAsyncStatus status, void* userdata) {
+ mockBufferMapAsyncCallback->Call(status, userdata);
+}
+
class BufferValidationTest : public ValidationTest {
protected:
wgpu::Buffer CreateMapReadBuffer(uint64_t size) {
@@ -67,6 +77,7 @@
return device.CreateBuffer(&descriptor);
}
+
wgpu::Buffer CreateMapWriteBuffer(uint64_t size) {
wgpu::BufferDescriptor descriptor;
descriptor.size = size;
@@ -92,6 +103,13 @@
return device.CreateBuffer(&descriptor);
}
+ void AssertMapAsyncError(wgpu::Buffer buffer, wgpu::MapMode mode, size_t offset, size_t size) {
+ EXPECT_CALL(*mockBufferMapAsyncCallback, Call(WGPUBufferMapAsyncStatus_Error, _)).Times(1);
+
+ ASSERT_DEVICE_ERROR(
+ buffer.MapAsync(mode, offset, size, ToMockBufferMapAsyncCallback, nullptr));
+ }
+
wgpu::Queue queue;
private:
@@ -100,6 +118,7 @@
mockBufferMapReadCallback = std::make_unique<MockBufferMapReadCallback>();
mockBufferMapWriteCallback = std::make_unique<MockBufferMapWriteCallback>();
+ mockBufferMapAsyncCallback = std::make_unique<MockBufferMapAsyncCallback>();
queue = device.GetDefaultQueue();
}
@@ -107,6 +126,7 @@
// Delete mocks so that expectations are checked
mockBufferMapReadCallback = nullptr;
mockBufferMapWriteCallback = nullptr;
+ mockBufferMapAsyncCallback = nullptr;
ValidationTest::TearDown();
}
@@ -164,7 +184,272 @@
}
// Test the success case for mapping buffer for reading
-TEST_F(BufferValidationTest, MapReadSuccess) {
+TEST_F(BufferValidationTest, MapAsync_ReadSuccess) {
+ wgpu::Buffer buf = CreateMapReadBuffer(4);
+
+ buf.MapAsync(wgpu::MapMode::Read, 0, 4, ToMockBufferMapAsyncCallback, nullptr);
+
+ EXPECT_CALL(*mockBufferMapAsyncCallback, Call(WGPUBufferMapAsyncStatus_Success, _)).Times(1);
+ WaitForAllOperations(device);
+
+ buf.Unmap();
+}
+
+// Test the success case for mapping buffer for writing
+TEST_F(BufferValidationTest, MapAsync_WriteSuccess) {
+ wgpu::Buffer buf = CreateMapWriteBuffer(4);
+
+ buf.MapAsync(wgpu::MapMode::Write, 0, 4, ToMockBufferMapAsyncCallback, nullptr);
+
+ EXPECT_CALL(*mockBufferMapAsyncCallback, Call(WGPUBufferMapAsyncStatus_Success, _)).Times(1);
+ WaitForAllOperations(device);
+
+ buf.Unmap();
+}
+
+// Test map async with a buffer that's an error
+TEST_F(BufferValidationTest, MapAsync_ErrorBuffer) {
+ wgpu::BufferDescriptor desc;
+ desc.size = 4;
+ desc.usage = wgpu::BufferUsage::MapRead | wgpu::BufferUsage::MapWrite;
+ wgpu::Buffer buffer;
+ ASSERT_DEVICE_ERROR(buffer = device.CreateBuffer(&desc));
+
+ AssertMapAsyncError(buffer, wgpu::MapMode::Read, 0, 4);
+ AssertMapAsyncError(buffer, wgpu::MapMode::Write, 0, 4);
+}
+
+// Test map async with an invalid offset and size alignment.
+TEST_F(BufferValidationTest, MapAsync_OffsetSizeAlignment) {
+ // Control case, both aligned to 4 is ok.
+ {
+ wgpu::Buffer buffer = CreateMapReadBuffer(8);
+ buffer.MapAsync(wgpu::MapMode::Read, 4, 4, nullptr, nullptr);
+ }
+ {
+ wgpu::Buffer buffer = CreateMapWriteBuffer(8);
+ buffer.MapAsync(wgpu::MapMode::Write, 4, 4, nullptr, nullptr);
+ }
+
+ // Error case, offset aligned to 2 is an error.
+ {
+ wgpu::Buffer buffer = CreateMapReadBuffer(8);
+ AssertMapAsyncError(buffer, wgpu::MapMode::Read, 2, 4);
+ }
+ {
+ wgpu::Buffer buffer = CreateMapWriteBuffer(8);
+ AssertMapAsyncError(buffer, wgpu::MapMode::Write, 2, 4);
+ }
+
+ // Error case, size aligned to 2 is an error.
+ {
+ wgpu::Buffer buffer = CreateMapReadBuffer(8);
+ AssertMapAsyncError(buffer, wgpu::MapMode::Read, 0, 6);
+ }
+ {
+ wgpu::Buffer buffer = CreateMapWriteBuffer(8);
+ AssertMapAsyncError(buffer, wgpu::MapMode::Write, 0, 6);
+ }
+}
+
+// Test map async with a buffer that has the wrong usage
+TEST_F(BufferValidationTest, MapAsync_WrongUsage) {
+ {
+ wgpu::BufferDescriptor desc;
+ desc.usage = wgpu::BufferUsage::Vertex;
+ desc.size = 4;
+ wgpu::Buffer buffer = device.CreateBuffer(&desc);
+
+ AssertMapAsyncError(buffer, wgpu::MapMode::Read, 0, 4);
+ AssertMapAsyncError(buffer, wgpu::MapMode::Write, 0, 4);
+ }
+ {
+ wgpu::Buffer buffer = CreateMapReadBuffer(4);
+ AssertMapAsyncError(buffer, wgpu::MapMode::Write, 0, 4);
+ }
+ {
+ wgpu::Buffer buffer = CreateMapWriteBuffer(4);
+ AssertMapAsyncError(buffer, wgpu::MapMode::Read, 0, 4);
+ }
+}
+
+// Test map async with a wrong mode
+TEST_F(BufferValidationTest, MapAsync_WrongMode) {
+ {
+ wgpu::Buffer buffer = CreateMapReadBuffer(4);
+ AssertMapAsyncError(buffer, wgpu::MapMode::None, 0, 4);
+ }
+ {
+ wgpu::Buffer buffer = CreateMapReadBuffer(4);
+ AssertMapAsyncError(buffer, wgpu::MapMode::Read | wgpu::MapMode::Write, 0, 4);
+ }
+}
+
+// Test map async with a buffer that's already mapped
+TEST_F(BufferValidationTest, MapAsync_AlreadyMapped) {
+ {
+ wgpu::Buffer buffer = CreateMapReadBuffer(4);
+ buffer.MapAsync(wgpu::MapMode::Read, 0, 4, nullptr, nullptr);
+ AssertMapAsyncError(buffer, wgpu::MapMode::Read, 0, 4);
+ }
+ {
+ wgpu::Buffer buffer = BufferMappedAtCreation(4, wgpu::BufferUsage::MapRead);
+ AssertMapAsyncError(buffer, wgpu::MapMode::Read, 0, 4);
+ }
+ {
+ wgpu::Buffer buffer = CreateMapWriteBuffer(4);
+ buffer.MapAsync(wgpu::MapMode::Write, 0, 4, nullptr, nullptr);
+ AssertMapAsyncError(buffer, wgpu::MapMode::Write, 0, 4);
+ }
+ {
+ wgpu::Buffer buffer = BufferMappedAtCreation(4, wgpu::BufferUsage::MapWrite);
+ AssertMapAsyncError(buffer, wgpu::MapMode::Write, 0, 4);
+ }
+}
+
+// Test map async with a buffer that's destroyed
+TEST_F(BufferValidationTest, MapAsync_Destroy) {
+ {
+ wgpu::Buffer buffer = CreateMapReadBuffer(4);
+ buffer.Destroy();
+ AssertMapAsyncError(buffer, wgpu::MapMode::Read, 0, 4);
+ }
+ {
+ wgpu::Buffer buffer = CreateMapWriteBuffer(4);
+ buffer.Destroy();
+ AssertMapAsyncError(buffer, wgpu::MapMode::Write, 0, 4);
+ }
+}
+
+// Test map async but unmapping before the result is ready.
+TEST_F(BufferValidationTest, MapAsync_UnmapBeforeResult) {
+ {
+ wgpu::Buffer buf = CreateMapReadBuffer(4);
+ buf.MapAsync(wgpu::MapMode::Read, 0, 4, ToMockBufferMapAsyncCallback, nullptr);
+
+ EXPECT_CALL(*mockBufferMapAsyncCallback, Call(WGPUBufferMapAsyncStatus_Unknown, _))
+ .Times(1);
+ buf.Unmap();
+
+ // The callback shouldn't be called again.
+ WaitForAllOperations(device);
+ }
+ {
+ wgpu::Buffer buf = CreateMapWriteBuffer(4);
+ buf.MapAsync(wgpu::MapMode::Write, 0, 4, ToMockBufferMapAsyncCallback, nullptr);
+
+ EXPECT_CALL(*mockBufferMapAsyncCallback, Call(WGPUBufferMapAsyncStatus_Unknown, _))
+ .Times(1);
+ buf.Unmap();
+
+ // The callback shouldn't be called again.
+ WaitForAllOperations(device);
+ }
+}
+
+// When a MapAsync is cancelled with Unmap it might still be in flight, test doing a new request
+// works as expected and we don't get the cancelled request's data.
+TEST_F(BufferValidationTest, MapAsync_UnmapBeforeResultAndMapAgain) {
+ {
+ wgpu::Buffer buf = CreateMapReadBuffer(4);
+ buf.MapAsync(wgpu::MapMode::Read, 0, 4, ToMockBufferMapAsyncCallback, this + 0);
+
+ EXPECT_CALL(*mockBufferMapAsyncCallback, Call(WGPUBufferMapAsyncStatus_Unknown, this + 0))
+ .Times(1);
+ buf.Unmap();
+
+ buf.MapAsync(wgpu::MapMode::Read, 0, 4, ToMockBufferMapAsyncCallback, this + 1);
+ EXPECT_CALL(*mockBufferMapAsyncCallback, Call(WGPUBufferMapAsyncStatus_Success, this + 1))
+ .Times(1);
+ WaitForAllOperations(device);
+ }
+ {
+ wgpu::Buffer buf = CreateMapWriteBuffer(4);
+ buf.MapAsync(wgpu::MapMode::Write, 0, 4, ToMockBufferMapAsyncCallback, this + 0);
+
+ EXPECT_CALL(*mockBufferMapAsyncCallback, Call(WGPUBufferMapAsyncStatus_Unknown, this + 0))
+ .Times(1);
+ buf.Unmap();
+
+ buf.MapAsync(wgpu::MapMode::Write, 0, 4, ToMockBufferMapAsyncCallback, this + 1);
+ EXPECT_CALL(*mockBufferMapAsyncCallback, Call(WGPUBufferMapAsyncStatus_Success, this + 1))
+ .Times(1);
+ WaitForAllOperations(device);
+ }
+}
+
+// Test map async but destroying before the result is ready.
+TEST_F(BufferValidationTest, MapAsync_DestroyBeforeResult) {
+ {
+ wgpu::Buffer buf = CreateMapReadBuffer(4);
+ buf.MapAsync(wgpu::MapMode::Read, 0, 4, ToMockBufferMapAsyncCallback, nullptr);
+
+ EXPECT_CALL(*mockBufferMapAsyncCallback, Call(WGPUBufferMapAsyncStatus_Unknown, _))
+ .Times(1);
+ buf.Destroy();
+
+ // The callback shouldn't be called again.
+ WaitForAllOperations(device);
+ }
+ {
+ wgpu::Buffer buf = CreateMapWriteBuffer(4);
+ buf.MapAsync(wgpu::MapMode::Write, 0, 4, ToMockBufferMapAsyncCallback, nullptr);
+
+ EXPECT_CALL(*mockBufferMapAsyncCallback, Call(WGPUBufferMapAsyncStatus_Unknown, _))
+ .Times(1);
+ buf.Destroy();
+
+ // The callback shouldn't be called again.
+ WaitForAllOperations(device);
+ }
+}
+
+// Test that the MapCallback isn't fired twice when unmap() is called inside the callback
+TEST_F(BufferValidationTest, MapAsync_UnmapCalledInCallback) {
+ {
+ wgpu::Buffer buf = CreateMapReadBuffer(4);
+ buf.MapAsync(wgpu::MapMode::Read, 0, 4, ToMockBufferMapAsyncCallback, nullptr);
+
+ EXPECT_CALL(*mockBufferMapAsyncCallback, Call(WGPUBufferMapAsyncStatus_Success, _))
+ .WillOnce(InvokeWithoutArgs([&]() { buf.Unmap(); }));
+
+ WaitForAllOperations(device);
+ }
+ {
+ wgpu::Buffer buf = CreateMapWriteBuffer(4);
+ buf.MapAsync(wgpu::MapMode::Write, 0, 4, ToMockBufferMapAsyncCallback, nullptr);
+
+ EXPECT_CALL(*mockBufferMapAsyncCallback, Call(WGPUBufferMapAsyncStatus_Success, _))
+ .WillOnce(InvokeWithoutArgs([&]() { buf.Unmap(); }));
+
+ WaitForAllOperations(device);
+ }
+}
+
+// Test that the MapCallback isn't fired twice when destroy() is called inside the callback
+TEST_F(BufferValidationTest, MapAsync_DestroyCalledInCallback) {
+ {
+ wgpu::Buffer buf = CreateMapReadBuffer(4);
+ buf.MapAsync(wgpu::MapMode::Read, 0, 4, ToMockBufferMapAsyncCallback, nullptr);
+
+ EXPECT_CALL(*mockBufferMapAsyncCallback, Call(WGPUBufferMapAsyncStatus_Success, _))
+ .WillOnce(InvokeWithoutArgs([&]() { buf.Destroy(); }));
+
+ WaitForAllOperations(device);
+ }
+ {
+ wgpu::Buffer buf = CreateMapWriteBuffer(4);
+ buf.MapAsync(wgpu::MapMode::Write, 0, 4, ToMockBufferMapAsyncCallback, nullptr);
+
+ EXPECT_CALL(*mockBufferMapAsyncCallback, Call(WGPUBufferMapAsyncStatus_Success, _))
+ .WillOnce(InvokeWithoutArgs([&]() { buf.Destroy(); }));
+
+ WaitForAllOperations(device);
+ }
+}
+
+// Test the success case for mapping buffer for reading
+TEST_F(BufferValidationTest, MapReadAsyncSuccess) {
wgpu::Buffer buf = CreateMapReadBuffer(4);
buf.MapReadAsync(ToMockBufferMapReadCallback, nullptr);
@@ -178,7 +463,7 @@
}
// Test the success case for mapping buffer for writing
-TEST_F(BufferValidationTest, MapWriteSuccess) {
+TEST_F(BufferValidationTest, MapWriteAsyncSuccess) {
wgpu::Buffer buf = CreateMapWriteBuffer(4);
buf.MapWriteAsync(ToMockBufferMapWriteCallback, nullptr);
@@ -605,6 +890,30 @@
wgpu::Buffer bufA = device.CreateBuffer(&descriptorA);
wgpu::Buffer bufB = device.CreateBuffer(&descriptorB);
+ bufA.MapAsync(wgpu::MapMode::Write, 0, 4, nullptr, nullptr);
+
+ wgpu::CommandEncoder encoder = device.CreateCommandEncoder();
+ encoder.CopyBufferToBuffer(bufA, 0, bufB, 0, 4);
+ wgpu::CommandBuffer commands = encoder.Finish();
+ ASSERT_DEVICE_ERROR(queue.Submit(1, &commands));
+ WaitForAllOperations(device);
+ }
+ {
+ wgpu::Buffer bufA = device.CreateBuffer(&descriptorA);
+ wgpu::Buffer bufB = device.CreateBuffer(&descriptorB);
+
+ bufB.MapAsync(wgpu::MapMode::Read, 0, 4, nullptr, nullptr);
+
+ wgpu::CommandEncoder encoder = device.CreateCommandEncoder();
+ encoder.CopyBufferToBuffer(bufA, 0, bufB, 0, 4);
+ wgpu::CommandBuffer commands = encoder.Finish();
+ ASSERT_DEVICE_ERROR(queue.Submit(1, &commands));
+ WaitForAllOperations(device);
+ }
+ {
+ wgpu::Buffer bufA = device.CreateBuffer(&descriptorA);
+ wgpu::Buffer bufB = device.CreateBuffer(&descriptorB);
+
bufA.MapWriteAsync(ToMockBufferMapWriteCallback, nullptr);
wgpu::CommandEncoder encoder = device.CreateCommandEncoder();
@@ -721,6 +1030,24 @@
buf.Unmap();
buf.Unmap();
}
+ {
+ wgpu::Buffer buf = CreateMapReadBuffer(4);
+ // Buffer starts unmapped. Unmap should succeed.
+ buf.Unmap();
+ buf.MapAsync(wgpu::MapMode::Read, 0, 4, nullptr, nullptr);
+ buf.Unmap();
+ // Unmapping twice should succeed
+ buf.Unmap();
+ }
+ {
+ wgpu::Buffer buf = CreateMapWriteBuffer(4);
+ // Buffer starts unmapped. Unmap should succeed.
+ buf.Unmap();
+ buf.MapAsync(wgpu::MapMode::Write, 0, 4, nullptr, nullptr);
+ // Unmapping twice should succeed
+ buf.Unmap();
+ buf.Unmap();
+ }
}
// Test that it is invalid to call GetMappedRange on an unmapped buffer.
@@ -782,6 +1109,32 @@
ASSERT_EQ(nullptr, buf.GetMappedRange());
ASSERT_EQ(nullptr, buf.GetConstMappedRange());
}
+ // Unmapped after MapAsync read case.
+ {
+ wgpu::Buffer buf = CreateMapReadBuffer(4);
+
+ buf.MapAsync(wgpu::MapMode::Read, 0, 4, ToMockBufferMapAsyncCallback, nullptr);
+ EXPECT_CALL(*mockBufferMapAsyncCallback, Call(WGPUBufferMapAsyncStatus_Success, _))
+ .Times(1);
+ WaitForAllOperations(device);
+ buf.Unmap();
+
+ ASSERT_EQ(nullptr, buf.GetMappedRange());
+ ASSERT_EQ(nullptr, buf.GetConstMappedRange());
+ }
+
+ // Unmapped after MapAsync write case.
+ {
+ wgpu::Buffer buf = CreateMapWriteBuffer(4);
+ buf.MapAsync(wgpu::MapMode::Write, 0, 4, ToMockBufferMapAsyncCallback, nullptr);
+ EXPECT_CALL(*mockBufferMapAsyncCallback, Call(WGPUBufferMapAsyncStatus_Success, _))
+ .Times(1);
+ WaitForAllOperations(device);
+ buf.Unmap();
+
+ ASSERT_EQ(nullptr, buf.GetMappedRange());
+ ASSERT_EQ(nullptr, buf.GetConstMappedRange());
+ }
}
// Test that it is invalid to call GetMappedRange on a destroyed buffer.
@@ -844,19 +1197,57 @@
ASSERT_EQ(nullptr, buf.GetMappedRange());
ASSERT_EQ(nullptr, buf.GetConstMappedRange());
}
+ // Destroyed after MapAsync read case.
+ {
+ wgpu::Buffer buf = CreateMapReadBuffer(4);
+
+ buf.MapAsync(wgpu::MapMode::Read, 0, 4, ToMockBufferMapAsyncCallback, nullptr);
+ EXPECT_CALL(*mockBufferMapAsyncCallback, Call(WGPUBufferMapAsyncStatus_Success, _))
+ .Times(1);
+ WaitForAllOperations(device);
+ buf.Destroy();
+
+ ASSERT_EQ(nullptr, buf.GetMappedRange());
+ ASSERT_EQ(nullptr, buf.GetConstMappedRange());
+ }
+
+ // Destroyed after MapAsync write case.
+ {
+ wgpu::Buffer buf = CreateMapWriteBuffer(4);
+ buf.MapAsync(wgpu::MapMode::Write, 0, 4, ToMockBufferMapAsyncCallback, nullptr);
+ EXPECT_CALL(*mockBufferMapAsyncCallback, Call(WGPUBufferMapAsyncStatus_Success, _))
+ .Times(1);
+ WaitForAllOperations(device);
+ buf.Destroy();
+
+ ASSERT_EQ(nullptr, buf.GetMappedRange());
+ ASSERT_EQ(nullptr, buf.GetConstMappedRange());
+ }
}
// Test that it is invalid to call GetMappedRange on a buffer afterMapReadAsync
TEST_F(BufferValidationTest, GetMappedRangeOnMappedForReading) {
- wgpu::Buffer buf = CreateMapReadBuffer(4);
+ {
+ wgpu::Buffer buf = CreateMapReadBuffer(4);
- buf.MapReadAsync(ToMockBufferMapReadCallback, nullptr);
- EXPECT_CALL(*mockBufferMapReadCallback,
- Call(WGPUBufferMapAsyncStatus_Success, Ne(nullptr), 4u, _))
- .Times(1);
- WaitForAllOperations(device);
+ buf.MapReadAsync(ToMockBufferMapReadCallback, nullptr);
+ EXPECT_CALL(*mockBufferMapReadCallback,
+ Call(WGPUBufferMapAsyncStatus_Success, Ne(nullptr), 4u, _))
+ .Times(1);
+ WaitForAllOperations(device);
- ASSERT_EQ(nullptr, buf.GetMappedRange());
+ ASSERT_EQ(nullptr, buf.GetMappedRange());
+ }
+ {
+ wgpu::Buffer buf = CreateMapReadBuffer(4);
+
+ buf.MapAsync(wgpu::MapMode::Read, 0, 4, ToMockBufferMapAsyncCallback, nullptr);
+ EXPECT_CALL(*mockBufferMapAsyncCallback, Call(WGPUBufferMapAsyncStatus_Success, _))
+ .Times(1);
+ WaitForAllOperations(device);
+
+ ASSERT_EQ(nullptr, buf.GetMappedRange());
+ }
}
// Test valid cases to call GetMappedRange on a buffer.