blob: bbb1afef92b159c3933404fb44a4a57ee4bd664f [file] [log] [blame]
// Copyright 2018 The Chromium Authors. All rights reserved.
// Use of this source code is governed by a BSD-style license that can be
// found in the LICENSE file.
#include "gpu/command_buffer/client/webgpu_implementation.h"
#include <algorithm>
#include <vector>
#include "base/numerics/checked_math.h"
#include "base/trace_event/trace_event.h"
#include "gpu/command_buffer/client/dawn_client_memory_transfer_service.h"
#include "gpu/command_buffer/client/gpu_control.h"
#include "gpu/command_buffer/client/shared_memory_limits.h"
#define GPU_CLIENT_SINGLE_THREAD_CHECK()
namespace gpu {
namespace webgpu {
#if BUILDFLAG(USE_DAWN)
WebGPUCommandSerializer::WebGPUCommandSerializer(
WebGPUCmdHelper* helper,
DawnClientMemoryTransferService* memory_transfer_service)
: helper_(helper), memory_transfer_service_(memory_transfer_service) {
c2s_transfer_buffer_.reset(new TransferBuffer(helper_));
const SharedMemoryLimits& limits = SharedMemoryLimits::ForWebGPUContext();
c2s_transfer_buffer_->Initialize(
limits.start_transfer_buffer_size, ImplementationBase::kStartingOffset,
limits.min_transfer_buffer_size, limits.max_transfer_buffer_size,
ImplementationBase::kAlignment);
DCHECK(helper_);
c2s_buffer_.reset(
new ScopedTransferBufferPtr(helper_, c2s_transfer_buffer_.get()));
c2s_buffer_default_size_ = limits.start_transfer_buffer_size;
DCHECK_GT(c2s_buffer_default_size_, 0u);
DCHECK(memory_transfer_service_);
dawn_wire::WireClientDescriptor descriptor = {};
descriptor.serializer = this;
descriptor.memoryTransferService = memory_transfer_service_;
wire_client_.reset(new dawn_wire::WireClient(descriptor));
}
WebGPUCommandSerializer::~WebGPUCommandSerializer() {}
void* WebGPUCommandSerializer::GetCmdSpace(size_t size) {
// The buffer size must be initialized before any commands are serialized.
if (c2s_buffer_default_size_ == 0u) {
NOTREACHED();
return nullptr;
}
base::CheckedNumeric<uint32_t> checked_next_offset(c2s_put_offset_);
checked_next_offset += size;
uint32_t next_offset;
bool next_offset_valid = checked_next_offset.AssignIfValid(&next_offset);
// If the buffer does not have enough space, or if the buffer is not
// initialized, flush and reset the command stream.
if (!next_offset_valid || next_offset > c2s_buffer_->size() ||
!c2s_buffer_->valid()) {
Flush();
uint32_t max_allocation = c2s_transfer_buffer_->GetMaxSize();
// TODO(crbug.com/951558): Handle command chunking or ensure commands aren't
// this large.
CHECK_LE(size, max_allocation);
uint32_t allocation_size =
std::max(c2s_buffer_default_size_, static_cast<uint32_t>(size));
TRACE_EVENT1(TRACE_DISABLED_BY_DEFAULT("gpu.dawn"),
"WebGPUCommandSerializer::GetCmdSpace", "bytes",
allocation_size);
c2s_buffer_->Reset(allocation_size);
c2s_put_offset_ = 0;
next_offset = size;
// TODO(crbug.com/951558): Handle OOM.
CHECK(c2s_buffer_->valid());
CHECK_LE(size, c2s_buffer_->size());
}
DCHECK(c2s_buffer_->valid());
uint8_t* ptr = static_cast<uint8_t*>(c2s_buffer_->address());
ptr += c2s_put_offset_;
c2s_put_offset_ = next_offset;
return ptr;
}
bool WebGPUCommandSerializer::Flush() {
if (c2s_buffer_->valid()) {
TRACE_EVENT1(TRACE_DISABLED_BY_DEFAULT("gpu.dawn"),
"WebGPUCommandSerializer::Flush", "bytes", c2s_put_offset_);
TRACE_EVENT_FLOW_BEGIN0(
TRACE_DISABLED_BY_DEFAULT("gpu.dawn"), "DawnCommands",
(static_cast<uint64_t>(c2s_buffer_->shm_id()) << 32) +
c2s_buffer_->offset());
c2s_buffer_->Shrink(c2s_put_offset_);
helper_->DawnCommands(c2s_buffer_->shm_id(), c2s_buffer_->offset(),
c2s_put_offset_);
c2s_put_offset_ = 0;
c2s_buffer_->Release();
}
memory_transfer_service_->FreeHandlesPendingToken(helper_->InsertToken());
return true;
}
WGPUDevice WebGPUCommandSerializer::GetDevice() const {
return wire_client_->GetDevice();
}
ReservedTexture WebGPUCommandSerializer::ReserveTexture(WGPUDevice device) {
dawn_wire::ReservedTexture reservation = wire_client_->ReserveTexture(device);
return {reservation.texture, reservation.id, reservation.generation};
}
bool WebGPUCommandSerializer::HandleCommands(const char* commands,
size_t command_size) {
return wire_client_->HandleCommands(commands, command_size);
}
#endif
// Include the auto-generated part of this file. We split this because it means
// we can easily edit the non-auto generated parts right here in this file
// instead of having to edit some template or the code generator.
#include "gpu/command_buffer/client/webgpu_implementation_impl_autogen.h"
WebGPUImplementation::WebGPUImplementation(
WebGPUCmdHelper* helper,
TransferBufferInterface* transfer_buffer,
GpuControl* gpu_control)
: ImplementationBase(helper, transfer_buffer, gpu_control),
helper_(helper) {}
WebGPUImplementation::~WebGPUImplementation() {
#if BUILDFLAG(USE_DAWN)
// Wait for all commands to finish or we may free shared memory while
// commands are still in flight.
if (command_serializer_) {
command_serializer_->Flush();
}
#endif
helper_->Finish();
#if BUILDFLAG(USE_DAWN)
// Now that commands are finished, free the wire client.
command_serializer_.reset();
// All client-side Dawn objects are now destroyed.
// Shared memory allocations for buffers that were still mapped at the time
// of destruction can now be safely freed.
memory_transfer_service_->FreeHandlesPendingToken(helper_->InsertToken());
helper_->Finish();
#endif
}
gpu::ContextResult WebGPUImplementation::Initialize(
const SharedMemoryLimits& limits) {
TRACE_EVENT0("gpu", "WebGPUImplementation::Initialize");
auto result = ImplementationBase::Initialize(limits);
if (result != gpu::ContextResult::kSuccess) {
return result;
}
#if BUILDFLAG(USE_DAWN)
memory_transfer_service_.reset(
new DawnClientMemoryTransferService(mapped_memory_.get()));
procs_ = dawn_wire::WireClient::GetProcs();
#endif
return gpu::ContextResult::kSuccess;
}
// ContextSupport implementation.
void WebGPUImplementation::SetAggressivelyFreeResources(
bool aggressively_free_resources) {
NOTIMPLEMENTED();
}
void WebGPUImplementation::Swap(uint32_t flags,
SwapCompletedCallback complete_callback,
PresentationCallback presentation_callback) {
NOTIMPLEMENTED();
}
void WebGPUImplementation::SwapWithBounds(
const std::vector<gfx::Rect>& rects,
uint32_t flags,
SwapCompletedCallback swap_completed,
PresentationCallback presentation_callback) {
NOTIMPLEMENTED();
}
void WebGPUImplementation::PartialSwapBuffers(
const gfx::Rect& sub_buffer,
uint32_t flags,
SwapCompletedCallback swap_completed,
PresentationCallback presentation_callback) {
NOTIMPLEMENTED();
}
void WebGPUImplementation::CommitOverlayPlanes(
uint32_t flags,
SwapCompletedCallback swap_completed,
PresentationCallback presentation_callback) {
NOTREACHED();
}
void WebGPUImplementation::ScheduleOverlayPlane(
int plane_z_order,
gfx::OverlayTransform plane_transform,
unsigned overlay_texture_id,
const gfx::Rect& display_bounds,
const gfx::RectF& uv_rect,
bool enable_blend,
unsigned gpu_fence_id) {
NOTREACHED();
}
uint64_t WebGPUImplementation::ShareGroupTracingGUID() const {
NOTIMPLEMENTED();
return 0;
}
void WebGPUImplementation::SetErrorMessageCallback(
base::RepeatingCallback<void(const char*, int32_t)> callback) {
NOTIMPLEMENTED();
}
bool WebGPUImplementation::ThreadSafeShallowLockDiscardableTexture(
uint32_t texture_id) {
NOTREACHED();
return false;
}
void WebGPUImplementation::CompleteLockDiscardableTexureOnContextThread(
uint32_t texture_id) {
NOTREACHED();
}
bool WebGPUImplementation::ThreadsafeDiscardableTextureIsDeletedForTracing(
uint32_t texture_id) {
NOTREACHED();
return false;
}
void* WebGPUImplementation::MapTransferCacheEntry(uint32_t serialized_size) {
NOTREACHED();
return nullptr;
}
void WebGPUImplementation::UnmapAndCreateTransferCacheEntry(uint32_t type,
uint32_t id) {
NOTREACHED();
}
bool WebGPUImplementation::ThreadsafeLockTransferCacheEntry(uint32_t type,
uint32_t id) {
NOTREACHED();
return false;
}
void WebGPUImplementation::UnlockTransferCacheEntries(
const std::vector<std::pair<uint32_t, uint32_t>>& entries) {
NOTREACHED();
}
void WebGPUImplementation::DeleteTransferCacheEntry(uint32_t type,
uint32_t id) {
NOTREACHED();
}
unsigned int WebGPUImplementation::GetTransferBufferFreeSize() const {
NOTREACHED();
return 0;
}
bool WebGPUImplementation::IsJpegDecodeAccelerationSupported() const {
NOTREACHED();
return false;
}
bool WebGPUImplementation::IsWebPDecodeAccelerationSupported() const {
NOTREACHED();
return false;
}
bool WebGPUImplementation::CanDecodeWithHardwareAcceleration(
const cc::ImageHeaderMetadata* image_metadata) const {
NOTREACHED();
return false;
}
// InterfaceBase implementation.
void WebGPUImplementation::GenSyncTokenCHROMIUM(GLbyte* sync_token) {
ImplementationBase::GenSyncToken(sync_token);
}
void WebGPUImplementation::GenUnverifiedSyncTokenCHROMIUM(GLbyte* sync_token) {
ImplementationBase::GenUnverifiedSyncToken(sync_token);
}
void WebGPUImplementation::VerifySyncTokensCHROMIUM(GLbyte** sync_tokens,
GLsizei count) {
ImplementationBase::VerifySyncTokens(sync_tokens, count);
}
void WebGPUImplementation::WaitSyncTokenCHROMIUM(const GLbyte* sync_token) {
ImplementationBase::WaitSyncToken(sync_token);
}
// ImplementationBase implementation.
void WebGPUImplementation::IssueShallowFlush() {
NOTIMPLEMENTED();
}
void WebGPUImplementation::SetGLError(GLenum error,
const char* function_name,
const char* msg) {
GPU_CLIENT_LOG("[" << GetLogPrefix() << "] Client Synthesized Error: "
<< gles2::GLES2Util::GetStringError(error) << ": "
<< function_name << ": " << msg);
NOTIMPLEMENTED();
}
// GpuControlClient implementation.
void WebGPUImplementation::OnGpuControlLostContext() {
NOTIMPLEMENTED();
}
void WebGPUImplementation::OnGpuControlLostContextMaybeReentrant() {
NOTIMPLEMENTED();
}
void WebGPUImplementation::OnGpuControlErrorMessage(const char* message,
int32_t id) {
NOTIMPLEMENTED();
}
void WebGPUImplementation::OnGpuControlSwapBuffersCompleted(
const SwapBuffersCompleteParams& params) {
NOTIMPLEMENTED();
}
void WebGPUImplementation::OnSwapBufferPresented(
uint64_t swap_id,
const gfx::PresentationFeedback& feedback) {
NOTIMPLEMENTED();
}
void WebGPUImplementation::OnGpuControlReturnData(
base::span<const uint8_t> data) {
#if BUILDFLAG(USE_DAWN)
static uint32_t return_trace_id = 0;
TRACE_EVENT_FLOW_END0(TRACE_DISABLED_BY_DEFAULT("gpu.dawn"),
"DawnReturnCommands", return_trace_id++);
TRACE_EVENT1(TRACE_DISABLED_BY_DEFAULT("gpu.dawn"),
"WebGPUImplementation::OnGpuControlReturnData", "bytes",
data.size());
if (data.size() <= sizeof(cmds::DawnReturnDataHeader)) {
// TODO(jiawei.shao@intel.com): Lose the context.
NOTREACHED();
return;
}
const cmds::DawnReturnDataHeader& dawnReturnDataHeader =
*reinterpret_cast<const cmds::DawnReturnDataHeader*>(data.data());
switch (dawnReturnDataHeader.return_data_type) {
case DawnReturnDataType::kDawnCommands: {
if (!command_serializer_) {
// TODO(jiawei.shao@intel.com): Lose the context.
NOTREACHED();
break;
}
if (data.size() < sizeof(cmds::DawnReturnCommandsInfo)) {
// TODO(jiawei.shao@intel.com): Lose the context.
NOTREACHED();
break;
}
const cmds::DawnReturnCommandsInfo* dawn_return_commands_info =
reinterpret_cast<const cmds::DawnReturnCommandsInfo*>(data.data());
if (!command_serializer_->HandleCommands(
reinterpret_cast<const char*>(
dawn_return_commands_info->deserialized_buffer),
data.size() - offsetof(cmds::DawnReturnCommandsInfo,
deserialized_buffer))) {
// TODO(enga): Lose the context.
NOTREACHED();
}
} break;
case DawnReturnDataType::kRequestedDawnAdapterProperties: {
if (data.size() < sizeof(cmds::DawnReturnAdapterInfo)) {
// TODO(jiawei.shao@intel.com): Lose the context.
NOTREACHED();
break;
}
const cmds::DawnReturnAdapterInfo* returned_adapter_info =
reinterpret_cast<const cmds::DawnReturnAdapterInfo*>(data.data());
GLuint request_adapter_serial =
returned_adapter_info->header.request_adapter_serial;
auto request_callback_iter =
request_adapter_callback_map_.find(request_adapter_serial);
if (request_callback_iter == request_adapter_callback_map_.end()) {
// TODO(jiawei.shao@intel.com): Lose the context.
NOTREACHED();
break;
}
auto& request_callback = request_callback_iter->second;
GLuint adapter_service_id =
returned_adapter_info->header.adapter_service_id;
WGPUDeviceProperties adapter_properties = {};
const volatile char* deserialized_buffer =
reinterpret_cast<const volatile char*>(
returned_adapter_info->deserialized_buffer);
dawn_wire::DeserializeWGPUDeviceProperties(&adapter_properties,
deserialized_buffer);
std::move(request_callback).Run(adapter_service_id, adapter_properties);
request_adapter_callback_map_.erase(request_callback_iter);
} break;
case DawnReturnDataType::kRequestedDeviceReturnInfo: {
if (data.size() < sizeof(cmds::DawnReturnRequestDeviceInfo)) {
// TODO(jiawei.shao@intel.com): Lose the context.
NOTREACHED();
break;
}
const cmds::DawnReturnRequestDeviceInfo* returned_request_device_info =
reinterpret_cast<const cmds::DawnReturnRequestDeviceInfo*>(
data.data());
GLuint request_device_serial =
returned_request_device_info->request_device_serial;
auto request_callback_iter =
request_device_callback_map_.find(request_device_serial);
if (request_callback_iter == request_device_callback_map_.end()) {
// TODO(jiawei.shao@intel.com): Lose the context.
NOTREACHED();
break;
}
auto& request_callback = request_callback_iter->second;
bool is_request_device_success =
returned_request_device_info->is_request_device_success;
std::move(request_callback).Run(is_request_device_success);
request_device_callback_map_.erase(request_callback_iter);
} break;
default:
// TODO(jiawei.shao@intel.com): Lose the context.
NOTREACHED();
break;
}
#endif
}
const DawnProcTable& WebGPUImplementation::GetProcs() const {
#if !BUILDFLAG(USE_DAWN)
NOTREACHED();
#endif
return procs_;
}
void WebGPUImplementation::FlushCommands() {
#if BUILDFLAG(USE_DAWN)
if (command_serializer_) {
command_serializer_->Flush();
}
#endif
helper_->Flush();
}
WGPUDevice WebGPUImplementation::GetDefaultDevice() {
#if BUILDFLAG(USE_DAWN)
DCHECK(command_serializer_);
return command_serializer_->GetDevice();
#else
NOTREACHED();
return {};
#endif
}
ReservedTexture WebGPUImplementation::ReserveTexture(WGPUDevice device) {
#if BUILDFLAG(USE_DAWN)
return command_serializer_->ReserveTexture(device);
#else
NOTREACHED();
return {};
#endif
}
uint32_t WebGPUImplementation::NextRequestAdapterSerial() {
return ++request_adapter_serial_;
}
bool WebGPUImplementation::RequestAdapterAsync(
PowerPreference power_preference,
base::OnceCallback<void(uint32_t, const WGPUDeviceProperties&)>
request_adapter_callback) {
uint32_t request_adapter_serial = NextRequestAdapterSerial();
// Avoid the overflow of request_adapter_serial and old slot being reused.
if (request_adapter_callback_map_.find(request_adapter_serial) !=
request_adapter_callback_map_.end()) {
return false;
}
helper_->RequestAdapter(request_adapter_serial,
static_cast<uint32_t>(power_preference));
helper_->Flush();
request_adapter_callback_map_[request_adapter_serial] =
std::move(request_adapter_callback);
return true;
}
uint32_t WebGPUImplementation::NextRequestDeviceSerial() {
return ++request_device_serial_;
}
bool WebGPUImplementation::RequestDeviceAsync(
uint32_t requested_adapter_id,
const WGPUDeviceProperties* requested_device_properties,
base::OnceCallback<void(bool)> request_device_callback) {
#if BUILDFLAG(USE_DAWN)
uint32_t request_device_serial = NextRequestDeviceSerial();
// Avoid the overflow of request_device_serial and old slot being reused.
if (request_device_callback_map_.find(request_device_serial) !=
request_device_callback_map_.end()) {
return false;
}
// TODO(jiawei.shao@intel.com): support multiple WebGPU devices. Each WebGPU
// device corresponds to a unique WebGPUCommandSerializer.
if (!command_serializer_) {
command_serializer_.reset(
new WebGPUCommandSerializer(helper_, memory_transfer_service_.get()));
}
request_device_callback_map_[request_device_serial] =
std::move(request_device_callback);
if (!requested_device_properties) {
helper_->RequestDevice(request_device_serial_, requested_adapter_id, 0, 0,
0);
return true;
}
size_t serialized_device_properties_size =
dawn_wire::SerializedWGPUDevicePropertiesSize(
requested_device_properties);
DCHECK_NE(0u, serialized_device_properties_size);
// TODO(jiawei.shao@intel.com): transfer requested_device_properties with
// c2s_buffer_ in the WebGPUCommandSerializer directly when we support
// multiple WebGPU devices.
ScopedTransferBufferPtr transfer_buffer(serialized_device_properties_size,
helper_, transfer_buffer_);
dawn_wire::SerializeWGPUDeviceProperties(
requested_device_properties,
reinterpret_cast<char*>(transfer_buffer.address()));
helper_->RequestDevice(request_device_serial, requested_adapter_id,
transfer_buffer.shm_id(), transfer_buffer.offset(),
serialized_device_properties_size);
transfer_buffer.Release();
helper_->Flush();
return true;
#else
NOTREACHED();
return false;
#endif
}
} // namespace webgpu
} // namespace gpu