blob: e7fdebf6895b8cfa11c0a1793d1f3a358089ac01 [file] [log] [blame]
// Copyright 2022 The Chromium Authors
// Use of this source code is governed by a BSD-style license that can be
// found in the LICENSE file.
#include "gpu/command_buffer/service/shared_image/compound_image_backing.h"
#include "base/feature_list.h"
#include "base/functional/bind.h"
#include "base/logging.h"
#include "base/memory/ptr_util.h"
#include "base/memory/scoped_refptr.h"
#include "base/metrics/histogram_macros.h"
#include "base/strings/stringprintf.h"
#include "base/trace_event/memory_allocator_dump_guid.h"
#include "base/trace_event/process_memory_dump.h"
#include "components/viz/common/resources/shared_image_format.h"
#include "components/viz/common/resources/shared_image_format_utils.h"
#include "gpu/command_buffer/common/mailbox.h"
#include "gpu/command_buffer/common/shared_image_usage.h"
#include "gpu/command_buffer/service/memory_tracking.h"
#include "gpu/command_buffer/service/shared_context_state.h"
#include "gpu/command_buffer/service/shared_image/shared_image_backing_factory.h"
#include "gpu/command_buffer/service/shared_image/shared_image_copy_manager.h"
#include "gpu/command_buffer/service/shared_image/shared_image_factory.h"
#include "gpu/command_buffer/service/shared_image/shared_image_format_service_utils.h"
#include "gpu/command_buffer/service/shared_image/shared_image_manager.h"
#include "gpu/command_buffer/service/shared_image/shared_image_representation.h"
#include "gpu/command_buffer/service/shared_image/shared_memory_image_backing.h"
#include "gpu/command_buffer/service/shared_image/shared_memory_image_backing_factory.h"
#include "gpu/command_buffer/service/shared_memory_region_wrapper.h"
#include "gpu/config/gpu_finch_features.h"
#include "third_party/skia/include/core/SkAlphaType.h"
#include "third_party/skia/include/core/SkPixmap.h"
#include "third_party/skia/include/core/SkSurface.h"
#include "third_party/skia/include/core/SkSurfaceProps.h"
#include "third_party/skia/include/gpu/ganesh/GrDirectContext.h"
#include "third_party/skia/include/gpu/ganesh/GrTypes.h"
#include "third_party/skia/include/gpu/graphite/BackendTexture.h"
#include "third_party/skia/include/private/chromium/GrPromiseImageTexture.h"
#include "ui/gfx/buffer_format_util.h"
#include "ui/gfx/buffer_types.h"
#include "ui/gfx/color_space.h"
#include "ui/gfx/geometry/size.h"
#include "ui/gfx/gpu_memory_buffer_handle.h"
namespace gpu {
namespace {
// Allows CompoundImageBacking to allocate backings during runtime if a
// compatible backing to serve clients requested usage is not already present.
BASE_FEATURE(kUseDynamicBackingAllocations, base::FEATURE_DISABLED_BY_DEFAULT);
constexpr AccessStreamSet kMemoryStreamSet = {SharedImageAccessStream::kMemory};
// Unique GUIDs for child backings.
base::trace_event::MemoryAllocatorDumpGuid GetSubBackingGUIDForTracing(
const Mailbox& mailbox,
int backing_index) {
return base::trace_event::MemoryAllocatorDumpGuid(
base::StringPrintf("gpu-shared-image/%s/sub-backing/%d",
mailbox.ToDebugString().c_str(), backing_index));
}
#if BUILDFLAG(IS_WIN)
// Only allow shmem overlays for NV12 on Windows.
// This moves the SCANOUT flag from the GPU backing to the shmem backing in the
// CompoundImageBacking.
constexpr bool kAllowShmOverlays = true;
#else
constexpr bool kAllowShmOverlays = false;
#endif
gpu::SharedImageUsageSet GetShmSharedImageUsage(SharedImageUsageSet usage) {
gpu::SharedImageUsageSet new_usage = SHARED_IMAGE_USAGE_CPU_WRITE_ONLY;
if (usage.Has(SHARED_IMAGE_USAGE_CPU_READ)) {
new_usage |= SHARED_IMAGE_USAGE_CPU_READ;
}
if (kAllowShmOverlays && usage.Has(gpu::SHARED_IMAGE_USAGE_SCANOUT)) {
new_usage |= SHARED_IMAGE_USAGE_SCANOUT;
}
return new_usage;
}
// This might need further tweaking in order to be able to choose appropriate
// backing. Note that the backing usually doesnt know at this point if the
// access stream will be used for read or write.
SharedImageUsageSet GetUsageFromAccessStream(SharedImageAccessStream stream) {
switch (stream) {
case SharedImageAccessStream::kGL:
return SHARED_IMAGE_USAGE_GLES2_READ | SHARED_IMAGE_USAGE_GLES2_WRITE;
case SharedImageAccessStream::kSkia:
return SHARED_IMAGE_USAGE_RASTER_READ | SHARED_IMAGE_USAGE_RASTER_WRITE |
SHARED_IMAGE_USAGE_DISPLAY_READ | SHARED_IMAGE_USAGE_DISPLAY_WRITE;
case SharedImageAccessStream::kDawn:
return SHARED_IMAGE_USAGE_WEBGPU_READ | SHARED_IMAGE_USAGE_WEBGPU_WRITE;
case SharedImageAccessStream::kOverlay:
return SHARED_IMAGE_USAGE_SCANOUT;
case SharedImageAccessStream::kVaapi:
return SHARED_IMAGE_USAGE_VIDEO_DECODE;
case SharedImageAccessStream::kMemory:
// Below usage set ensures that only SharedMemoryImageBacking will be able
// to support this stream.
return SHARED_IMAGE_USAGE_CPU_WRITE_ONLY | SHARED_IMAGE_USAGE_CPU_READ |
SHARED_IMAGE_USAGE_RASTER_COPY_SOURCE;
default:
NOTREACHED();
}
}
} // namespace
// Wrapped representation types are not in the anonymous namespace because they
// need to be friend classes to real representations to access protected
// virtual functions.
class WrappedGLTextureCompoundImageRepresentation
: public GLTextureImageRepresentation {
public:
WrappedGLTextureCompoundImageRepresentation(
SharedImageManager* manager,
SharedImageBacking* backing,
MemoryTypeTracker* tracker,
std::unique_ptr<GLTextureImageRepresentation> wrapped)
: GLTextureImageRepresentation(manager, backing, tracker),
wrapped_(std::move(wrapped)) {
DCHECK(wrapped_);
}
CompoundImageBacking* compound_backing() {
return static_cast<CompoundImageBacking*>(backing());
}
// GLTextureImageRepresentation implementation.
bool BeginAccess(GLenum mode) final {
AccessMode access_mode =
mode == kReadAccessMode ? AccessMode::kRead : AccessMode::kWrite;
compound_backing()->NotifyBeginAccess(wrapped_->backing(), access_mode);
return wrapped_->BeginAccess(mode);
}
void EndAccess() final { wrapped_->EndAccess(); }
gpu::TextureBase* GetTextureBase(int plane_index) final {
return wrapped_->GetTextureBase(plane_index);
}
bool SupportsMultipleConcurrentReadAccess() final {
return wrapped_->SupportsMultipleConcurrentReadAccess();
}
gles2::Texture* GetTexture(int plane_index) final {
return wrapped_->GetTexture(plane_index);
}
private:
std::unique_ptr<GLTextureImageRepresentation> wrapped_;
};
class WrappedGLTexturePassthroughCompoundImageRepresentation
: public GLTexturePassthroughImageRepresentation {
public:
WrappedGLTexturePassthroughCompoundImageRepresentation(
SharedImageManager* manager,
SharedImageBacking* backing,
MemoryTypeTracker* tracker,
std::unique_ptr<GLTexturePassthroughImageRepresentation> wrapped)
: GLTexturePassthroughImageRepresentation(manager, backing, tracker),
wrapped_(std::move(wrapped)) {
DCHECK(wrapped_);
}
CompoundImageBacking* compound_backing() {
return static_cast<CompoundImageBacking*>(backing());
}
// GLTexturePassthroughImageRepresentation implementation.
bool BeginAccess(GLenum mode) final {
AccessMode access_mode =
mode == kReadAccessMode ? AccessMode::kRead : AccessMode::kWrite;
compound_backing()->NotifyBeginAccess(wrapped_->backing(), access_mode);
return wrapped_->BeginAccess(mode);
}
void EndAccess() final { wrapped_->EndAccess(); }
gpu::TextureBase* GetTextureBase(int plane_index) final {
return wrapped_->GetTextureBase(plane_index);
}
bool SupportsMultipleConcurrentReadAccess() final {
return wrapped_->SupportsMultipleConcurrentReadAccess();
}
const scoped_refptr<gles2::TexturePassthrough>& GetTexturePassthrough(
int plane_index) final {
return wrapped_->GetTexturePassthrough(plane_index);
}
private:
std::unique_ptr<GLTexturePassthroughImageRepresentation> wrapped_;
};
class WrappedSkiaGaneshCompoundImageRepresentation
: public SkiaGaneshImageRepresentation {
public:
WrappedSkiaGaneshCompoundImageRepresentation(
GrDirectContext* gr_context,
SharedImageManager* manager,
SharedImageBacking* backing,
MemoryTypeTracker* tracker,
std::unique_ptr<SkiaGaneshImageRepresentation> wrapped)
: SkiaGaneshImageRepresentation(gr_context, manager, backing, tracker),
wrapped_(std::move(wrapped)) {
DCHECK(wrapped_);
}
CompoundImageBacking* compound_backing() {
return static_cast<CompoundImageBacking*>(backing());
}
// SkiaImageRepresentation implementation.
bool SupportsMultipleConcurrentReadAccess() final {
return wrapped_->SupportsMultipleConcurrentReadAccess();
}
std::vector<sk_sp<SkSurface>> BeginWriteAccess(
int final_msaa_count,
const SkSurfaceProps& surface_props,
const gfx::Rect& update_rect,
std::vector<GrBackendSemaphore>* begin_semaphores,
std::vector<GrBackendSemaphore>* end_semaphores,
std::unique_ptr<skgpu::MutableTextureState>* end_state) final {
compound_backing()->NotifyBeginAccess(wrapped_->backing(),
AccessMode::kWrite);
return wrapped_->BeginWriteAccess(final_msaa_count, surface_props,
update_rect, begin_semaphores,
end_semaphores, end_state);
}
std::vector<sk_sp<GrPromiseImageTexture>> BeginWriteAccess(
std::vector<GrBackendSemaphore>* begin_semaphores,
std::vector<GrBackendSemaphore>* end_semaphores,
std::unique_ptr<skgpu::MutableTextureState>* end_state) final {
compound_backing()->NotifyBeginAccess(wrapped_->backing(),
AccessMode::kWrite);
return wrapped_->BeginWriteAccess(begin_semaphores, end_semaphores,
end_state);
}
void EndWriteAccess() final { wrapped_->EndWriteAccess(); }
std::vector<sk_sp<GrPromiseImageTexture>> BeginReadAccess(
std::vector<GrBackendSemaphore>* begin_semaphores,
std::vector<GrBackendSemaphore>* end_semaphores,
std::unique_ptr<skgpu::MutableTextureState>* end_state) final {
compound_backing()->NotifyBeginAccess(wrapped_->backing(),
AccessMode::kRead);
return wrapped_->BeginReadAccess(begin_semaphores, end_semaphores,
end_state);
}
void EndReadAccess() final { wrapped_->EndReadAccess(); }
private:
std::unique_ptr<SkiaGaneshImageRepresentation> wrapped_;
};
class WrappedSkiaGraphiteCompoundImageRepresentation
: public SkiaGraphiteImageRepresentation {
public:
WrappedSkiaGraphiteCompoundImageRepresentation(
SharedImageManager* manager,
SharedImageBacking* backing,
MemoryTypeTracker* tracker,
std::unique_ptr<SkiaGraphiteImageRepresentation> wrapped)
: SkiaGraphiteImageRepresentation(manager, backing, tracker),
wrapped_(std::move(wrapped)) {
CHECK(wrapped_);
}
CompoundImageBacking* compound_backing() {
return static_cast<CompoundImageBacking*>(backing());
}
// SkiaGraphiteImageRepresentation implementation.
bool SupportsMultipleConcurrentReadAccess() final {
return wrapped_->SupportsMultipleConcurrentReadAccess();
}
std::vector<sk_sp<SkSurface>> BeginWriteAccess(
const SkSurfaceProps& surface_props,
const gfx::Rect& update_rect) final {
compound_backing()->NotifyBeginAccess(wrapped_->backing(),
AccessMode::kWrite);
return wrapped_->BeginWriteAccess(surface_props, update_rect);
}
std::vector<scoped_refptr<GraphiteTextureHolder>> BeginWriteAccess() final {
compound_backing()->NotifyBeginAccess(wrapped_->backing(),
AccessMode::kWrite);
return wrapped_->BeginWriteAccess();
}
void EndWriteAccess() final { wrapped_->EndWriteAccess(); }
std::vector<scoped_refptr<GraphiteTextureHolder>> BeginReadAccess() final {
compound_backing()->NotifyBeginAccess(wrapped_->backing(),
AccessMode::kRead);
return wrapped_->BeginReadAccess();
}
void EndReadAccess() final { wrapped_->EndReadAccess(); }
private:
std::unique_ptr<SkiaGraphiteImageRepresentation> wrapped_;
};
class WrappedDawnCompoundImageRepresentation : public DawnImageRepresentation {
public:
WrappedDawnCompoundImageRepresentation(
SharedImageManager* manager,
SharedImageBacking* backing,
MemoryTypeTracker* tracker,
std::unique_ptr<DawnImageRepresentation> wrapped)
: DawnImageRepresentation(manager, backing, tracker),
wrapped_(std::move(wrapped)) {
DCHECK(wrapped_);
}
CompoundImageBacking* compound_backing() {
return static_cast<CompoundImageBacking*>(backing());
}
// DawnImageRepresentation implementation.
wgpu::Texture BeginAccess(wgpu::TextureUsage webgpu_usage,
wgpu::TextureUsage internal_usage) final {
AccessMode access_mode =
webgpu_usage & kWriteUsage ? AccessMode::kWrite : AccessMode::kRead;
if (internal_usage & kWriteUsage) {
access_mode = AccessMode::kWrite;
}
compound_backing()->NotifyBeginAccess(wrapped_->backing(), access_mode);
return wrapped_->BeginAccess(webgpu_usage, internal_usage);
}
void EndAccess() final { wrapped_->EndAccess(); }
private:
std::unique_ptr<DawnImageRepresentation> wrapped_;
};
class WrappedOverlayCompoundImageRepresentation
: public OverlayImageRepresentation {
public:
WrappedOverlayCompoundImageRepresentation(
SharedImageManager* manager,
SharedImageBacking* backing,
MemoryTypeTracker* tracker,
std::unique_ptr<OverlayImageRepresentation> wrapped)
: OverlayImageRepresentation(manager, backing, tracker),
wrapped_(std::move(wrapped)) {
DCHECK(wrapped_);
}
CompoundImageBacking* compound_backing() {
return static_cast<CompoundImageBacking*>(backing());
}
// OverlayImageRepresentation implementation.
bool BeginReadAccess(gfx::GpuFenceHandle& acquire_fence) final {
compound_backing()->NotifyBeginAccess(wrapped_->backing(),
AccessMode::kRead);
return wrapped_->BeginReadAccess(acquire_fence);
}
void EndReadAccess(gfx::GpuFenceHandle release_fence) final {
return wrapped_->EndReadAccess(std::move(release_fence));
}
#if BUILDFLAG(IS_ANDROID)
AHardwareBuffer* GetAHardwareBuffer() final {
return wrapped_->GetAHardwareBuffer();
}
std::unique_ptr<base::android::ScopedHardwareBufferFenceSync>
GetAHardwareBufferFenceSync() final {
return wrapped_->GetAHardwareBufferFenceSync();
}
#elif BUILDFLAG(IS_WIN)
std::optional<gl::DCLayerOverlayImage> GetDCLayerOverlayImage() final {
return wrapped_->GetDCLayerOverlayImage();
}
#elif BUILDFLAG(IS_APPLE)
gfx::ScopedIOSurface GetIOSurface() const final {
return wrapped_->GetIOSurface();
}
bool IsInUseByWindowServer() const final {
return wrapped_->IsInUseByWindowServer();
}
#endif
private:
std::unique_ptr<OverlayImageRepresentation> wrapped_;
};
// static
bool CompoundImageBacking::IsValidSharedMemoryBufferFormat(
const gfx::Size& size,
viz::SharedImageFormat format) {
if (format.PrefersExternalSampler() ||
!viz::HasEquivalentBufferFormat(format)) {
DVLOG(1) << "Not a valid format: " << format.ToString();
return false;
}
if (!IsSizeForBufferHandleValid(size, format)) {
DVLOG(1) << "Invalid image size: " << size.ToString()
<< " for format: " << format.ToString();
return false;
}
return true;
}
// static
SharedImageUsageSet CompoundImageBacking::GetGpuSharedImageUsage(
SharedImageUsageSet usage) {
// Add allow copying from the shmem backing to the gpu backing.
usage |= SHARED_IMAGE_USAGE_CPU_UPLOAD;
if (kAllowShmOverlays) {
// Remove SCANOUT usage since it was previously moved to the shmem backing.
// See: |GetShmSharedImageUsage|
usage.RemoveAll(SharedImageUsageSet(gpu::SHARED_IMAGE_USAGE_SCANOUT));
}
if (usage.Has(SHARED_IMAGE_USAGE_CPU_READ)) {
// Remove CPU_READ usage since it was previously moved to the shmem backing.
// See: |GetShmSharedImageUsage|
usage.RemoveAll(SharedImageUsageSet(gpu::SHARED_IMAGE_USAGE_CPU_READ));
}
if (usage.Has(SHARED_IMAGE_USAGE_CPU_WRITE_ONLY)) {
// Remove CPU_WRITE usage since it was previously moved to the shmem
// backing. See: |GetShmSharedImageUsage|
usage.RemoveAll(
SharedImageUsageSet(gpu::SHARED_IMAGE_USAGE_CPU_WRITE_ONLY));
}
return usage;
}
// static
std::unique_ptr<SharedImageBacking> CompoundImageBacking::Create(
SharedImageFactory* shared_image_factory,
scoped_refptr<SharedImageCopyManager> copy_manager,
const Mailbox& mailbox,
gfx::GpuMemoryBufferHandle handle,
viz::SharedImageFormat format,
const gfx::Size& size,
const gfx::ColorSpace& color_space,
GrSurfaceOrigin surface_origin,
SkAlphaType alpha_type,
SharedImageUsageSet usage,
std::string debug_label) {
if (!IsValidSharedMemoryBufferFormat(size, format)) {
return nullptr;
}
auto* gpu_backing_factory = shared_image_factory->GetFactoryByUsage(
GetGpuSharedImageUsage(SharedImageUsageSet(usage)), format, size,
/*pixel_data=*/{}, gfx::EMPTY_BUFFER);
if (!gpu_backing_factory) {
return nullptr;
}
auto shm_backing = SharedMemoryImageBackingFactory().CreateSharedImage(
mailbox, format, size, color_space, surface_origin, alpha_type,
GetShmSharedImageUsage(usage), debug_label,
/*is_thread_safe=*/false, std::move(handle));
if (!shm_backing) {
return nullptr;
}
shm_backing->SetNotRefCounted();
return base::WrapUnique(new CompoundImageBacking(
mailbox, format, size, color_space, surface_origin, alpha_type, usage,
std::move(debug_label), std::move(shm_backing),
shared_image_factory->GetWeakPtr(), gpu_backing_factory->GetWeakPtr(),
std::move(copy_manager)));
}
// static
std::unique_ptr<SharedImageBacking> CompoundImageBacking::Create(
SharedImageFactory* shared_image_factory,
scoped_refptr<SharedImageCopyManager> copy_manager,
const Mailbox& mailbox,
viz::SharedImageFormat format,
const gfx::Size& size,
const gfx::ColorSpace& color_space,
GrSurfaceOrigin surface_origin,
SkAlphaType alpha_type,
SharedImageUsageSet usage,
std::string debug_label,
gfx::BufferUsage buffer_usage) {
if (!IsValidSharedMemoryBufferFormat(size, format)) {
return nullptr;
}
auto* gpu_backing_factory = shared_image_factory->GetFactoryByUsage(
GetGpuSharedImageUsage(SharedImageUsageSet(usage)), format, size,
/*pixel_data=*/{}, gfx::EMPTY_BUFFER);
if (!gpu_backing_factory) {
return nullptr;
}
auto shm_backing = SharedMemoryImageBackingFactory().CreateSharedImage(
mailbox, format, kNullSurfaceHandle, size, color_space, surface_origin,
alpha_type, GetShmSharedImageUsage(usage), debug_label,
/*is_thread_safe=*/false, buffer_usage);
if (!shm_backing) {
return nullptr;
}
shm_backing->SetNotRefCounted();
return base::WrapUnique(new CompoundImageBacking(
mailbox, format, size, color_space, surface_origin, alpha_type, usage,
std::move(debug_label), std::move(shm_backing),
shared_image_factory->GetWeakPtr(), gpu_backing_factory->GetWeakPtr(),
std::move(copy_manager), std::move(buffer_usage)));
}
// static
std::unique_ptr<SharedImageBacking>
CompoundImageBacking::CreateSharedMemoryForTesting(
SharedImageBackingFactory* gpu_backing_factory,
scoped_refptr<SharedImageCopyManager> copy_manager,
const Mailbox& mailbox,
gfx::GpuMemoryBufferHandle handle,
viz::SharedImageFormat format,
const gfx::Size& size,
const gfx::ColorSpace& color_space,
GrSurfaceOrigin surface_origin,
SkAlphaType alpha_type,
SharedImageUsageSet usage,
std::string debug_label) {
DCHECK(IsValidSharedMemoryBufferFormat(size, format));
auto shm_backing = SharedMemoryImageBackingFactory().CreateSharedImage(
mailbox, format, size, color_space, surface_origin, alpha_type,
GetShmSharedImageUsage(usage), debug_label,
/*is_thread_safe=*/false, std::move(handle));
if (!shm_backing) {
return nullptr;
}
shm_backing->SetNotRefCounted();
return base::WrapUnique(new CompoundImageBacking(
mailbox, format, size, color_space, surface_origin, alpha_type, usage,
std::move(debug_label), std::move(shm_backing),
/*shared_image_factory=*/nullptr, gpu_backing_factory->GetWeakPtr(),
std::move(copy_manager)));
}
// static
std::unique_ptr<SharedImageBacking>
CompoundImageBacking::CreateSharedMemoryForTesting(
SharedImageBackingFactory* gpu_backing_factory,
scoped_refptr<SharedImageCopyManager> copy_manager,
const Mailbox& mailbox,
viz::SharedImageFormat format,
const gfx::Size& size,
const gfx::ColorSpace& color_space,
GrSurfaceOrigin surface_origin,
SkAlphaType alpha_type,
SharedImageUsageSet usage,
std::string debug_label,
gfx::BufferUsage buffer_usage) {
DCHECK(IsValidSharedMemoryBufferFormat(size, format));
auto shm_backing = SharedMemoryImageBackingFactory().CreateSharedImage(
mailbox, format, kNullSurfaceHandle, size, color_space, surface_origin,
alpha_type, GetShmSharedImageUsage(usage), debug_label,
/*is_thread_safe=*/false, buffer_usage);
if (!shm_backing) {
return nullptr;
}
shm_backing->SetNotRefCounted();
return base::WrapUnique(new CompoundImageBacking(
mailbox, format, size, color_space, surface_origin, alpha_type, usage,
std::move(debug_label), std::move(shm_backing),
/*shared_image_factory=*/nullptr, gpu_backing_factory->GetWeakPtr(),
std::move(copy_manager), std::move(buffer_usage)));
}
CompoundImageBacking::CompoundImageBacking(
const Mailbox& mailbox,
viz::SharedImageFormat format,
const gfx::Size& size,
const gfx::ColorSpace& color_space,
GrSurfaceOrigin surface_origin,
SkAlphaType alpha_type,
SharedImageUsageSet usage,
std::string debug_label,
std::unique_ptr<SharedImageBacking> shm_backing,
base::WeakPtr<SharedImageFactory> shared_image_factory,
base::WeakPtr<SharedImageBackingFactory> gpu_backing_factory,
scoped_refptr<SharedImageCopyManager> copy_manager,
std::optional<gfx::BufferUsage> buffer_usage)
: SharedImageBacking(mailbox,
format,
size,
color_space,
surface_origin,
alpha_type,
usage,
debug_label,
shm_backing->GetEstimatedSize(),
/*is_thread_safe=*/false,
std::move(buffer_usage)),
shared_image_factory_(std::move(shared_image_factory)),
copy_manager_(std::move(copy_manager)) {
DCHECK(shm_backing);
DCHECK_EQ(size, shm_backing->size());
// Create shared-memory element (stream = kMemory).
ElementHolder shm_element;
shm_element.access_streams = kMemoryStreamSet;
if (kAllowShmOverlays && usage.Has(gpu::SHARED_IMAGE_USAGE_SCANOUT)) {
shm_element.access_streams.Put(SharedImageAccessStream::kOverlay);
}
shm_element.content_id_ = latest_content_id_;
shm_element.backing = std::move(shm_backing);
elements_.push_back(std::move(shm_element));
// Create placeholder for GPU-backed element (streams = all except kMemory).
ElementHolder gpu_element;
gpu_element.access_streams =
base::Difference(AccessStreamSet::All(), kMemoryStreamSet);
// CreateBackingFromBackingFactory will be called on demand. Hence this is
// lazy backing creation.
gpu_element.create_callback =
base::BindOnce(&CompoundImageBacking::CreateBackingFromBackingFactory,
base::Unretained(this), std::move(gpu_backing_factory),
std::move(debug_label));
elements_.push_back(std::move(gpu_element));
}
CompoundImageBacking::~CompoundImageBacking() {
if (pending_copy_to_gmb_callback_) {
std::move(pending_copy_to_gmb_callback_).Run(/*success=*/false);
}
}
void CompoundImageBacking::NotifyBeginAccess(SharedImageBacking* backing,
RepresentationAccessMode mode) {
ElementHolder* access_element = GetElement(backing);
if (!access_element) {
LOG(ERROR) << "backing not in the element list.";
return;
}
// If this element already has the latest content, we're good for read access.
if (access_element->content_id_ == latest_content_id_) {
if (mode == RepresentationAccessMode::kWrite) {
// For write access, this backing is about to become the new latest.
++latest_content_id_;
access_element->content_id_ = latest_content_id_;
}
return;
}
// This backing is stale. We need to find the element which has the latest
// content and copy from it.
ElementHolder* latest_content_element = GetElementWithLatestContent();
bool updated_backing = false;
if (latest_content_element &&
copy_manager_->CopyImage(
/*src_backing=*/latest_content_element->GetBacking(),
/*dst_backing=*/access_element->GetBacking())) {
updated_backing = true;
} else {
LOG(ERROR)
<< "Failed to copy between backings. Backing can be using stale data";
}
// Update content IDs. In case of write, we are updating the
// |latest_content_id_| as well as marking the |access_element| as having
// latest content irrespective of above copy failures since write will likely
// overwrite all of the previous content. Although not necessarily true for
// partial writes. For read, we only mark the |access_element| as having
// latest content if the copy succeeded.
if (mode == RepresentationAccessMode::kWrite) {
++latest_content_id_;
access_element->content_id_ = latest_content_id_;
} else if (updated_backing) {
access_element->content_id_ = latest_content_id_;
}
}
SharedImageBackingType CompoundImageBacking::GetType() const {
return SharedImageBackingType::kCompound;
}
void CompoundImageBacking::Update(std::unique_ptr<gfx::GpuFence> in_fence) {
CHECK(!in_fence);
// Find a shared memory backing to update.
auto& shm_element = GetShmElement();
CHECK(shm_element.backing);
++latest_content_id_;
shm_element.content_id_ = latest_content_id_;
}
bool CompoundImageBacking::CopyToGpuMemoryBuffer() {
auto& shm_element = GetShmElement();
if (HasLatestContent(shm_element)) {
return true;
}
auto* gpu_backing = GetGpuBacking();
if (!gpu_backing ||
!copy_manager_->CopyImage(gpu_backing, shm_element.GetBacking())) {
DLOG(ERROR) << "Failed to copy from GPU backing to shared memory";
return false;
}
shm_element.content_id_ = latest_content_id_;
return true;
}
void CompoundImageBacking::CopyToGpuMemoryBufferAsync(
base::OnceCallback<void(bool)> callback) {
auto& shm_element = GetShmElement();
if (HasLatestContent(shm_element)) {
std::move(callback).Run(true);
return;
}
if (pending_copy_to_gmb_callback_) {
DLOG(ERROR) << "Existing CopyToGpuMemoryBuffer operation pending";
std::move(callback).Run(false);
return;
}
auto* gpu_backing = GetGpuBacking();
if (!gpu_backing) {
DLOG(ERROR) << "Failed to copy from GPU backing to shared memory";
std::move(callback).Run(false);
return;
}
pending_copy_to_gmb_callback_ = std::move(callback);
gpu_backing->ReadbackToMemoryAsync(
GetSharedMemoryPixmaps(),
base::BindOnce(&CompoundImageBacking::OnCopyToGpuMemoryBufferComplete,
base::Unretained(this)));
}
void CompoundImageBacking::OnCopyToGpuMemoryBufferComplete(bool success) {
if (success) {
auto& shm_element = GetShmElement();
shm_element.content_id_ = latest_content_id_;
}
std::move(pending_copy_to_gmb_callback_).Run(success);
}
gfx::Rect CompoundImageBacking::ClearedRect() const {
// Copy on access will always ensure backing is cleared by first access.
return gfx::Rect(size());
}
void CompoundImageBacking::SetClearedRect(const gfx::Rect& cleared_rect) {}
gfx::GpuMemoryBufferHandle CompoundImageBacking::GetGpuMemoryBufferHandle() {
auto& element = GetShmElement();
CHECK(element.backing);
return element.backing->GetGpuMemoryBufferHandle();
}
std::unique_ptr<DawnImageRepresentation> CompoundImageBacking::ProduceDawn(
SharedImageManager* manager,
MemoryTypeTracker* tracker,
const wgpu::Device& device,
wgpu::BackendType backend_type,
std::vector<wgpu::TextureFormat> view_formats,
scoped_refptr<SharedContextState> context_state) {
auto* backing = GetOrAllocateBacking(SharedImageAccessStream::kDawn);
if (!backing)
return nullptr;
auto real_rep = backing->ProduceDawn(manager, tracker, device, backend_type,
std::move(view_formats), context_state);
if (!real_rep)
return nullptr;
return std::make_unique<WrappedDawnCompoundImageRepresentation>(
manager, this, tracker, std::move(real_rep));
}
std::unique_ptr<GLTextureImageRepresentation>
CompoundImageBacking::ProduceGLTexture(SharedImageManager* manager,
MemoryTypeTracker* tracker) {
auto* backing = GetOrAllocateBacking(SharedImageAccessStream::kGL);
if (!backing)
return nullptr;
auto real_rep = backing->ProduceGLTexture(manager, tracker);
if (!real_rep)
return nullptr;
return std::make_unique<WrappedGLTextureCompoundImageRepresentation>(
manager, this, tracker, std::move(real_rep));
}
std::unique_ptr<GLTexturePassthroughImageRepresentation>
CompoundImageBacking::ProduceGLTexturePassthrough(SharedImageManager* manager,
MemoryTypeTracker* tracker) {
auto* backing = GetOrAllocateBacking(SharedImageAccessStream::kGL);
if (!backing)
return nullptr;
auto real_rep = backing->ProduceGLTexturePassthrough(manager, tracker);
if (!real_rep)
return nullptr;
return std::make_unique<
WrappedGLTexturePassthroughCompoundImageRepresentation>(
manager, this, tracker, std::move(real_rep));
}
std::unique_ptr<SkiaGaneshImageRepresentation>
CompoundImageBacking::ProduceSkiaGanesh(
SharedImageManager* manager,
MemoryTypeTracker* tracker,
scoped_refptr<SharedContextState> context_state) {
auto* backing = GetOrAllocateBacking(SharedImageAccessStream::kSkia);
if (!backing)
return nullptr;
auto real_rep = backing->ProduceSkiaGanesh(manager, tracker, context_state);
if (!real_rep)
return nullptr;
auto* gr_context = context_state ? context_state->gr_context() : nullptr;
return std::make_unique<WrappedSkiaGaneshCompoundImageRepresentation>(
gr_context, manager, this, tracker, std::move(real_rep));
}
std::unique_ptr<SkiaGraphiteImageRepresentation>
CompoundImageBacking::ProduceSkiaGraphite(
SharedImageManager* manager,
MemoryTypeTracker* tracker,
scoped_refptr<SharedContextState> context_state) {
auto* backing = GetOrAllocateBacking(SharedImageAccessStream::kSkia);
if (!backing) {
return nullptr;
}
auto real_rep = backing->ProduceSkiaGraphite(manager, tracker, context_state);
if (!real_rep) {
return nullptr;
}
return std::make_unique<WrappedSkiaGraphiteCompoundImageRepresentation>(
manager, this, tracker, std::move(real_rep));
}
std::unique_ptr<OverlayImageRepresentation>
CompoundImageBacking::ProduceOverlay(SharedImageManager* manager,
MemoryTypeTracker* tracker) {
auto* backing = GetOrAllocateBacking(SharedImageAccessStream::kOverlay);
if (!backing)
return nullptr;
auto real_rep = backing->ProduceOverlay(manager, tracker);
if (!real_rep)
return nullptr;
return std::make_unique<WrappedOverlayCompoundImageRepresentation>(
manager, this, tracker, std::move(real_rep));
}
base::trace_event::MemoryAllocatorDump* CompoundImageBacking::OnMemoryDump(
const std::string& dump_name,
base::trace_event::MemoryAllocatorDumpGuid client_guid,
base::trace_event::ProcessMemoryDump* pmd,
uint64_t client_tracing_id) {
// Create dump but don't add scalar size. The size will be inferred from the
// sizes of the sub-backings.
base::trace_event::MemoryAllocatorDump* dump =
pmd->CreateAllocatorDump(dump_name);
dump->AddString("type", "", GetName());
dump->AddString("dimensions", "", size().ToString());
dump->AddString("format", "", format().ToString());
dump->AddString("usage", "", CreateLabelForSharedImageUsage(usage()));
// Add ownership edge to `client_guid` which expresses shared ownership with
// the client process for the top level dump.
pmd->CreateSharedGlobalAllocatorDump(client_guid);
pmd->AddOwnershipEdge(dump->guid(), client_guid,
static_cast<int>(TracingImportance::kNotOwner));
// Add dumps nested under `dump_name` for child backings owned by compound
// image. These get different shared GUIDs to add ownership edges with GPU
// texture or shared memory.
for (int i = 0; i < static_cast<int>(elements_.size()); ++i) {
auto* backing = elements_[i].backing.get();
if (!backing)
continue;
auto element_client_guid = GetSubBackingGUIDForTracing(mailbox(), i + 1);
std::string element_dump_name =
base::StringPrintf("%s/element_%d", dump_name.c_str(), i);
backing->OnMemoryDump(element_dump_name, element_client_guid, pmd,
client_tracing_id);
}
return dump;
}
const std::vector<SkPixmap>& CompoundImageBacking::GetSharedMemoryPixmaps() {
auto* shm_backing = GetShmElement().GetBacking();
DCHECK(shm_backing);
return static_cast<SharedMemoryImageBacking*>(shm_backing)->pixmaps();
}
CompoundImageBacking::ElementHolder& CompoundImageBacking::GetShmElement() {
for (auto& element : elements_) {
// There should be exactly one element where this returns true.
if (element.access_streams.Has(SharedImageAccessStream::kMemory)) {
return element;
}
}
NOTREACHED();
}
CompoundImageBacking::ElementHolder* CompoundImageBacking::GetElement(
const SharedImageBacking* backing) {
for (auto& element : elements_) {
if (element.GetBacking() == backing) {
return &element;
}
}
return nullptr;
}
CompoundImageBacking::ElementHolder*
CompoundImageBacking::GetElementWithLatestContent() {
// Note that for now iterating over all elements should be fine since we would
// likely not ever had too many backings existing concurrently. We can
// optimize this code by using better algorithm or more suitable data
// structure later if needed.
for (auto& element : elements_) {
if (element.content_id_ == latest_content_id_ && element.GetBacking()) {
return &element;
}
}
return nullptr;
}
SharedImageBacking* CompoundImageBacking::GetOrAllocateBacking(
SharedImageAccessStream stream) {
ElementHolder* best_match = nullptr;
ElementHolder* any_match = nullptr;
// Note that for now iterating over all elements should be fine since we would
// likely not ever had too many backings existing concurrently. We can
// optimize this code by using better algorithm or more suitable data
// structure later if needed.
for (auto& element : elements_) {
if (element.access_streams.Has(stream) && element.GetBacking()) {
if (element.content_id_ == latest_content_id_) {
best_match = &element;
break;
}
if (!any_match) {
any_match = &element;
}
}
}
ElementHolder* target_element = best_match ? best_match : any_match;
if (target_element) {
return target_element->GetBacking();
}
// If no backing is found, we will try to create a new one. This feature is
// disabled by default currently until SharedImageCopyManager is fully ready
// to support all the existing gpu-gpu copy usages.
if (base::FeatureList::IsEnabled(kUseDynamicBackingAllocations) &&
shared_image_factory_) {
SharedImageUsageSet usage = GetUsageFromAccessStream(stream);
auto* gpu_backing_factory = shared_image_factory_->GetFactoryByUsage(
usage, format(), size(),
/*pixel_data=*/{}, gfx::EMPTY_BUFFER);
if (gpu_backing_factory) {
ElementHolder element;
element.access_streams.Put(stream);
CreateBackingFromBackingFactory(gpu_backing_factory->GetWeakPtr(),
debug_label(), element.backing);
if (element.backing) {
elements_.push_back(std::move(element));
return elements_.back().GetBacking();
}
}
}
LOG(ERROR) << "Could not find or create a backing for representation.";
return nullptr;
}
SharedImageBacking* CompoundImageBacking::GetGpuBacking() {
for (auto& element : elements_) {
if (!element.access_streams.Has(SharedImageAccessStream::kMemory)) {
return element.GetBacking();
}
}
LOG(ERROR) << "No GPU backing found.";
return nullptr;
}
void CompoundImageBacking::CreateBackingFromBackingFactory(
base::WeakPtr<SharedImageBackingFactory> factory,
std::string debug_label,
std::unique_ptr<SharedImageBacking>& backing) {
if (!factory) {
DLOG(ERROR) << "Can't allocate backing after image has been destroyed";
return;
}
backing = factory->CreateSharedImage(
mailbox(), format(), kNullSurfaceHandle, size(), color_space(),
surface_origin(), alpha_type(),
GetGpuSharedImageUsage(SharedImageUsageSet(usage())),
std::move(debug_label), /*is_thread_safe=*/false);
if (!backing) {
DLOG(ERROR) << "Failed to allocate GPU backing";
return;
}
// Since the owned GPU backing is never registered with SharedImageManager
// it's not recorded in UMA histogram there.
UMA_HISTOGRAM_ENUMERATION("GPU.SharedImage.BackingType", backing->GetType());
backing->SetNotRefCounted();
backing->SetCleared();
// Update peak GPU memory tracking with the new estimated size.
size_t estimated_size = 0;
for (auto& element : elements_) {
if (element.backing)
estimated_size += element.backing->GetEstimatedSize();
}
AutoLock auto_lock(this);
UpdateEstimatedSize(estimated_size);
}
bool CompoundImageBacking::HasLatestContent(ElementHolder& element) {
return element.content_id_ == latest_content_id_;
}
void CompoundImageBacking::OnAddSecondaryReference() {
// When client adds a reference from another processes it expects this
// SharedImage can outlive original factory ref and so potentially
// SharedimageFactory. We should create all backings now as we might not have
// access to corresponding SharedImageBackingFactories later.
for (auto& element : elements_) {
element.CreateBackingIfNecessary();
}
}
CompoundImageBacking::ElementHolder::ElementHolder() = default;
CompoundImageBacking::ElementHolder::~ElementHolder() = default;
CompoundImageBacking::ElementHolder::ElementHolder(ElementHolder&& other) =
default;
CompoundImageBacking::ElementHolder&
CompoundImageBacking::ElementHolder::operator=(ElementHolder&& other) = default;
void CompoundImageBacking::ElementHolder::CreateBackingIfNecessary() {
if (create_callback) {
std::move(create_callback).Run(backing);
}
}
SharedImageBacking* CompoundImageBacking::ElementHolder::GetBacking() {
CreateBackingIfNecessary();
return backing.get();
}
} // namespace gpu