| // Copyright 2018 The Chromium Authors. All rights reserved. |
| // Use of this source code is governed by a BSD-style license that can be |
| // found in the LICENSE file. |
| |
| #ifndef GPU_COMMAND_BUFFER_SERVICE_SHARED_CONTEXT_STATE_H_ |
| #define GPU_COMMAND_BUFFER_SERVICE_SHARED_CONTEXT_STATE_H_ |
| |
| #include <memory> |
| #include <vector> |
| |
| #include "base/containers/mru_cache.h" |
| #include "base/memory/memory_pressure_listener.h" |
| #include "base/memory/ref_counted.h" |
| #include "base/memory/weak_ptr.h" |
| #include "base/observer_list.h" |
| #include "base/task/single_thread_task_runner_forward.h" |
| #include "base/time/time.h" |
| #include "base/trace_event/memory_dump_provider.h" |
| #include "build/build_config.h" |
| #include "gpu/command_buffer/common/constants.h" |
| #include "gpu/command_buffer/common/gl2_types.h" |
| #include "gpu/command_buffer/common/skia_utils.h" |
| #include "gpu/command_buffer/service/gl_context_virtual_delegate.h" |
| #include "gpu/command_buffer/service/gr_cache_controller.h" |
| #include "gpu/command_buffer/service/memory_tracking.h" |
| #include "gpu/config/gpu_preferences.h" |
| #include "gpu/gpu_gles2_export.h" |
| #include "gpu/ipc/common/command_buffer_id.h" |
| #include "gpu/ipc/common/gpu_peak_memory.h" |
| #include "gpu/vulkan/buildflags.h" |
| #include "third_party/abseil-cpp/absl/types/optional.h" |
| #include "third_party/skia/include/core/SkSurface.h" |
| #include "third_party/skia/include/gpu/GrDirectContext.h" |
| #include "ui/gl/progress_reporter.h" |
| |
| namespace gl { |
| class GLContext; |
| class GLShareGroup; |
| class GLSurface; |
| } // namespace gl |
| |
| namespace viz { |
| class DawnContextProvider; |
| class MetalContextProvider; |
| class VulkanContextProvider; |
| } // namespace viz |
| |
| namespace gpu { |
| class ExternalSemaphorePool; |
| class GpuDriverBugWorkarounds; |
| class GpuProcessActivityFlags; |
| class ServiceTransferCache; |
| |
| namespace gles2 { |
| class FeatureInfo; |
| struct ContextState; |
| } // namespace gles2 |
| |
| namespace raster { |
| class GrShaderCache; |
| class RasterDecoderTestBase; |
| } // namespace raster |
| |
| class GPU_GLES2_EXPORT SharedContextState |
| : public base::trace_event::MemoryDumpProvider, |
| public gpu::GLContextVirtualDelegate, |
| public base::RefCounted<SharedContextState>, |
| public GrContextOptions::ShaderErrorHandler { |
| public: |
| using ContextLostCallback = base::OnceCallback<void(bool)>; |
| |
| // TODO(vikassoni): Refactor code to have seperate constructor for GL and |
| // Vulkan and not initialize/use GL related info for vulkan and vice-versa. |
| SharedContextState( |
| scoped_refptr<gl::GLShareGroup> share_group, |
| scoped_refptr<gl::GLSurface> surface, |
| scoped_refptr<gl::GLContext> context, |
| bool use_virtualized_gl_contexts, |
| ContextLostCallback context_lost_callback, |
| GrContextType gr_context_type = GrContextType::kGL, |
| viz::VulkanContextProvider* vulkan_context_provider = nullptr, |
| viz::MetalContextProvider* metal_context_provider = nullptr, |
| viz::DawnContextProvider* dawn_context_provider = nullptr, |
| base::WeakPtr<gpu::MemoryTracker::Observer> peak_memory_monitor = |
| nullptr); |
| |
| SharedContextState(const SharedContextState&) = delete; |
| SharedContextState& operator=(const SharedContextState&) = delete; |
| |
| bool InitializeGrContext(const GpuPreferences& gpu_preferences, |
| const GpuDriverBugWorkarounds& workarounds, |
| gpu::raster::GrShaderCache* cache, |
| GpuProcessActivityFlags* activity_flags = nullptr, |
| gl::ProgressReporter* progress_reporter = nullptr); |
| bool GrContextIsGL() const { |
| return gr_context_type_ == GrContextType::kGL; |
| } |
| bool GrContextIsVulkan() const { |
| return gr_context_type_ == GrContextType::kVulkan; |
| } |
| bool GrContextIsMetal() const { |
| return gr_context_type_ == GrContextType::kMetal; |
| } |
| bool GrContextIsDawn() const { |
| return gr_context_type_ == GrContextType::kDawn; |
| } |
| |
| bool InitializeGL(const GpuPreferences& gpu_preferences, |
| scoped_refptr<gles2::FeatureInfo> feature_info); |
| bool IsGLInitialized() const { return !!feature_info_; } |
| |
| bool MakeCurrent(gl::GLSurface* surface, bool needs_gl = false); |
| void ReleaseCurrent(gl::GLSurface* surface); |
| void MarkContextLost(error::ContextLostReason reason = error::kUnknown); |
| bool IsCurrent(gl::GLSurface* surface, bool needs_gl = false); |
| |
| void PurgeMemory( |
| base::MemoryPressureListener::MemoryPressureLevel memory_pressure_level); |
| |
| void UpdateSkiaOwnedMemorySize(); |
| uint64_t GetMemoryUsage(); |
| |
| void PessimisticallyResetGrContext() const; |
| |
| void StoreVkPipelineCacheIfNeeded(); |
| |
| gl::GLShareGroup* share_group() { return share_group_.get(); } |
| gl::GLContext* context() { return context_.get(); } |
| gl::GLContext* real_context() { return real_context_.get(); } |
| gl::GLSurface* surface() { return surface_.get(); } |
| viz::VulkanContextProvider* vk_context_provider() { |
| return vk_context_provider_; |
| } |
| viz::MetalContextProvider* metal_context_provider() { |
| return metal_context_provider_; |
| } |
| viz::DawnContextProvider* dawn_context_provider() { |
| return dawn_context_provider_; |
| } |
| gl::ProgressReporter* progress_reporter() const { return progress_reporter_; } |
| GrDirectContext* gr_context() { return gr_context_; } |
| GrContextType gr_context_type() const { return gr_context_type_; } |
| // Handles Skia-reported shader compilation errors. |
| void compileError(const char* shader, const char* errors) override; |
| gles2::FeatureInfo* feature_info() { return feature_info_.get(); } |
| gles2::ContextState* context_state() const { return context_state_.get(); } |
| bool context_lost() const { return !!context_lost_reason_; } |
| absl::optional<error::ContextLostReason> context_lost_reason() { |
| return context_lost_reason_; |
| } |
| bool need_context_state_reset() const { return need_context_state_reset_; } |
| void set_need_context_state_reset(bool reset) { |
| need_context_state_reset_ = reset; |
| } |
| ServiceTransferCache* transfer_cache() { return transfer_cache_.get(); } |
| std::vector<uint8_t>* scratch_deserialization_buffer() { |
| return &scratch_deserialization_buffer_; |
| } |
| bool use_virtualized_gl_contexts() const { |
| return use_virtualized_gl_contexts_; |
| } |
| bool support_vulkan_external_object() const { |
| return support_vulkan_external_object_; |
| } |
| bool support_gl_external_object_flags() const { |
| return support_gl_external_object_flags_; |
| } |
| gpu::MemoryTracker::Observer* memory_tracker_observer() { |
| return &memory_tracker_observer_; |
| } |
| gpu::MemoryTracker* memory_tracker() { return &memory_tracker_; } |
| gpu::MemoryTypeTracker* memory_type_tracker() { |
| return &memory_type_tracker_; |
| } |
| ExternalSemaphorePool* external_semaphore_pool() { |
| #if BUILDFLAG(ENABLE_VULKAN) |
| return external_semaphore_pool_.get(); |
| #else |
| return nullptr; |
| #endif |
| } |
| |
| // base::trace_event::MemoryDumpProvider implementation. |
| bool OnMemoryDump(const base::trace_event::MemoryDumpArgs& args, |
| base::trace_event::ProcessMemoryDump* pmd) override; |
| |
| // Observer class which is notified when the context is lost. |
| class ContextLostObserver { |
| public: |
| virtual void OnContextLost() = 0; |
| |
| protected: |
| virtual ~ContextLostObserver() = default; |
| }; |
| void AddContextLostObserver(ContextLostObserver* obs); |
| void RemoveContextLostObserver(ContextLostObserver* obs); |
| |
| // Creating a SkSurface backed by FBO takes ~500 usec and holds ~50KB of heap |
| // on Android circa 2020. Caching them is a memory/CPU tradeoff. |
| void CacheSkSurface(void* key, sk_sp<SkSurface> surface) { |
| sk_surface_cache_.Put(key, surface); |
| } |
| sk_sp<SkSurface> GetCachedSkSurface(void* key) { |
| auto found = sk_surface_cache_.Get(key); |
| if (found == sk_surface_cache_.end()) |
| return nullptr; |
| return found->second; |
| } |
| void EraseCachedSkSurface(void* key) { |
| auto found = sk_surface_cache_.Peek(key); |
| if (found != sk_surface_cache_.end()) |
| sk_surface_cache_.Erase(found); |
| } |
| // Supports DCHECKs. OK to be approximate. |
| bool CachedSkSurfaceIsUnique(void* key) { |
| auto found = sk_surface_cache_.Peek(key); |
| // It was purged. Assume it was unique. |
| if (found == sk_surface_cache_.end()) |
| return true; |
| return found->second->unique(); |
| } |
| |
| // Updates |context_lost_reason| and returns true if lost |
| // (e.g. VK_ERROR_DEVICE_LOST or GL_UNKNOWN_CONTEXT_RESET_ARB). |
| bool CheckResetStatus(bool needs_gl); |
| bool device_needs_reset() { return device_needs_reset_; } |
| |
| void ScheduleGrContextCleanup(); |
| |
| private: |
| friend class base::RefCounted<SharedContextState>; |
| friend class raster::RasterDecoderTestBase; |
| |
| // Observer which is notified when SkiaOutputSurfaceImpl takes ownership of a |
| // shared image, and forward information to both histograms and task manager. |
| class GPU_GLES2_EXPORT MemoryTrackerObserver |
| : public gpu::MemoryTracker::Observer { |
| public: |
| explicit MemoryTrackerObserver( |
| base::WeakPtr<gpu::MemoryTracker::Observer> peak_memory_monitor); |
| MemoryTrackerObserver(MemoryTrackerObserver&) = delete; |
| MemoryTrackerObserver& operator=(MemoryTrackerObserver&) = delete; |
| ~MemoryTrackerObserver() override; |
| |
| // gpu::MemoryTracker::Observer implementation: |
| void OnMemoryAllocatedChange( |
| CommandBufferId id, |
| uint64_t old_size, |
| uint64_t new_size, |
| GpuPeakMemoryAllocationSource source = |
| GpuPeakMemoryAllocationSource::UNKNOWN) override; |
| |
| // Reports to GpuServiceImpl::GetVideoMemoryUsageStats() |
| uint64_t GetMemoryUsage() const { return size_; } |
| |
| private: |
| uint64_t size_ = 0; |
| base::WeakPtr<gpu::MemoryTracker::Observer> const peak_memory_monitor_; |
| }; |
| |
| // MemoryTracker implementation used to track SharedImages owned by |
| // SkiaOutputSurfaceImpl. |
| class MemoryTracker : public gpu::MemoryTracker { |
| public: |
| explicit MemoryTracker(gpu::MemoryTracker::Observer* observer); |
| MemoryTracker(const MemoryTracker&) = delete; |
| MemoryTracker& operator=(const MemoryTracker&) = delete; |
| ~MemoryTracker() override; |
| |
| // MemoryTracker implementation: |
| void TrackMemoryAllocatedChange(int64_t delta) override; |
| uint64_t GetSize() const override; |
| uint64_t ClientTracingId() const override; |
| int ClientId() const override; |
| uint64_t ContextGroupTracingId() const override; |
| |
| private: |
| gpu::CommandBufferId command_buffer_id_; |
| const uint64_t client_tracing_id_; |
| gpu::MemoryTracker::Observer* const observer_; |
| uint64_t size_ = 0; |
| }; |
| |
| ~SharedContextState() override; |
| |
| absl::optional<error::ContextLostReason> GetResetStatus(bool needs_gl); |
| |
| // gpu::GLContextVirtualDelegate implementation. |
| bool initialized() const override; |
| const gles2::ContextState* GetContextState() override; |
| void RestoreState(const gles2::ContextState* prev_state) override; |
| void RestoreGlobalState() const override; |
| void ClearAllAttributes() const override; |
| void RestoreActiveTexture() const override; |
| void RestoreAllTextureUnitAndSamplerBindings( |
| const gles2::ContextState* prev_state) const override; |
| void RestoreActiveTextureUnitBinding(unsigned int target) const override; |
| void RestoreBufferBinding(unsigned int target) override; |
| void RestoreBufferBindings() const override; |
| void RestoreFramebufferBindings() const override; |
| void RestoreRenderbufferBindings() override; |
| void RestoreProgramBindings() const override; |
| void RestoreTextureUnitBindings(unsigned unit) const override; |
| void RestoreVertexAttribArray(unsigned index) override; |
| void RestoreAllExternalTextureBindingsIfNeeded() override; |
| QueryManager* GetQueryManager() override; |
| |
| bool use_virtualized_gl_contexts_ = false; |
| bool support_vulkan_external_object_ = false; |
| bool support_gl_external_object_flags_ = false; |
| ContextLostCallback context_lost_callback_; |
| GrContextType gr_context_type_ = GrContextType::kGL; |
| MemoryTrackerObserver memory_tracker_observer_; |
| MemoryTracker memory_tracker_; |
| gpu::MemoryTypeTracker memory_type_tracker_; |
| viz::VulkanContextProvider* const vk_context_provider_; |
| viz::MetalContextProvider* const metal_context_provider_; |
| viz::DawnContextProvider* const dawn_context_provider_; |
| GrDirectContext* gr_context_ = nullptr; |
| |
| scoped_refptr<gl::GLShareGroup> share_group_; |
| scoped_refptr<gl::GLContext> context_; |
| scoped_refptr<gl::GLContext> real_context_; |
| scoped_refptr<gl::GLSurface> surface_; |
| |
| // Most recent surface that this ShareContextState was made current with. |
| // Avoids a call to MakeCurrent with a different surface, if we don't |
| // care which surface is current. |
| gl::GLSurface* last_current_surface_ = nullptr; |
| |
| scoped_refptr<gles2::FeatureInfo> feature_info_; |
| |
| // raster decoders and display compositor share this context_state_. |
| std::unique_ptr<gles2::ContextState> context_state_; |
| |
| gl::ProgressReporter* progress_reporter_ = nullptr; |
| sk_sp<GrDirectContext> owned_gr_context_; |
| std::unique_ptr<ServiceTransferCache> transfer_cache_; |
| uint64_t skia_gr_cache_size_ = 0; |
| std::vector<uint8_t> scratch_deserialization_buffer_; |
| gpu::raster::GrShaderCache* gr_shader_cache_ = nullptr; |
| |
| // |need_context_state_reset| is set whenever Skia may have altered the |
| // driver's GL state. |
| bool need_context_state_reset_ = false; |
| |
| absl::optional<error::ContextLostReason> context_lost_reason_; |
| base::ObserverList<ContextLostObserver>::Unchecked context_lost_observers_; |
| |
| base::MRUCache<void*, sk_sp<SkSurface>> sk_surface_cache_; |
| |
| bool device_needs_reset_ = false; |
| base::Time last_gl_check_graphics_reset_status_; |
| bool disable_check_reset_status_throttling_for_test_ = false; |
| |
| #if BUILDFLAG(ENABLE_VULKAN) |
| std::unique_ptr<ExternalSemaphorePool> external_semaphore_pool_; |
| #endif |
| |
| absl::optional<raster::GrCacheController> gr_cache_controller_; |
| |
| base::WeakPtrFactory<SharedContextState> weak_ptr_factory_{this}; |
| }; |
| |
| } // namespace gpu |
| |
| #endif // GPU_COMMAND_BUFFER_SERVICE_SHARED_CONTEXT_STATE_H_ |