| // Copyright (c) 2013 The Chromium Authors. All rights reserved. |
| // Use of this source code is governed by a BSD-style license that can be |
| // found in the LICENSE file. |
| |
| #include "media/gpu/android_video_decode_accelerator.h" |
| |
| #include <stddef.h> |
| |
| #include <memory> |
| |
| #include "base/android/build_info.h" |
| #include "base/auto_reset.h" |
| #include "base/bind.h" |
| #include "base/bind_helpers.h" |
| #include "base/callback_helpers.h" |
| #include "base/command_line.h" |
| #include "base/lazy_instance.h" |
| #include "base/logging.h" |
| #include "base/metrics/histogram.h" |
| #include "base/task_runner_util.h" |
| #include "base/threading/thread_checker.h" |
| #include "base/threading/thread_task_runner_handle.h" |
| #include "base/trace_event/trace_event.h" |
| #include "gpu/command_buffer/service/gles2_cmd_decoder.h" |
| #include "gpu/command_buffer/service/mailbox_manager.h" |
| #include "gpu/ipc/service/gpu_channel.h" |
| #include "media/base/android/media_codec_bridge.h" |
| #include "media/base/android/media_codec_util.h" |
| #include "media/base/bind_to_current_loop.h" |
| #include "media/base/bitstream_buffer.h" |
| #include "media/base/limits.h" |
| #include "media/base/media.h" |
| #include "media/base/timestamp_constants.h" |
| #include "media/base/video_decoder_config.h" |
| #include "media/gpu/android_copying_backing_strategy.h" |
| #include "media/gpu/android_deferred_rendering_backing_strategy.h" |
| #include "media/gpu/avda_return_on_failure.h" |
| #include "media/gpu/shared_memory_region.h" |
| #include "media/video/picture.h" |
| #include "ui/gl/android/scoped_java_surface.h" |
| #include "ui/gl/android/surface_texture.h" |
| #include "ui/gl/gl_bindings.h" |
| |
| #if defined(ENABLE_MOJO_MEDIA_IN_GPU_PROCESS) |
| #include "media/mojo/services/mojo_cdm_service.h" |
| #endif |
| |
| #define POST_ERROR(error_code, error_message) \ |
| do { \ |
| DLOG(ERROR) << error_message; \ |
| PostError(FROM_HERE, VideoDecodeAccelerator::error_code); \ |
| } while (0) |
| |
| namespace media { |
| |
| enum { kNumPictureBuffers = limits::kMaxVideoFrames + 1 }; |
| |
| // Max number of bitstreams notified to the client with |
| // NotifyEndOfBitstreamBuffer() before getting output from the bitstream. |
| enum { kMaxBitstreamsNotifiedInAdvance = 32 }; |
| |
| // MediaCodec is only guaranteed to support baseline, but some devices may |
| // support others. Advertise support for all H264 profiles and let the |
| // MediaCodec fail when decoding if it's not actually supported. It's assumed |
| // that consumers won't have software fallback for H264 on Android anyway. |
| static const VideoCodecProfile kSupportedH264Profiles[] = { |
| H264PROFILE_BASELINE, |
| H264PROFILE_MAIN, |
| H264PROFILE_EXTENDED, |
| H264PROFILE_HIGH, |
| H264PROFILE_HIGH10PROFILE, |
| H264PROFILE_HIGH422PROFILE, |
| H264PROFILE_HIGH444PREDICTIVEPROFILE, |
| H264PROFILE_SCALABLEBASELINE, |
| H264PROFILE_SCALABLEHIGH, |
| H264PROFILE_STEREOHIGH, |
| H264PROFILE_MULTIVIEWHIGH}; |
| |
| // Because MediaCodec is thread-hostile (must be poked on a single thread) and |
| // has no callback mechanism (b/11990118), we must drive it by polling for |
| // complete frames (and available input buffers, when the codec is fully |
| // saturated). This function defines the polling delay. The value used is an |
| // arbitrary choice that trades off CPU utilization (spinning) against latency. |
| // Mirrors android_video_encode_accelerator.cc:EncodePollDelay(). |
| // |
| // An alternative to this polling scheme could be to dedicate a new thread |
| // (instead of using the ChildThread) to run the MediaCodec, and make that |
| // thread use the timeout-based flavor of MediaCodec's dequeue methods when it |
| // believes the codec should complete "soon" (e.g. waiting for an input |
| // buffer, or waiting for a picture when it knows enough complete input |
| // pictures have been fed to saturate any internal buffering). This is |
| // speculative and it's unclear that this would be a win (nor that there's a |
| // reasonably device-agnostic way to fill in the "believes" above). |
| constexpr base::TimeDelta DecodePollDelay = |
| base::TimeDelta::FromMilliseconds(10); |
| |
| constexpr base::TimeDelta NoWaitTimeOut = base::TimeDelta::FromMicroseconds(0); |
| |
| constexpr base::TimeDelta IdleTimerTimeOut = base::TimeDelta::FromSeconds(1); |
| |
| // Time between when we notice an error, and when we actually notify somebody. |
| // This is to prevent codec errors caused by SurfaceView fullscreen transitions |
| // from breaking the pipeline, if we're about to be reset anyway. |
| constexpr base::TimeDelta ErrorPostingDelay = base::TimeDelta::FromSeconds(2); |
| |
| // For RecordFormatChangedMetric. |
| enum FormatChangedValue { |
| CodecInitialized = false, |
| MissingFormatChanged = true |
| }; |
| |
| // Maximum number of concurrent, incomplete codec creations that we'll allow |
| // before turning off autodection of codec type. |
| enum { kMaxConcurrentCodecAutodetections = 4 }; |
| |
| static inline void RecordFormatChangedMetric(FormatChangedValue value) { |
| UMA_HISTOGRAM_BOOLEAN("Media.AVDA.MissingFormatChanged", !!value); |
| } |
| |
| // Handle OnFrameAvailable callbacks safely. Since they occur asynchronously, |
| // we take care that the AVDA that wants them still exists. A WeakPtr to |
| // the AVDA would be preferable, except that OnFrameAvailable callbacks can |
| // occur off the gpu main thread. We also can't guarantee when the |
| // SurfaceTexture will quit sending callbacks to coordinate with the |
| // destruction of the AVDA, so we have a separate object that the cb can own. |
| class AndroidVideoDecodeAccelerator::OnFrameAvailableHandler |
| : public base::RefCountedThreadSafe<OnFrameAvailableHandler> { |
| public: |
| // We do not retain ownership of |owner|. It must remain valid until |
| // after ClearOwner() is called. This will register with |
| // |surface_texture| to receive OnFrameAvailable callbacks. |
| OnFrameAvailableHandler( |
| AndroidVideoDecodeAccelerator* owner, |
| const scoped_refptr<gl::SurfaceTexture>& surface_texture) |
| : owner_(owner) { |
| // Note that the callback owns a strong ref to us. |
| surface_texture->SetFrameAvailableCallbackOnAnyThread( |
| base::Bind(&OnFrameAvailableHandler::OnFrameAvailable, |
| scoped_refptr<OnFrameAvailableHandler>(this))); |
| } |
| |
| // Forget about our owner, which is required before one deletes it. |
| // No further callbacks will happen once this completes. |
| void ClearOwner() { |
| base::AutoLock lock(lock_); |
| // No callback can happen until we release the lock. |
| owner_ = nullptr; |
| } |
| |
| // Call back into our owner if it hasn't been deleted. |
| void OnFrameAvailable() { |
| base::AutoLock auto_lock(lock_); |
| // |owner_| can't be deleted while we have the lock. |
| if (owner_) |
| owner_->OnFrameAvailable(); |
| } |
| |
| private: |
| friend class base::RefCountedThreadSafe<OnFrameAvailableHandler>; |
| virtual ~OnFrameAvailableHandler() {} |
| |
| // Protects changes to owner_. |
| base::Lock lock_; |
| |
| // AVDA that wants the OnFrameAvailable callback. |
| AndroidVideoDecodeAccelerator* owner_; |
| |
| DISALLOW_COPY_AND_ASSIGN(OnFrameAvailableHandler); |
| }; |
| |
| // Helper class to share an IO timer for DoIOTask() execution; prevents each |
| // AVDA instance from starting its own high frequency timer. The intuition |
| // behind this is that, if we're waiting for long enough, then either (a) |
| // MediaCodec is broken or (b) MediaCodec is waiting on us to change state |
| // (e.g., get new demuxed data / get a free picture buffer / return an output |
| // buffer to MediaCodec). This is inherently a race, since we don't know if |
| // MediaCodec is broken or just slow. Since the MediaCodec API doesn't let |
| // us wait on MediaCodec state changes prior to L, we more or less have to |
| // time out or keep polling forever in some common cases. |
| class AVDATimerManager { |
| public: |
| // Make sure that the construction thread is started for |avda_instance|. |
| bool StartThread(AndroidVideoDecodeAccelerator* avda_instance) { |
| DCHECK(thread_checker_.CalledOnValidThread()); |
| |
| // If we chose not to shut it down due to pending codec constructions, then |
| // the thread might already be started even if there are no avda instances. |
| // Plus, sometimes we just fail to start the thread. |
| if (!construction_thread_.IsRunning()) { |
| if (!construction_thread_.Start()) { |
| LOG(ERROR) << "Failed to start construction thread."; |
| return false; |
| } |
| } |
| |
| thread_avda_instances_.insert(avda_instance); |
| UMA_HISTOGRAM_ENUMERATION("Media.AVDA.NumAVDAInstances", |
| thread_avda_instances_.size(), |
| 31); // PRESUBMIT_IGNORE_UMA_MAX |
| return true; |
| } |
| |
| // |avda_instance| will no longer need the construction thread. Stop the |
| // thread if this is the last instance. |
| void StopThread(AndroidVideoDecodeAccelerator* avda_instance) { |
| DCHECK(thread_checker_.CalledOnValidThread()); |
| |
| thread_avda_instances_.erase(avda_instance); |
| if (!thread_avda_instances_.empty()) |
| return; |
| |
| // Don't stop the thread if there are outstanding requests, since they |
| // might be hung. They also might simply be incomplete, and the thread |
| // will stay running until we try to shut it down again. |
| base::AutoLock auto_lock(autodetection_info_.lock_); |
| if (autodetection_info_.outstanding_) |
| return; |
| |
| construction_thread_.Stop(); |
| } |
| |
| // Request periodic callback of |avda_instance|->DoIOTask(). Does nothing if |
| // the instance is already registered and the timer started. The first request |
| // will start the repeating timer on an interval of DecodePollDelay. |
| void StartTimer(AndroidVideoDecodeAccelerator* avda_instance) { |
| DCHECK(thread_checker_.CalledOnValidThread()); |
| |
| timer_avda_instances_.insert(avda_instance); |
| |
| // If the timer is running, StopTimer() might have been called earlier, if |
| // so remove the instance from the pending erasures. |
| if (timer_running_) |
| pending_erase_.erase(avda_instance); |
| |
| if (io_timer_.IsRunning()) |
| return; |
| io_timer_.Start(FROM_HERE, DecodePollDelay, this, |
| &AVDATimerManager::RunTimer); |
| } |
| |
| // Stop callbacks to |avda_instance|->DoIOTask(). Does nothing if the instance |
| // is not registered. If there are no instances left, the repeating timer will |
| // be stopped. |
| void StopTimer(AndroidVideoDecodeAccelerator* avda_instance) { |
| DCHECK(thread_checker_.CalledOnValidThread()); |
| |
| // If the timer is running, defer erasures to avoid iterator invalidation. |
| if (timer_running_) { |
| pending_erase_.insert(avda_instance); |
| return; |
| } |
| |
| timer_avda_instances_.erase(avda_instance); |
| if (timer_avda_instances_.empty()) |
| io_timer_.Stop(); |
| } |
| |
| // Eventually, we should run the timer on this thread. For now, we just keep |
| // it as a convenience for construction. |
| scoped_refptr<base::SingleThreadTaskRunner> ConstructionTaskRunner() { |
| DCHECK(thread_checker_.CalledOnValidThread()); |
| return construction_thread_.task_runner(); |
| } |
| |
| // Called on the main thread when the construction thread will be doing work |
| // that can potentially hang (e.g., autodetection). There may be several |
| // calls to this before any call to DoneUsingConstructionThread. |
| // Note that this should only be called from the main thread, else it's a race |
| // with IsCodecAutodetectionProbablySafe. |
| void StartUsingConstructionThread() { |
| DCHECK(thread_checker_.CalledOnValidThread()); |
| base::AutoLock auto_lock(autodetection_info_.lock_); |
| ++autodetection_info_.outstanding_; |
| } |
| |
| // Called on any thread after the potentially dangerous construction thread |
| // work completes safely. May be called on any thread, including the |
| // construction thread. |
| // This assumes that requests are ordered, so please don't mix sync and async |
| // codec construction here. |
| void DoneUsingConstructionThread() { |
| base::AutoLock auto_lock_l(autodetection_info_.lock_); |
| DCHECK_GT(autodetection_info_.outstanding_, 0); |
| --autodetection_info_.outstanding_; |
| } |
| |
| // Return a hint about whether autodetecting the codec type is safe or not. |
| bool IsCodecAutodetectionProbablySafe() { |
| base::AutoLock auto_lock_l(autodetection_info_.lock_); |
| |
| return autodetection_info_.outstanding_ < kMaxConcurrentCodecAutodetections; |
| } |
| |
| // |avda| would like to use |surface_id|. If it is not busy, then mark it |
| // as busy and return true. If it is busy, then replace any existing waiter, |
| // make |avda| the current waiter, and return false. Any existing waiter |
| // is assumed to be on the way out, so we fail its allocation request. |
| bool AllocateSurface(int surface_id, AndroidVideoDecodeAccelerator* avda) { |
| // Nobody has to wait for no surface. |
| if (surface_id == AndroidVideoDecodeAccelerator::Config::kNoSurfaceID) |
| return true; |
| |
| auto iter = surface_waiter_map_.find(surface_id); |
| if (iter == surface_waiter_map_.end()) { |
| // SurfaceView isn't allocated. Succeed. |
| surface_waiter_map_[surface_id].owner = avda; |
| return true; |
| } |
| |
| // SurfaceView is already allocated. |
| if (iter->second.waiter) { |
| // Some other AVDA is waiting. |avda| will replace it, so notify it |
| // that it will fail. |
| iter->second.waiter->OnSurfaceAvailable(false); |
| iter->second.waiter = nullptr; |
| } |
| |
| // |avda| is now waiting. |
| iter->second.waiter = avda; |
| return false; |
| } |
| |
| // Clear any waiting request for |surface_id| by |avda|. It is okay if |
| // |waiter| is not waiting and/or isn't the owner of |surface_id|. |
| void DeallocateSurface(int surface_id, AndroidVideoDecodeAccelerator* avda) { |
| SurfaceWaiterMap::iterator iter = surface_waiter_map_.find(surface_id); |
| if (iter == surface_waiter_map_.end()) |
| return; |
| |
| // If |avda| was waiting, then remove it without OnSurfaceAvailable. |
| if (iter->second.waiter == avda) |
| iter->second.waiter = nullptr; |
| |
| // If |avda| is the owner, then let the waiter have it. |
| if (iter->second.owner != avda) |
| return; |
| |
| AndroidVideoDecodeAccelerator* waiter = iter->second.waiter; |
| if (!waiter) { |
| // No waiter -- remove the record and return explicitly since |iter| is |
| // no longer valid. |
| surface_waiter_map_.erase(iter); |
| return; |
| } |
| |
| // Promote |waiter| to be the owner. |
| iter->second.owner = waiter; |
| iter->second.waiter = nullptr; |
| waiter->OnSurfaceAvailable(true); |
| } |
| |
| private: |
| friend struct base::DefaultLazyInstanceTraits<AVDATimerManager>; |
| |
| AVDATimerManager() : construction_thread_("AVDAThread") {} |
| ~AVDATimerManager() { NOTREACHED(); } |
| |
| void RunTimer() { |
| { |
| // Call out to all AVDA instances, some of which may attempt to remove |
| // themselves from the list during this operation; those removals will be |
| // deferred until after all iterations are complete. |
| base::AutoReset<bool> scoper(&timer_running_, true); |
| for (auto* avda : timer_avda_instances_) |
| avda->DoIOTask(false); |
| } |
| |
| // Take care of any deferred erasures. |
| for (auto* avda : pending_erase_) |
| StopTimer(avda); |
| pending_erase_.clear(); |
| |
| // TODO(dalecurtis): We may want to consider chunking this if task execution |
| // takes too long for the combined timer. |
| } |
| |
| // All AVDA instances that would like us to poll DoIOTask. |
| std::set<AndroidVideoDecodeAccelerator*> timer_avda_instances_; |
| |
| // All AVDA instances that might like to use the construction thread. |
| std::set<AndroidVideoDecodeAccelerator*> thread_avda_instances_; |
| |
| struct OwnerRecord { |
| AndroidVideoDecodeAccelerator* owner = nullptr; |
| AndroidVideoDecodeAccelerator* waiter = nullptr; |
| }; |
| // [surface id] = OwnerRecord for that surface. |
| using SurfaceWaiterMap = std::map<int, OwnerRecord>; |
| SurfaceWaiterMap surface_waiter_map_; |
| |
| // Since we can't delete while iterating when using a set, defer erasure until |
| // after iteration complete. |
| bool timer_running_ = false; |
| std::set<AndroidVideoDecodeAccelerator*> pending_erase_; |
| |
| // Repeating timer responsible for draining pending IO to the codecs. |
| base::RepeatingTimer io_timer_; |
| |
| // Data for determining if codec creation is hanging. |
| struct { |
| // Lock that protects other members of this struct. |
| base::Lock lock_; |
| |
| // Number of currently pending work items of the construction thread. |
| int outstanding_ = 0; |
| } autodetection_info_; |
| |
| base::Thread construction_thread_; |
| |
| base::ThreadChecker thread_checker_; |
| |
| DISALLOW_COPY_AND_ASSIGN(AVDATimerManager); |
| }; |
| |
| static base::LazyInstance<AVDATimerManager>::Leaky g_avda_timer = |
| LAZY_INSTANCE_INITIALIZER; |
| |
| AndroidVideoDecodeAccelerator::CodecConfig::CodecConfig() {} |
| |
| AndroidVideoDecodeAccelerator::CodecConfig::~CodecConfig() {} |
| |
| AndroidVideoDecodeAccelerator::BitstreamRecord::BitstreamRecord( |
| const BitstreamBuffer& bitstream_buffer) |
| : buffer(bitstream_buffer) { |
| if (buffer.id() != -1) |
| memory.reset(new SharedMemoryRegion(buffer, true)); |
| } |
| |
| AndroidVideoDecodeAccelerator::BitstreamRecord::BitstreamRecord( |
| BitstreamRecord&& other) |
| : buffer(std::move(other.buffer)), memory(std::move(other.memory)) {} |
| |
| AndroidVideoDecodeAccelerator::BitstreamRecord::~BitstreamRecord() {} |
| |
| AndroidVideoDecodeAccelerator::AndroidVideoDecodeAccelerator( |
| const MakeGLContextCurrentCallback& make_context_current_cb, |
| const GetGLES2DecoderCallback& get_gles2_decoder_cb) |
| : client_(NULL), |
| make_context_current_cb_(make_context_current_cb), |
| get_gles2_decoder_cb_(get_gles2_decoder_cb), |
| state_(NO_ERROR), |
| picturebuffers_requested_(false), |
| drain_type_(DRAIN_TYPE_NONE), |
| media_drm_bridge_cdm_context_(nullptr), |
| cdm_registration_id_(0), |
| pending_input_buf_index_(-1), |
| error_sequence_token_(0), |
| defer_errors_(false), |
| deferred_initialization_pending_(false), |
| codec_needs_reset_(false), |
| weak_this_factory_(this) {} |
| |
| AndroidVideoDecodeAccelerator::~AndroidVideoDecodeAccelerator() { |
| DCHECK(thread_checker_.CalledOnValidThread()); |
| g_avda_timer.Pointer()->StopTimer(this); |
| g_avda_timer.Pointer()->StopThread(this); |
| |
| #if defined(ENABLE_MOJO_MEDIA_IN_GPU_PROCESS) |
| if (!media_drm_bridge_cdm_context_) |
| return; |
| |
| DCHECK(cdm_registration_id_); |
| |
| // Cancel previously registered callback (if any). |
| media_drm_bridge_cdm_context_->SetMediaCryptoReadyCB( |
| MediaDrmBridgeCdmContext::MediaCryptoReadyCB()); |
| |
| media_drm_bridge_cdm_context_->UnregisterPlayer(cdm_registration_id_); |
| #endif // defined(ENABLE_MOJO_MEDIA_IN_GPU_PROCESS) |
| } |
| |
| bool AndroidVideoDecodeAccelerator::Initialize(const Config& config, |
| Client* client) { |
| DVLOG(1) << __FUNCTION__ << ": " << config.AsHumanReadableString(); |
| TRACE_EVENT0("media", "AVDA::Initialize"); |
| DCHECK(!media_codec_); |
| DCHECK(thread_checker_.CalledOnValidThread()); |
| |
| if (make_context_current_cb_.is_null() || get_gles2_decoder_cb_.is_null()) { |
| DLOG(ERROR) << "GL callbacks are required for this VDA"; |
| return false; |
| } |
| |
| if (config.output_mode != Config::OutputMode::ALLOCATE) { |
| DLOG(ERROR) << "Only ALLOCATE OutputMode is supported by this VDA"; |
| return false; |
| } |
| |
| DCHECK(client); |
| client_ = client; |
| config_ = config; |
| codec_config_ = new CodecConfig(); |
| codec_config_->codec_ = VideoCodecProfileToVideoCodec(config.profile); |
| codec_config_->initial_expected_coded_size_ = |
| config.initial_expected_coded_size; |
| |
| // We signalled that we support deferred initialization, so see if the client |
| // does also. |
| deferred_initialization_pending_ = config.is_deferred_initialization_allowed; |
| |
| if (config_.is_encrypted && !deferred_initialization_pending_) { |
| DLOG(ERROR) << "Deferred initialization must be used for encrypted streams"; |
| return false; |
| } |
| |
| if (codec_config_->codec_ != kCodecVP8 && |
| codec_config_->codec_ != kCodecVP9 && |
| codec_config_->codec_ != kCodecH264) { |
| LOG(ERROR) << "Unsupported profile: " << config.profile; |
| return false; |
| } |
| |
| // Only use MediaCodec for VP8/9 if it's likely backed by hardware |
| // or if the stream is encrypted. |
| if (IsMediaCodecSoftwareDecodingForbidden() && |
| VideoCodecBridge::IsKnownUnaccelerated(codec_config_->codec_, |
| MEDIA_CODEC_DECODER)) { |
| DVLOG(1) << "Initialization failed: " |
| << (codec_config_->codec_ == kCodecVP8 ? "vp8" : "vp9") |
| << " is not hardware accelerated"; |
| return false; |
| } |
| |
| auto gles_decoder = get_gles2_decoder_cb_.Run(); |
| if (!gles_decoder) { |
| LOG(ERROR) << "Failed to get gles2 decoder instance."; |
| return false; |
| } |
| |
| const gpu::GpuPreferences& gpu_preferences = |
| gles_decoder->GetContextGroup()->gpu_preferences(); |
| |
| if (UseDeferredRenderingStrategy(gpu_preferences)) { |
| DVLOG(1) << __FUNCTION__ << ", using deferred rendering strategy."; |
| strategy_.reset(new AndroidDeferredRenderingBackingStrategy(this)); |
| } else { |
| DVLOG(1) << __FUNCTION__ << ", using copy back strategy."; |
| strategy_.reset(new AndroidCopyingBackingStrategy(this)); |
| } |
| |
| if (!make_context_current_cb_.Run()) { |
| LOG(ERROR) << "Failed to make this decoder's GL context current."; |
| return false; |
| } |
| |
| if (g_avda_timer.Pointer()->AllocateSurface(config_.surface_id, this)) { |
| // We have succesfully owned the surface, so finish initialization now. |
| return InitializeStrategy(); |
| } |
| |
| // We have to wait for some other AVDA instance to free up the surface. |
| // OnSurfaceAvailable will be called when it's available. |
| return true; |
| } |
| |
| void AndroidVideoDecodeAccelerator::OnSurfaceAvailable(bool success) { |
| DCHECK(deferred_initialization_pending_); |
| |
| if (!success || !InitializeStrategy()) { |
| NotifyInitializationComplete(false); |
| deferred_initialization_pending_ = false; |
| } |
| } |
| |
| bool AndroidVideoDecodeAccelerator::InitializeStrategy() { |
| codec_config_->surface_ = strategy_->Initialize(config_.surface_id); |
| if (codec_config_->surface_.IsEmpty()) { |
| LOG(ERROR) << "Failed to initialize the backing strategy. The returned " |
| "Java surface is empty."; |
| return false; |
| } |
| |
| on_destroying_surface_cb_ = |
| base::Bind(&AndroidVideoDecodeAccelerator::OnDestroyingSurface, |
| weak_this_factory_.GetWeakPtr()); |
| AVDASurfaceTracker::GetInstance()->RegisterOnDestroyingSurfaceCallback( |
| on_destroying_surface_cb_); |
| |
| // TODO(watk,liberato): move this into the strategy. |
| scoped_refptr<gl::SurfaceTexture> surface_texture = |
| strategy_->GetSurfaceTexture(); |
| if (surface_texture) { |
| on_frame_available_handler_ = |
| new OnFrameAvailableHandler(this, surface_texture); |
| } |
| |
| // Start the thread for async configuration, even if we don't need it now. |
| // ResetCodecState might rebuild the codec later, for example. |
| if (!g_avda_timer.Pointer()->StartThread(this)) { |
| LOG(ERROR) << "Failed to start AVDA thread"; |
| return false; |
| } |
| |
| // If we are encrypted, then we aren't able to create the codec yet. |
| if (config_.is_encrypted) { |
| InitializeCdm(); |
| return true; |
| } |
| |
| if (deferred_initialization_pending_) { |
| ConfigureMediaCodecAsynchronously(); |
| return true; |
| } |
| |
| // If the client doesn't support deferred initialization (WebRTC), then we |
| // should complete it now and return a meaningful result. Note that it would |
| // be nice if we didn't have to worry about starting codec configuration at |
| // all (::Initialize or the wrapper can do it), but then they have to remember |
| // not to start codec config if we have to wait for the cdm. It's somewhat |
| // clearer for us to handle both cases. |
| return ConfigureMediaCodecSynchronously(); |
| } |
| |
| void AndroidVideoDecodeAccelerator::DoIOTask(bool start_timer) { |
| DCHECK(thread_checker_.CalledOnValidThread()); |
| TRACE_EVENT0("media", "AVDA::DoIOTask"); |
| if (state_ == ERROR || state_ == WAITING_FOR_CODEC || |
| state_ == SURFACE_DESTROYED) { |
| return; |
| } |
| |
| strategy_->MaybeRenderEarly(); |
| bool did_work = false, did_input = false, did_output = false; |
| do { |
| did_input = QueueInput(); |
| did_output = DequeueOutput(); |
| if (did_input || did_output) |
| did_work = true; |
| } while (did_input || did_output); |
| |
| ManageTimer(did_work || start_timer); |
| } |
| |
| bool AndroidVideoDecodeAccelerator::QueueInput() { |
| DCHECK(thread_checker_.CalledOnValidThread()); |
| TRACE_EVENT0("media", "AVDA::QueueInput"); |
| base::AutoReset<bool> auto_reset(&defer_errors_, true); |
| if (state_ == ERROR || state_ == WAITING_FOR_CODEC || |
| state_ == WAITING_FOR_KEY) { |
| return false; |
| } |
| if (bitstreams_notified_in_advance_.size() > kMaxBitstreamsNotifiedInAdvance) |
| return false; |
| if (pending_bitstream_records_.empty()) |
| return false; |
| |
| int input_buf_index = pending_input_buf_index_; |
| |
| // Do not dequeue a new input buffer if we failed with MEDIA_CODEC_NO_KEY. |
| // That status does not return this buffer back to the pool of |
| // available input buffers. We have to reuse it in QueueSecureInputBuffer(). |
| if (input_buf_index == -1) { |
| MediaCodecStatus status = |
| media_codec_->DequeueInputBuffer(NoWaitTimeOut, &input_buf_index); |
| switch (status) { |
| case MEDIA_CODEC_DEQUEUE_INPUT_AGAIN_LATER: |
| return false; |
| case MEDIA_CODEC_ERROR: |
| POST_ERROR(PLATFORM_FAILURE, "Failed to DequeueInputBuffer"); |
| return false; |
| case MEDIA_CODEC_OK: |
| break; |
| default: |
| NOTREACHED() << "Unknown DequeueInputBuffer status " << status; |
| return false; |
| } |
| } |
| |
| DCHECK_NE(input_buf_index, -1); |
| |
| BitstreamBuffer bitstream_buffer = pending_bitstream_records_.front().buffer; |
| |
| if (bitstream_buffer.id() == -1) { |
| pending_bitstream_records_.pop(); |
| TRACE_COUNTER1("media", "AVDA::PendingBitstreamBufferCount", |
| pending_bitstream_records_.size()); |
| |
| media_codec_->QueueEOS(input_buf_index); |
| return true; |
| } |
| |
| std::unique_ptr<SharedMemoryRegion> shm; |
| |
| if (pending_input_buf_index_ == -1) { |
| // When |pending_input_buf_index_| is not -1, the buffer is already dequeued |
| // from MediaCodec, filled with data and bitstream_buffer.handle() is |
| // closed. |
| shm = std::move(pending_bitstream_records_.front().memory); |
| |
| if (!shm->Map()) { |
| POST_ERROR(UNREADABLE_INPUT, "Failed to SharedMemoryRegion::Map()"); |
| return false; |
| } |
| } |
| |
| const base::TimeDelta presentation_timestamp = |
| bitstream_buffer.presentation_timestamp(); |
| DCHECK(presentation_timestamp != kNoTimestamp) |
| << "Bitstream buffers must have valid presentation timestamps"; |
| |
| // There may already be a bitstream buffer with this timestamp, e.g., VP9 alt |
| // ref frames, but it's OK to overwrite it because we only expect a single |
| // output frame to have that timestamp. AVDA clients only use the bitstream |
| // buffer id in the returned Pictures to map a bitstream buffer back to a |
| // timestamp on their side, so either one of the bitstream buffer ids will |
| // result in them finding the right timestamp. |
| bitstream_buffers_in_decoder_[presentation_timestamp] = bitstream_buffer.id(); |
| |
| // Notice that |memory| will be null if we repeatedly enqueue the same buffer, |
| // this happens after MEDIA_CODEC_NO_KEY. |
| const uint8_t* memory = |
| shm ? static_cast<const uint8_t*>(shm->memory()) : nullptr; |
| const std::string& key_id = bitstream_buffer.key_id(); |
| const std::string& iv = bitstream_buffer.iv(); |
| const std::vector<SubsampleEntry>& subsamples = bitstream_buffer.subsamples(); |
| |
| MediaCodecStatus status; |
| if (key_id.empty() || iv.empty()) { |
| status = media_codec_->QueueInputBuffer(input_buf_index, memory, |
| bitstream_buffer.size(), |
| presentation_timestamp); |
| } else { |
| status = media_codec_->QueueSecureInputBuffer( |
| input_buf_index, memory, bitstream_buffer.size(), key_id, iv, |
| subsamples, presentation_timestamp); |
| } |
| |
| DVLOG(2) << __FUNCTION__ |
| << ": Queue(Secure)InputBuffer: pts:" << presentation_timestamp |
| << " status:" << status; |
| |
| if (status == MEDIA_CODEC_NO_KEY) { |
| // Keep trying to enqueue the same input buffer. |
| // The buffer is owned by us (not the MediaCodec) and is filled with data. |
| DVLOG(1) << "QueueSecureInputBuffer failed: NO_KEY"; |
| pending_input_buf_index_ = input_buf_index; |
| state_ = WAITING_FOR_KEY; |
| return false; |
| } |
| |
| pending_input_buf_index_ = -1; |
| pending_bitstream_records_.pop(); |
| TRACE_COUNTER1("media", "AVDA::PendingBitstreamBufferCount", |
| pending_bitstream_records_.size()); |
| // We should call NotifyEndOfBitstreamBuffer(), when no more decoded output |
| // will be returned from the bitstream buffer. However, MediaCodec API is |
| // not enough to guarantee it. |
| // So, here, we calls NotifyEndOfBitstreamBuffer() in advance in order to |
| // keep getting more bitstreams from the client, and throttle them by using |
| // |bitstreams_notified_in_advance_|. |
| // TODO(dwkang): check if there is a way to remove this workaround. |
| base::ThreadTaskRunnerHandle::Get()->PostTask( |
| FROM_HERE, |
| base::Bind(&AndroidVideoDecodeAccelerator::NotifyEndOfBitstreamBuffer, |
| weak_this_factory_.GetWeakPtr(), bitstream_buffer.id())); |
| bitstreams_notified_in_advance_.push_back(bitstream_buffer.id()); |
| |
| if (status != MEDIA_CODEC_OK) { |
| POST_ERROR(PLATFORM_FAILURE, "Failed to QueueInputBuffer: " << status); |
| return false; |
| } |
| |
| return true; |
| } |
| |
| bool AndroidVideoDecodeAccelerator::DequeueOutput() { |
| DCHECK(thread_checker_.CalledOnValidThread()); |
| TRACE_EVENT0("media", "AVDA::DequeueOutput"); |
| base::AutoReset<bool> auto_reset(&defer_errors_, true); |
| if (state_ == ERROR || state_ == WAITING_FOR_CODEC) |
| return false; |
| if (picturebuffers_requested_ && output_picture_buffers_.empty()) |
| return false; |
| if (!output_picture_buffers_.empty() && free_picture_ids_.empty()) { |
| // Don't have any picture buffer to send. Need to wait. |
| return false; |
| } |
| |
| bool eos = false; |
| base::TimeDelta presentation_timestamp; |
| int32_t buf_index = 0; |
| do { |
| size_t offset = 0; |
| size_t size = 0; |
| |
| TRACE_EVENT_BEGIN0("media", "AVDA::DequeueOutput"); |
| MediaCodecStatus status = media_codec_->DequeueOutputBuffer( |
| NoWaitTimeOut, &buf_index, &offset, &size, &presentation_timestamp, |
| &eos, NULL); |
| TRACE_EVENT_END2("media", "AVDA::DequeueOutput", "status", status, |
| "presentation_timestamp (ms)", |
| presentation_timestamp.InMilliseconds()); |
| |
| switch (status) { |
| case MEDIA_CODEC_ERROR: |
| // Do not post an error if we are draining for reset and destroy. |
| // Instead, run the drain completion task. |
| if (IsDrainingForResetOrDestroy()) { |
| DVLOG(1) << __FUNCTION__ << ": error while codec draining"; |
| state_ = ERROR; |
| OnDrainCompleted(); |
| } else { |
| POST_ERROR(PLATFORM_FAILURE, "DequeueOutputBuffer failed."); |
| } |
| return false; |
| |
| case MEDIA_CODEC_DEQUEUE_OUTPUT_AGAIN_LATER: |
| return false; |
| |
| case MEDIA_CODEC_OUTPUT_FORMAT_CHANGED: { |
| // An OUTPUT_FORMAT_CHANGED is not reported after flush() if the frame |
| // size does not change. Therefore we have to keep track on the format |
| // even if draining, unless we are draining for destroy. |
| if (drain_type_ == DRAIN_FOR_DESTROY) |
| return true; // ignore |
| |
| if (media_codec_->GetOutputSize(&size_) != MEDIA_CODEC_OK) { |
| POST_ERROR(PLATFORM_FAILURE, "GetOutputSize failed."); |
| return false; |
| } |
| |
| DVLOG(3) << __FUNCTION__ |
| << " OUTPUT_FORMAT_CHANGED, new size: " << size_.ToString(); |
| |
| // Don't request picture buffers if we already have some. This avoids |
| // having to dismiss the existing buffers which may actively reference |
| // decoded images. Breaking their connection to the decoded image will |
| // cause rendering of black frames. Instead, we let the existing |
| // PictureBuffers live on and we simply update their size the next time |
| // they're attachted to an image of the new resolution. See the |
| // size update in |SendDecodedFrameToClient| and https://crbug/587994. |
| if (output_picture_buffers_.empty() && !picturebuffers_requested_) { |
| picturebuffers_requested_ = true; |
| base::ThreadTaskRunnerHandle::Get()->PostTask( |
| FROM_HERE, |
| base::Bind(&AndroidVideoDecodeAccelerator::RequestPictureBuffers, |
| weak_this_factory_.GetWeakPtr())); |
| return false; |
| } |
| |
| return true; |
| } |
| |
| case MEDIA_CODEC_OUTPUT_BUFFERS_CHANGED: |
| break; |
| |
| case MEDIA_CODEC_OK: |
| DCHECK_GE(buf_index, 0); |
| DVLOG(3) << __FUNCTION__ << ": pts:" << presentation_timestamp |
| << " buf_index:" << buf_index << " offset:" << offset |
| << " size:" << size << " eos:" << eos; |
| break; |
| |
| default: |
| NOTREACHED(); |
| break; |
| } |
| } while (buf_index < 0); |
| |
| if (eos) { |
| OnDrainCompleted(); |
| return false; |
| } |
| |
| if (IsDrainingForResetOrDestroy()) { |
| media_codec_->ReleaseOutputBuffer(buf_index, false); |
| return true; |
| } |
| |
| if (!picturebuffers_requested_) { |
| // If, somehow, we get a decoded frame back before a FORMAT_CHANGED |
| // message, then we might not have any picture buffers to use. This |
| // isn't supposed to happen (see EncodeDecodeTest.java#617). |
| // Log a metric to see how common this is. |
| RecordFormatChangedMetric(FormatChangedValue::MissingFormatChanged); |
| media_codec_->ReleaseOutputBuffer(buf_index, false); |
| POST_ERROR(PLATFORM_FAILURE, "Dequeued buffers before FORMAT_CHANGED."); |
| return false; |
| } |
| |
| // Get the bitstream buffer id from the timestamp. |
| auto it = bitstream_buffers_in_decoder_.find(presentation_timestamp); |
| |
| if (it != bitstream_buffers_in_decoder_.end()) { |
| const int32_t bitstream_buffer_id = it->second; |
| bitstream_buffers_in_decoder_.erase(bitstream_buffers_in_decoder_.begin(), |
| ++it); |
| SendDecodedFrameToClient(buf_index, bitstream_buffer_id); |
| |
| // Removes ids former or equal than the id from decoder. Note that |
| // |bitstreams_notified_in_advance_| does not mean bitstream ids in decoder |
| // because of frame reordering issue. We just maintain this roughly and use |
| // it for throttling. |
| for (auto bitstream_it = bitstreams_notified_in_advance_.begin(); |
| bitstream_it != bitstreams_notified_in_advance_.end(); |
| ++bitstream_it) { |
| if (*bitstream_it == bitstream_buffer_id) { |
| bitstreams_notified_in_advance_.erase( |
| bitstreams_notified_in_advance_.begin(), ++bitstream_it); |
| break; |
| } |
| } |
| } else { |
| // Normally we assume that the decoder makes at most one output frame for |
| // each distinct input timestamp. However MediaCodecBridge uses timestamp |
| // correction and provides a non-decreasing timestamp sequence, which might |
| // result in timestamp duplicates. Discard the frame if we cannot get the |
| // corresponding buffer id. |
| DVLOG(3) << __FUNCTION__ << ": Releasing buffer with unexpected PTS: " |
| << presentation_timestamp; |
| media_codec_->ReleaseOutputBuffer(buf_index, false); |
| } |
| |
| // We got a decoded frame, so try for another. |
| return true; |
| } |
| |
| void AndroidVideoDecodeAccelerator::SendDecodedFrameToClient( |
| int32_t codec_buffer_index, |
| int32_t bitstream_id) { |
| DCHECK(thread_checker_.CalledOnValidThread()); |
| DCHECK_NE(bitstream_id, -1); |
| DCHECK(!free_picture_ids_.empty()); |
| TRACE_EVENT0("media", "AVDA::SendDecodedFrameToClient"); |
| |
| if (!make_context_current_cb_.Run()) { |
| POST_ERROR(PLATFORM_FAILURE, "Failed to make the GL context current."); |
| return; |
| } |
| |
| int32_t picture_buffer_id = free_picture_ids_.front(); |
| free_picture_ids_.pop(); |
| TRACE_COUNTER1("media", "AVDA::FreePictureIds", free_picture_ids_.size()); |
| |
| const auto& i = output_picture_buffers_.find(picture_buffer_id); |
| if (i == output_picture_buffers_.end()) { |
| POST_ERROR(PLATFORM_FAILURE, |
| "Can't find PictureBuffer id: " << picture_buffer_id); |
| return; |
| } |
| |
| bool size_changed = false; |
| if (i->second.size() != size_) { |
| // Size may have changed due to resolution change since the last time this |
| // PictureBuffer was used. |
| strategy_->UpdatePictureBufferSize(&i->second, size_); |
| size_changed = true; |
| } |
| |
| const bool allow_overlay = strategy_->ArePicturesOverlayable(); |
| UMA_HISTOGRAM_BOOLEAN("Media.AVDA.FrameSentAsOverlay", allow_overlay); |
| Picture picture(picture_buffer_id, bitstream_id, gfx::Rect(size_), |
| allow_overlay); |
| picture.set_size_changed(size_changed); |
| |
| // Notify picture ready before calling UseCodecBufferForPictureBuffer() since |
| // that process may be slow and shouldn't delay delivery of the frame to the |
| // renderer. The picture is only used on the same thread as this method is |
| // called, so it is safe to do this. |
| NotifyPictureReady(picture); |
| |
| // Connect the PictureBuffer to the decoded frame, via whatever mechanism the |
| // strategy likes. |
| strategy_->UseCodecBufferForPictureBuffer(codec_buffer_index, i->second); |
| } |
| |
| void AndroidVideoDecodeAccelerator::Decode( |
| const BitstreamBuffer& bitstream_buffer) { |
| DCHECK(thread_checker_.CalledOnValidThread()); |
| |
| // If we previously deferred a codec restart, take care of it now. This can |
| // happen on older devices where configuration changes require a codec reset. |
| if (codec_needs_reset_) { |
| DCHECK_EQ(drain_type_, DRAIN_TYPE_NONE); |
| ResetCodecState(); |
| } |
| |
| if (bitstream_buffer.id() >= 0 && bitstream_buffer.size() > 0) { |
| DecodeBuffer(bitstream_buffer); |
| return; |
| } |
| |
| if (base::SharedMemory::IsHandleValid(bitstream_buffer.handle())) |
| base::SharedMemory::CloseHandle(bitstream_buffer.handle()); |
| |
| if (bitstream_buffer.id() < 0) { |
| POST_ERROR(INVALID_ARGUMENT, |
| "Invalid bistream_buffer, id: " << bitstream_buffer.id()); |
| } else { |
| base::ThreadTaskRunnerHandle::Get()->PostTask( |
| FROM_HERE, |
| base::Bind(&AndroidVideoDecodeAccelerator::NotifyEndOfBitstreamBuffer, |
| weak_this_factory_.GetWeakPtr(), bitstream_buffer.id())); |
| } |
| } |
| |
| void AndroidVideoDecodeAccelerator::DecodeBuffer( |
| const BitstreamBuffer& bitstream_buffer) { |
| pending_bitstream_records_.push(BitstreamRecord(bitstream_buffer)); |
| TRACE_COUNTER1("media", "AVDA::PendingBitstreamBufferCount", |
| pending_bitstream_records_.size()); |
| |
| DoIOTask(true); |
| } |
| |
| void AndroidVideoDecodeAccelerator::RequestPictureBuffers() { |
| if (client_) { |
| client_->ProvidePictureBuffers(kNumPictureBuffers, PIXEL_FORMAT_UNKNOWN, 1, |
| strategy_->GetPictureBufferSize(), |
| strategy_->GetTextureTarget()); |
| } |
| } |
| |
| void AndroidVideoDecodeAccelerator::AssignPictureBuffers( |
| const std::vector<PictureBuffer>& buffers) { |
| DCHECK(thread_checker_.CalledOnValidThread()); |
| DCHECK(output_picture_buffers_.empty()); |
| DCHECK(free_picture_ids_.empty()); |
| |
| if (buffers.size() < kNumPictureBuffers) { |
| POST_ERROR(INVALID_ARGUMENT, "Not enough picture buffers assigned."); |
| return; |
| } |
| |
| const bool have_context = make_context_current_cb_.Run(); |
| LOG_IF(WARNING, !have_context) |
| << "Failed to make GL context current for Assign, continuing."; |
| |
| for (size_t i = 0; i < buffers.size(); ++i) { |
| if (buffers[i].size() != strategy_->GetPictureBufferSize()) { |
| POST_ERROR(INVALID_ARGUMENT, |
| "Invalid picture buffer size assigned. Wanted " |
| << size_.ToString() << ", but got " |
| << buffers[i].size().ToString()); |
| return; |
| } |
| int32_t id = buffers[i].id(); |
| output_picture_buffers_.insert(std::make_pair(id, buffers[i])); |
| free_picture_ids_.push(id); |
| |
| strategy_->AssignOnePictureBuffer(buffers[i], have_context); |
| } |
| TRACE_COUNTER1("media", "AVDA::FreePictureIds", free_picture_ids_.size()); |
| DoIOTask(true); |
| } |
| |
| void AndroidVideoDecodeAccelerator::ReusePictureBuffer( |
| int32_t picture_buffer_id) { |
| DCHECK(thread_checker_.CalledOnValidThread()); |
| |
| free_picture_ids_.push(picture_buffer_id); |
| TRACE_COUNTER1("media", "AVDA::FreePictureIds", free_picture_ids_.size()); |
| |
| OutputBufferMap::const_iterator i = |
| output_picture_buffers_.find(picture_buffer_id); |
| if (i == output_picture_buffers_.end()) { |
| POST_ERROR(PLATFORM_FAILURE, "Can't find PictureBuffer id " |
| << picture_buffer_id); |
| return; |
| } |
| |
| strategy_->ReuseOnePictureBuffer(i->second); |
| DoIOTask(true); |
| } |
| |
| void AndroidVideoDecodeAccelerator::Flush() { |
| DVLOG(1) << __FUNCTION__; |
| DCHECK(thread_checker_.CalledOnValidThread()); |
| |
| if (state_ == SURFACE_DESTROYED) |
| NotifyFlushDone(); |
| else |
| StartCodecDrain(DRAIN_FOR_FLUSH); |
| } |
| |
| void AndroidVideoDecodeAccelerator::ConfigureMediaCodecAsynchronously() { |
| DCHECK(thread_checker_.CalledOnValidThread()); |
| |
| // It's probably okay just to return here, since the codec will be configured |
| // asynchronously. It's unclear that any state for the new request could |
| // be different, unless somebody modifies |codec_config_| while we're already |
| // waiting for a codec. One shouldn't do that for thread safety. |
| DCHECK_NE(state_, WAITING_FOR_CODEC); |
| |
| state_ = WAITING_FOR_CODEC; |
| |
| // Tell the strategy that we're changing codecs. The codec itself could be |
| // used normally, since we don't replace it until we're back on the main |
| // thread. However, if we're using an output surface, then the incoming codec |
| // might access that surface while the main thread is drawing. Telling the |
| // strategy to forget the codec avoids this. |
| if (media_codec_) { |
| ReleaseMediaCodec(); |
| strategy_->CodecChanged(nullptr); |
| } |
| |
| // Choose whether to autodetect the codec type. Note that we do this after |
| // releasing any outgoing codec, so that |codec_config_| still matches the |
| // outgoing codec for ReleaseMediaCodec(). |
| codec_config_->allow_autodetection_ = |
| g_avda_timer.Pointer()->IsCodecAutodetectionProbablySafe(); |
| |
| // If autodetection is disallowed, fall back to Chrome's software decoders |
| // instead of using the software decoders provided by MediaCodec. |
| if (!codec_config_->allow_autodetection_ && |
| IsMediaCodecSoftwareDecodingForbidden()) { |
| OnCodecConfigured(nullptr); |
| return; |
| } |
| |
| codec_config_->notify_completion_ = codec_config_->allow_autodetection_; |
| if (codec_config_->allow_autodetection_) |
| g_avda_timer.Pointer()->StartUsingConstructionThread(); |
| |
| // If we're not trying autodetection, then use the main thread. The original |
| // might be blocked. |
| scoped_refptr<base::SingleThreadTaskRunner> task_runner = |
| codec_config_->allow_autodetection_ |
| ? g_avda_timer.Pointer()->ConstructionTaskRunner() |
| : base::ThreadTaskRunnerHandle::Get(); |
| CHECK(task_runner); |
| |
| base::PostTaskAndReplyWithResult( |
| task_runner.get(), FROM_HERE, |
| base::Bind(&AndroidVideoDecodeAccelerator::ConfigureMediaCodecOnAnyThread, |
| codec_config_), |
| base::Bind(&AndroidVideoDecodeAccelerator::OnCodecConfigured, |
| weak_this_factory_.GetWeakPtr())); |
| } |
| |
| bool AndroidVideoDecodeAccelerator::ConfigureMediaCodecSynchronously() { |
| state_ = WAITING_FOR_CODEC; |
| |
| // Decide whether to allow autodetection or not. Since we're on the main |
| // thread, and this request is unordered with respect to pending async config |
| // attempts, don't record it. It may break book-keeping, and there's not |
| // much we can do anyway. |
| codec_config_->allow_autodetection_ = |
| g_avda_timer.Pointer()->IsCodecAutodetectionProbablySafe(); |
| codec_config_->notify_completion_ = false; |
| |
| ReleaseMediaCodec(); |
| std::unique_ptr<VideoCodecBridge> media_codec = |
| ConfigureMediaCodecOnAnyThread(codec_config_); |
| OnCodecConfigured(std::move(media_codec)); |
| return !!media_codec_; |
| } |
| |
| std::unique_ptr<VideoCodecBridge> |
| AndroidVideoDecodeAccelerator::ConfigureMediaCodecOnAnyThread( |
| scoped_refptr<CodecConfig> codec_config) { |
| TRACE_EVENT0("media", "AVDA::ConfigureMediaCodec"); |
| |
| jobject media_crypto = codec_config->media_crypto_ |
| ? codec_config->media_crypto_->obj() |
| : nullptr; |
| |
| // |needs_protected_surface_| implies encrypted stream. |
| DCHECK(!codec_config->needs_protected_surface_ || media_crypto); |
| |
| const bool require_software_codec = !codec_config->allow_autodetection_; |
| |
| std::unique_ptr<VideoCodecBridge> codec(VideoCodecBridge::CreateDecoder( |
| codec_config->codec_, codec_config->needs_protected_surface_, |
| codec_config->initial_expected_coded_size_, |
| codec_config->surface_.j_surface().obj(), media_crypto, true, |
| require_software_codec)); |
| |
| // If we successfully completed after an autodetect, then let the other |
| // instances know that we didn't get stuck. |
| if (codec_config->notify_completion_) |
| g_avda_timer.Pointer()->DoneUsingConstructionThread(); |
| |
| return codec; |
| } |
| |
| void AndroidVideoDecodeAccelerator::OnCodecConfigured( |
| std::unique_ptr<VideoCodecBridge> media_codec) { |
| DCHECK(thread_checker_.CalledOnValidThread()); |
| DCHECK(state_ == WAITING_FOR_CODEC || state_ == SURFACE_DESTROYED); |
| |
| // Record one instance of the codec being initialized. |
| RecordFormatChangedMetric(FormatChangedValue::CodecInitialized); |
| |
| // If we are supposed to notify that initialization is complete, then do so |
| // now. Otherwise, this is a reconfiguration. |
| if (deferred_initialization_pending_) { |
| // Losing the output surface is not considered an error state, so notify |
| // success. The client will destroy this soon. |
| NotifyInitializationComplete(state_ == SURFACE_DESTROYED ? true |
| : !!media_codec); |
| deferred_initialization_pending_ = false; |
| } |
| |
| // If |state_| changed to SURFACE_DESTROYED while we were configuring a codec, |
| // then the codec is already invalid so we return early and drop it. |
| if (state_ == SURFACE_DESTROYED) |
| return; |
| |
| DCHECK(!media_codec_); |
| media_codec_ = std::move(media_codec); |
| strategy_->CodecChanged(media_codec_.get()); |
| if (!media_codec_) { |
| POST_ERROR(PLATFORM_FAILURE, "Failed to create MediaCodec."); |
| return; |
| } |
| |
| state_ = NO_ERROR; |
| |
| ManageTimer(true); |
| } |
| |
| void AndroidVideoDecodeAccelerator::StartCodecDrain(DrainType drain_type) { |
| DVLOG(2) << __FUNCTION__ << " drain_type:" << drain_type; |
| DCHECK(thread_checker_.CalledOnValidThread()); |
| |
| // We assume that DRAIN_FOR_FLUSH and DRAIN_FOR_RESET cannot come while |
| // another drain request is present, but DRAIN_FOR_DESTROY can. |
| DCHECK_NE(drain_type, DRAIN_TYPE_NONE); |
| DCHECK(drain_type_ == DRAIN_TYPE_NONE || drain_type == DRAIN_FOR_DESTROY) |
| << "Unexpected StartCodecDrain() with drain type " << drain_type |
| << " while already draining with drain type " << drain_type_; |
| |
| const bool enqueue_eos = drain_type_ == DRAIN_TYPE_NONE; |
| drain_type_ = drain_type; |
| |
| if (enqueue_eos) |
| DecodeBuffer(BitstreamBuffer(-1, base::SharedMemoryHandle(), 0)); |
| } |
| |
| bool AndroidVideoDecodeAccelerator::IsDrainingForResetOrDestroy() const { |
| return drain_type_ == DRAIN_FOR_RESET || drain_type_ == DRAIN_FOR_DESTROY; |
| } |
| |
| void AndroidVideoDecodeAccelerator::OnDrainCompleted() { |
| DVLOG(2) << __FUNCTION__; |
| DCHECK(thread_checker_.CalledOnValidThread()); |
| |
| // If we were waiting for an EOS, clear the state and reset the MediaCodec |
| // as normal. |
| // |
| // Some Android platforms seem to send an EOS buffer even when we're not |
| // expecting it. In this case, destroy and reset the codec but don't notify |
| // flush done since it violates the state machine. http://crbug.com/585959. |
| |
| switch (drain_type_) { |
| case DRAIN_TYPE_NONE: |
| // Unexpected EOS. |
| state_ = ERROR; |
| ResetCodecState(); |
| break; |
| case DRAIN_FOR_FLUSH: |
| ResetCodecState(); |
| base::ThreadTaskRunnerHandle::Get()->PostTask( |
| FROM_HERE, base::Bind(&AndroidVideoDecodeAccelerator::NotifyFlushDone, |
| weak_this_factory_.GetWeakPtr())); |
| break; |
| case DRAIN_FOR_RESET: |
| ResetCodecState(); |
| base::ThreadTaskRunnerHandle::Get()->PostTask( |
| FROM_HERE, base::Bind(&AndroidVideoDecodeAccelerator::NotifyResetDone, |
| weak_this_factory_.GetWeakPtr())); |
| break; |
| case DRAIN_FOR_DESTROY: |
| ResetCodecState(); |
| base::ThreadTaskRunnerHandle::Get()->PostTask( |
| FROM_HERE, base::Bind(&AndroidVideoDecodeAccelerator::ActualDestroy, |
| weak_this_factory_.GetWeakPtr())); |
| break; |
| } |
| drain_type_ = DRAIN_TYPE_NONE; |
| } |
| |
| void AndroidVideoDecodeAccelerator::ResetCodecState() { |
| DCHECK(thread_checker_.CalledOnValidThread()); |
| |
| // If there is already a reset in flight, then that counts. This can really |
| // only happen if somebody calls Reset. |
| // If the surface is destroyed there's nothing to do. |
| if (state_ == WAITING_FOR_CODEC || state_ == SURFACE_DESTROYED) |
| return; |
| |
| bitstream_buffers_in_decoder_.clear(); |
| |
| if (pending_input_buf_index_ != -1) { |
| // The data for that index exists in the input buffer, but corresponding |
| // shm block been deleted. Check that it is safe to flush the codec, i.e. |
| // |pending_bitstream_records_| is empty. |
| // TODO(timav): keep shm block for that buffer and remove this restriction. |
| DCHECK(pending_bitstream_records_.empty()); |
| pending_input_buf_index_ = -1; |
| } |
| |
| const bool did_codec_error_happen = state_ == ERROR; |
| state_ = NO_ERROR; |
| |
| // Don't reset the codec here if there's no error and we're only flushing; |
| // instead defer until the next decode call; this prevents us from unbacking |
| // frames that might be out for display at end of stream. |
| codec_needs_reset_ = false; |
| if (drain_type_ == DRAIN_FOR_FLUSH && !did_codec_error_happen) { |
| codec_needs_reset_ = true; |
| return; |
| } |
| |
| // We might increment error_sequence_token here to cancel any delayed errors, |
| // but right now it's unclear that it's safe to do so. If we are in an error |
| // state because of a codec error, then it would be okay. Otherwise, it's |
| // less obvious that we are exiting the error state. Since deferred errors |
| // are only intended for fullscreen transitions right now, we take the more |
| // conservative approach and let the errors post. |
| // TODO(liberato): revisit this once we sort out the error state a bit more. |
| |
| // Flush the codec if possible, or create a new one if not. |
| if (!did_codec_error_happen && |
| !media::MediaCodecUtil::CodecNeedsFlushWorkaround(media_codec_.get())) { |
| DVLOG(3) << __FUNCTION__ << " Flushing MediaCodec."; |
| media_codec_->Flush(); |
| // Since we just flushed all the output buffers, make sure that nothing is |
| // using them. |
| strategy_->CodecChanged(media_codec_.get()); |
| } else { |
| DVLOG(3) << __FUNCTION__ |
| << " Deleting the MediaCodec and creating a new one."; |
| g_avda_timer.Pointer()->StopTimer(this); |
| ConfigureMediaCodecAsynchronously(); |
| } |
| } |
| |
| void AndroidVideoDecodeAccelerator::Reset() { |
| DVLOG(1) << __FUNCTION__; |
| DCHECK(thread_checker_.CalledOnValidThread()); |
| TRACE_EVENT0("media", "AVDA::Reset"); |
| |
| while (!pending_bitstream_records_.empty()) { |
| int32_t bitstream_buffer_id = |
| pending_bitstream_records_.front().buffer.id(); |
| pending_bitstream_records_.pop(); |
| |
| if (bitstream_buffer_id != -1) { |
| base::ThreadTaskRunnerHandle::Get()->PostTask( |
| FROM_HERE, |
| base::Bind(&AndroidVideoDecodeAccelerator::NotifyEndOfBitstreamBuffer, |
| weak_this_factory_.GetWeakPtr(), bitstream_buffer_id)); |
| } |
| } |
| TRACE_COUNTER1("media", "AVDA::PendingBitstreamBufferCount", 0); |
| bitstreams_notified_in_advance_.clear(); |
| |
| // Any error that is waiting to post can be ignored. |
| error_sequence_token_++; |
| |
| DCHECK(strategy_); |
| strategy_->ReleaseCodecBuffers(output_picture_buffers_); |
| |
| // Some VP8 files require complete MediaCodec drain before we can call |
| // MediaCodec.flush() or MediaCodec.reset(). http://crbug.com/598963. |
| if (media_codec_ && codec_config_->codec_ == kCodecVP8 && |
| !bitstream_buffers_in_decoder_.empty()) { |
| // Postpone ResetCodecState() after the drain. |
| StartCodecDrain(DRAIN_FOR_RESET); |
| } else { |
| ResetCodecState(); |
| base::ThreadTaskRunnerHandle::Get()->PostTask( |
| FROM_HERE, base::Bind(&AndroidVideoDecodeAccelerator::NotifyResetDone, |
| weak_this_factory_.GetWeakPtr())); |
| } |
| } |
| |
| void AndroidVideoDecodeAccelerator::Destroy() { |
| DVLOG(1) << __FUNCTION__; |
| DCHECK(thread_checker_.CalledOnValidThread()); |
| |
| bool have_context = make_context_current_cb_.Run(); |
| if (!have_context) |
| LOG(WARNING) << "Failed make GL context current for Destroy, continuing."; |
| |
| if (strategy_) |
| strategy_->BeginCleanup(have_context, output_picture_buffers_); |
| |
| // If we have an OnFrameAvailable handler, tell it that we're going away. |
| if (on_frame_available_handler_) { |
| on_frame_available_handler_->ClearOwner(); |
| on_frame_available_handler_ = nullptr; |
| } |
| |
| client_ = nullptr; |
| |
| // Some VP8 files require complete MediaCodec drain before we can call |
| // MediaCodec.flush() or MediaCodec.reset(). http://crbug.com/598963. |
| if (media_codec_ && codec_config_->codec_ == kCodecVP8) { |
| // Clear pending_bitstream_records_. |
| while (!pending_bitstream_records_.empty()) |
| pending_bitstream_records_.pop(); |
| |
| // Postpone ActualDestroy after the drain. |
| StartCodecDrain(DRAIN_FOR_DESTROY); |
| } else { |
| ActualDestroy(); |
| } |
| } |
| |
| void AndroidVideoDecodeAccelerator::ActualDestroy() { |
| DVLOG(1) << __FUNCTION__; |
| DCHECK(thread_checker_.CalledOnValidThread()); |
| |
| if (!on_destroying_surface_cb_.is_null()) { |
| AVDASurfaceTracker::GetInstance()->UnregisterOnDestroyingSurfaceCallback( |
| on_destroying_surface_cb_); |
| } |
| |
| if (strategy_) |
| strategy_->EndCleanup(); |
| |
| AVDATimerManager* manager = g_avda_timer.Pointer(); |
| |
| // We no longer care about |surface_id|, in case we did before. It's okay |
| // if we have no surface and/or weren't the owner or a waiter. |
| manager->DeallocateSurface(config_.surface_id, this); |
| |
| // Note that async codec construction might still be in progress. In that |
| // case, the codec will be deleted when it completes once we invalidate all |
| // our weak refs. |
| weak_this_factory_.InvalidateWeakPtrs(); |
| if (media_codec_) { |
| manager->StopTimer(this); |
| ReleaseMediaCodec(); |
| } |
| delete this; |
| } |
| |
| bool AndroidVideoDecodeAccelerator::TryToSetupDecodeOnSeparateThread( |
| const base::WeakPtr<Client>& decode_client, |
| const scoped_refptr<base::SingleThreadTaskRunner>& decode_task_runner) { |
| return false; |
| } |
| |
| const gfx::Size& AndroidVideoDecodeAccelerator::GetSize() const { |
| return size_; |
| } |
| |
| const base::ThreadChecker& AndroidVideoDecodeAccelerator::ThreadChecker() |
| const { |
| return thread_checker_; |
| } |
| |
| base::WeakPtr<gpu::gles2::GLES2Decoder> |
| AndroidVideoDecodeAccelerator::GetGlDecoder() const { |
| return get_gles2_decoder_cb_.Run(); |
| } |
| |
| gpu::gles2::TextureRef* AndroidVideoDecodeAccelerator::GetTextureForPicture( |
| const PictureBuffer& picture_buffer) { |
| auto gles_decoder = GetGlDecoder(); |
| RETURN_ON_FAILURE(this, gles_decoder, "Failed to get GL decoder", |
| ILLEGAL_STATE, nullptr); |
| RETURN_ON_FAILURE(this, gles_decoder->GetContextGroup(), |
| "Null gles_decoder->GetContextGroup()", ILLEGAL_STATE, |
| nullptr); |
| gpu::gles2::TextureManager* texture_manager = |
| gles_decoder->GetContextGroup()->texture_manager(); |
| RETURN_ON_FAILURE(this, texture_manager, "Null texture_manager", |
| ILLEGAL_STATE, nullptr); |
| |
| DCHECK_LE(1u, picture_buffer.internal_texture_ids().size()); |
| gpu::gles2::TextureRef* texture_ref = |
| texture_manager->GetTexture(picture_buffer.internal_texture_ids()[0]); |
| RETURN_ON_FAILURE(this, texture_manager, "Null texture_ref", ILLEGAL_STATE, |
| nullptr); |
| |
| return texture_ref; |
| } |
| |
| scoped_refptr<gl::SurfaceTexture> |
| AndroidVideoDecodeAccelerator::CreateAttachedSurfaceTexture( |
| GLuint* service_id) { |
| GLuint texture_id; |
| glGenTextures(1, &texture_id); |
| |
| glActiveTexture(GL_TEXTURE0); |
| glBindTexture(GL_TEXTURE_EXTERNAL_OES, texture_id); |
| glTexParameteri(GL_TEXTURE_EXTERNAL_OES, GL_TEXTURE_MAG_FILTER, GL_LINEAR); |
| glTexParameteri(GL_TEXTURE_EXTERNAL_OES, GL_TEXTURE_MIN_FILTER, GL_LINEAR); |
| glTexParameteri(GL_TEXTURE_EXTERNAL_OES, GL_TEXTURE_WRAP_S, GL_CLAMP_TO_EDGE); |
| glTexParameteri(GL_TEXTURE_EXTERNAL_OES, GL_TEXTURE_WRAP_T, GL_CLAMP_TO_EDGE); |
| |
| auto gl_decoder = GetGlDecoder(); |
| gl_decoder->RestoreTextureUnitBindings(0); |
| gl_decoder->RestoreActiveTexture(); |
| DCHECK_EQ(static_cast<GLenum>(GL_NO_ERROR), glGetError()); |
| |
| *service_id = texture_id; |
| // Previously, to reduce context switching, we used to create an unattached |
| // SurfaceTexture and attach it lazily in the compositor's context. But that |
| // was flaky because SurfaceTexture#detachFromGLContext() is buggy on a lot of |
| // devices. Now we attach it to the current context, which means we might have |
| // to context switch later to call updateTexImage(). Fortunately, if virtual |
| // contexts are in use, we won't have to context switch. |
| return gl::SurfaceTexture::Create(texture_id); |
| } |
| |
| void AndroidVideoDecodeAccelerator::OnDestroyingSurface(int surface_id) { |
| DCHECK(thread_checker_.CalledOnValidThread()); |
| TRACE_EVENT0("media", "AVDA::OnDestroyingSurface"); |
| DVLOG(1) << __FUNCTION__ << " surface_id: " << surface_id; |
| |
| if (surface_id != config_.surface_id) |
| return; |
| |
| // If we're currently asynchronously configuring a codec, it will be destroyed |
| // when configuration completes and it notices that |state_| has changed to |
| // SURFACE_DESTROYED. |
| state_ = SURFACE_DESTROYED; |
| if (media_codec_) { |
| ReleaseMediaCodec(); |
| strategy_->CodecChanged(media_codec_.get()); |
| } |
| // If we're draining, signal completion now because the drain can no longer |
| // proceed. |
| if (drain_type_ != DRAIN_TYPE_NONE) |
| OnDrainCompleted(); |
| } |
| |
| void AndroidVideoDecodeAccelerator::OnFrameAvailable() { |
| // Remember: this may be on any thread. |
| DCHECK(strategy_); |
| strategy_->OnFrameAvailable(); |
| } |
| |
| void AndroidVideoDecodeAccelerator::PostError( |
| const ::tracked_objects::Location& from_here, |
| VideoDecodeAccelerator::Error error) { |
| base::ThreadTaskRunnerHandle::Get()->PostDelayedTask( |
| from_here, |
| base::Bind(&AndroidVideoDecodeAccelerator::NotifyError, |
| weak_this_factory_.GetWeakPtr(), error, error_sequence_token_), |
| (defer_errors_ ? ErrorPostingDelay : base::TimeDelta())); |
| state_ = ERROR; |
| } |
| |
| void AndroidVideoDecodeAccelerator::InitializeCdm() { |
| DVLOG(2) << __FUNCTION__ << ": " << config_.cdm_id; |
| |
| #if !defined(ENABLE_MOJO_MEDIA_IN_GPU_PROCESS) |
| NOTIMPLEMENTED(); |
| NotifyInitializationComplete(false); |
| #else |
| // Store the CDM to hold a reference to it. |
| cdm_for_reference_holding_only_ = |
| MojoCdmService::LegacyGetCdm(config_.cdm_id); |
| DCHECK(cdm_for_reference_holding_only_); |
| |
| // On Android platform the CdmContext must be a MediaDrmBridgeCdmContext. |
| media_drm_bridge_cdm_context_ = static_cast<MediaDrmBridgeCdmContext*>( |
| cdm_for_reference_holding_only_->GetCdmContext()); |
| DCHECK(media_drm_bridge_cdm_context_); |
| |
| // Register CDM callbacks. The callbacks registered will be posted back to |
| // this thread via BindToCurrentLoop. |
| |
| // Since |this| holds a reference to the |cdm_|, by the time the CDM is |
| // destructed, UnregisterPlayer() must have been called and |this| has been |
| // destructed as well. So the |cdm_unset_cb| will never have a chance to be |
| // called. |
| // TODO(xhwang): Remove |cdm_unset_cb| after it's not used on all platforms. |
| cdm_registration_id_ = media_drm_bridge_cdm_context_->RegisterPlayer( |
| BindToCurrentLoop(base::Bind(&AndroidVideoDecodeAccelerator::OnKeyAdded, |
| weak_this_factory_.GetWeakPtr())), |
| base::Bind(&base::DoNothing)); |
| |
| // Deferred initialization will continue in OnMediaCryptoReady(). |
| media_drm_bridge_cdm_context_->SetMediaCryptoReadyCB(BindToCurrentLoop( |
| base::Bind(&AndroidVideoDecodeAccelerator::OnMediaCryptoReady, |
| weak_this_factory_.GetWeakPtr()))); |
| #endif // !defined(ENABLE_MOJO_MEDIA_IN_GPU_PROCESS) |
| } |
| |
| void AndroidVideoDecodeAccelerator::OnMediaCryptoReady( |
| MediaDrmBridgeCdmContext::JavaObjectPtr media_crypto, |
| bool needs_protected_surface) { |
| DVLOG(1) << __FUNCTION__; |
| |
| if (!media_crypto) { |
| LOG(ERROR) << "MediaCrypto is not available, can't play encrypted stream."; |
| cdm_for_reference_holding_only_ = nullptr; |
| media_drm_bridge_cdm_context_ = nullptr; |
| NotifyInitializationComplete(false); |
| return; |
| } |
| |
| DCHECK(!media_crypto->is_null()); |
| |
| // We assume this is a part of the initialization process, thus MediaCodec |
| // is not created yet. |
| DCHECK(!media_codec_); |
| |
| codec_config_->media_crypto_ = std::move(media_crypto); |
| codec_config_->needs_protected_surface_ = needs_protected_surface; |
| |
| // After receiving |media_crypto_| we can configure MediaCodec. |
| ConfigureMediaCodecAsynchronously(); |
| } |
| |
| void AndroidVideoDecodeAccelerator::OnKeyAdded() { |
| DVLOG(1) << __FUNCTION__; |
| |
| if (state_ == WAITING_FOR_KEY) |
| state_ = NO_ERROR; |
| |
| DoIOTask(true); |
| } |
| |
| void AndroidVideoDecodeAccelerator::NotifyInitializationComplete(bool success) { |
| if (client_) |
| client_->NotifyInitializationComplete(success); |
| } |
| |
| void AndroidVideoDecodeAccelerator::NotifyPictureReady(const Picture& picture) { |
| if (client_) |
| client_->PictureReady(picture); |
| } |
| |
| void AndroidVideoDecodeAccelerator::NotifyEndOfBitstreamBuffer( |
| int input_buffer_id) { |
| if (client_) |
| client_->NotifyEndOfBitstreamBuffer(input_buffer_id); |
| } |
| |
| void AndroidVideoDecodeAccelerator::NotifyFlushDone() { |
| if (client_) |
| client_->NotifyFlushDone(); |
| } |
| |
| void AndroidVideoDecodeAccelerator::NotifyResetDone() { |
| if (client_) |
| client_->NotifyResetDone(); |
| } |
| |
| void AndroidVideoDecodeAccelerator::NotifyError( |
| VideoDecodeAccelerator::Error error, |
| int token) { |
| DVLOG(1) << __FUNCTION__ << ": error: " << error << " token: " << token |
| << " current: " << error_sequence_token_; |
| if (token != error_sequence_token_) |
| return; |
| |
| if (client_) |
| client_->NotifyError(error); |
| } |
| |
| void AndroidVideoDecodeAccelerator::ManageTimer(bool did_work) { |
| bool should_be_running = true; |
| |
| base::TimeTicks now = base::TimeTicks::Now(); |
| if (!did_work && !most_recent_work_.is_null()) { |
| // Make sure that we have done work recently enough, else stop the timer. |
| if (now - most_recent_work_ > IdleTimerTimeOut) { |
| most_recent_work_ = base::TimeTicks(); |
| should_be_running = false; |
| } |
| } else { |
| most_recent_work_ = now; |
| } |
| |
| if (should_be_running) |
| g_avda_timer.Pointer()->StartTimer(this); |
| else |
| g_avda_timer.Pointer()->StopTimer(this); |
| } |
| |
| void AndroidVideoDecodeAccelerator::ReleaseMediaCodec() { |
| if (!media_codec_) |
| return; |
| |
| // If codec construction is broken, then we can't release this codec if it's |
| // backed by hardware, else it may hang too. Post it to the construction |
| // thread, and it'll get freed if things start working. If things are |
| // already working, then it'll be freed soon. |
| // |
| // We require software codecs when |allow_autodetection_| is false, so use |
| // the stored value as a proxy for whether the MediaCodec is software backed |
| // or not. |
| if (!codec_config_->allow_autodetection_) { |
| media_codec_.reset(); |
| } else { |
| g_avda_timer.Pointer()->StartUsingConstructionThread(); |
| scoped_refptr<base::SingleThreadTaskRunner> task_runner = |
| g_avda_timer.Pointer()->ConstructionTaskRunner(); |
| task_runner->DeleteSoon(FROM_HERE, media_codec_.release()); |
| task_runner->PostTask( |
| FROM_HERE, base::Bind(&AVDATimerManager::DoneUsingConstructionThread, |
| base::Unretained(g_avda_timer.Pointer()))); |
| } |
| } |
| |
| // static |
| bool AndroidVideoDecodeAccelerator::UseDeferredRenderingStrategy( |
| const gpu::GpuPreferences& gpu_preferences) { |
| return true; |
| } |
| |
| // static |
| bool AndroidVideoDecodeAccelerator::UseTextureCopyForDeferredStrategy( |
| const gpu::GpuPreferences& gpu_preferences) { |
| // http://crbug.com/582170 |
| return gpu_preferences.enable_threaded_texture_mailboxes; |
| } |
| |
| // static |
| VideoDecodeAccelerator::Capabilities |
| AndroidVideoDecodeAccelerator::GetCapabilities( |
| const gpu::GpuPreferences& gpu_preferences) { |
| Capabilities capabilities; |
| SupportedProfiles& profiles = capabilities.supported_profiles; |
| |
| // Only support VP8 on Android versions where we don't have to synchronously |
| // tear down the MediaCodec on surface destruction because VP8 requires |
| // us to completely drain the decoder before releasing it, which is difficult |
| // and time consuming to do while the surface is being destroyed. |
| if (base::android::BuildInfo::GetInstance()->sdk_int() >= 18 && |
| MediaCodecUtil::IsVp8DecoderAvailable()) { |
| SupportedProfile profile; |
| profile.profile = VP8PROFILE_ANY; |
| profile.min_resolution.SetSize(0, 0); |
| profile.max_resolution.SetSize(3840, 2160); |
| // If we know MediaCodec will just create a software codec, prefer our |
| // internal software decoder instead. It's more up to date and secured |
| // within the renderer sandbox. However if the content is encrypted, we |
| // must use MediaCodec anyways since MediaDrm offers no way to decrypt |
| // the buffers and let us use our internal software decoders. |
| profile.encrypted_only = |
| VideoCodecBridge::IsKnownUnaccelerated(kCodecVP8, MEDIA_CODEC_DECODER); |
| profiles.push_back(profile); |
| } |
| |
| if (MediaCodecUtil::IsVp9DecoderAvailable()) { |
| SupportedProfile profile; |
| profile.min_resolution.SetSize(0, 0); |
| profile.max_resolution.SetSize(3840, 2160); |
| // If we know MediaCodec will just create a software codec, prefer our |
| // internal software decoder instead. It's more up to date and secured |
| // within the renderer sandbox. However if the content is encrypted, we |
| // must use MediaCodec anyways since MediaDrm offers no way to decrypt |
| // the buffers and let us use our internal software decoders. |
| profile.encrypted_only = |
| VideoCodecBridge::IsKnownUnaccelerated(kCodecVP9, MEDIA_CODEC_DECODER); |
| profile.profile = VP9PROFILE_PROFILE0; |
| profiles.push_back(profile); |
| profile.profile = VP9PROFILE_PROFILE1; |
| profiles.push_back(profile); |
| profile.profile = VP9PROFILE_PROFILE2; |
| profiles.push_back(profile); |
| profile.profile = VP9PROFILE_PROFILE3; |
| profiles.push_back(profile); |
| } |
| |
| for (const auto& supported_profile : kSupportedH264Profiles) { |
| SupportedProfile profile; |
| profile.profile = supported_profile; |
| profile.min_resolution.SetSize(0, 0); |
| // Advertise support for 4k and let the MediaCodec fail when decoding if it |
| // doesn't support the resolution. It's assumed that consumers won't have |
| // software fallback for H264 on Android anyway. |
| profile.max_resolution.SetSize(3840, 2160); |
| profiles.push_back(profile); |
| } |
| |
| capabilities.flags = |
| VideoDecodeAccelerator::Capabilities::SUPPORTS_DEFERRED_INITIALIZATION; |
| if (UseDeferredRenderingStrategy(gpu_preferences)) { |
| capabilities.flags |= VideoDecodeAccelerator::Capabilities:: |
| NEEDS_ALL_PICTURE_BUFFERS_TO_DECODE; |
| if (UseTextureCopyForDeferredStrategy(gpu_preferences)) { |
| capabilities.flags |= |
| media::VideoDecodeAccelerator::Capabilities::REQUIRES_TEXTURE_COPY; |
| } else if (media::MediaCodecUtil::IsSurfaceViewOutputSupported()) { |
| // Fullscreen external SurfaceView is disabled for WebView. |
| // http://crbug.com/582170 |
| capabilities.flags |= media::VideoDecodeAccelerator::Capabilities:: |
| SUPPORTS_EXTERNAL_OUTPUT_SURFACE; |
| } |
| } |
| |
| return capabilities; |
| } |
| |
| bool AndroidVideoDecodeAccelerator::IsMediaCodecSoftwareDecodingForbidden() |
| const { |
| // Prevent MediaCodec from using its internal software decoders when we have |
| // more secure and up to date versions in the renderer process. |
| return !config_.is_encrypted && (codec_config_->codec_ == media::kCodecVP8 || |
| codec_config_->codec_ == media::kCodecVP9); |
| } |
| |
| } // namespace media |