| // Copyright 2014 The Chromium Authors. All rights reserved. |
| // Use of this source code is governed by a BSD-style license that can be |
| // found in the LICENSE file. |
| |
| #include "media/renderers/renderer_impl.h" |
| |
| #include <utility> |
| |
| #include "base/bind.h" |
| #include "base/callback.h" |
| #include "base/callback_helpers.h" |
| #include "base/command_line.h" |
| #include "base/compiler_specific.h" |
| #include "base/location.h" |
| #include "base/single_thread_task_runner.h" |
| #include "base/strings/string_number_conversions.h" |
| #include "media/base/audio_decoder_config.h" |
| #include "media/base/audio_renderer.h" |
| #include "media/base/bind_to_current_loop.h" |
| #include "media/base/media_log.h" |
| #include "media/base/media_resource.h" |
| #include "media/base/media_switches.h" |
| #include "media/base/renderer_client.h" |
| #include "media/base/time_source.h" |
| #include "media/base/video_decoder_config.h" |
| #include "media/base/video_renderer.h" |
| #include "media/base/wall_clock_time_source.h" |
| |
| namespace media { |
| |
| // See |video_underflow_threshold_|. |
| static const int kDefaultVideoUnderflowThresholdMs = 3000; |
| |
| static const int kAudioRestartUnderflowThresholdMs = 2000; |
| |
| class RendererImpl::RendererClientInternal : public RendererClient { |
| public: |
| RendererClientInternal(DemuxerStream::Type type, RendererImpl* renderer) |
| : type_(type), renderer_(renderer) { |
| DCHECK((type_ == DemuxerStream::AUDIO) || (type_ == DemuxerStream::VIDEO)); |
| } |
| |
| void OnError(PipelineStatus error) override { renderer_->OnError(error); } |
| void OnEnded() override { renderer_->OnRendererEnded(type_); } |
| void OnStatisticsUpdate(const PipelineStatistics& stats) override { |
| renderer_->OnStatisticsUpdate(stats); |
| } |
| void OnBufferingStateChange(BufferingState state) override { |
| renderer_->OnBufferingStateChange(type_, state); |
| } |
| void OnWaitingForDecryptionKey() override { |
| renderer_->OnWaitingForDecryptionKey(); |
| } |
| void OnVideoNaturalSizeChange(const gfx::Size& size) override { |
| DCHECK(type_ == DemuxerStream::VIDEO); |
| renderer_->OnVideoNaturalSizeChange(size); |
| } |
| void OnVideoOpacityChange(bool opaque) override { |
| DCHECK(type_ == DemuxerStream::VIDEO); |
| renderer_->OnVideoOpacityChange(opaque); |
| } |
| void OnDurationChange(base::TimeDelta duration) override { |
| // RendererClients should only be notified of duration changes in certain |
| // scenarios, none of which should arise for RendererClientInternal. |
| // Duration changes should be sent to the pipeline by the DemuxerStream, via |
| // the DemuxerHost interface. |
| NOTREACHED(); |
| } |
| |
| private: |
| DemuxerStream::Type type_; |
| RendererImpl* renderer_; |
| }; |
| |
| RendererImpl::RendererImpl( |
| const scoped_refptr<base::SingleThreadTaskRunner>& task_runner, |
| std::unique_ptr<AudioRenderer> audio_renderer, |
| std::unique_ptr<VideoRenderer> video_renderer) |
| : state_(STATE_UNINITIALIZED), |
| task_runner_(task_runner), |
| audio_renderer_(std::move(audio_renderer)), |
| video_renderer_(std::move(video_renderer)), |
| current_audio_stream_(nullptr), |
| current_video_stream_(nullptr), |
| time_source_(NULL), |
| time_ticking_(false), |
| playback_rate_(0.0), |
| audio_buffering_state_(BUFFERING_HAVE_NOTHING), |
| video_buffering_state_(BUFFERING_HAVE_NOTHING), |
| audio_ended_(false), |
| video_ended_(false), |
| cdm_context_(nullptr), |
| underflow_disabled_for_testing_(false), |
| clockless_video_playback_enabled_for_testing_(false), |
| video_underflow_threshold_( |
| base::TimeDelta::FromMilliseconds(kDefaultVideoUnderflowThresholdMs)), |
| weak_factory_(this) { |
| weak_this_ = weak_factory_.GetWeakPtr(); |
| DVLOG(1) << __func__; |
| |
| // TODO(dalecurtis): Remove once experiments for http://crbug.com/470940 are |
| // complete. |
| int threshold_ms = 0; |
| std::string threshold_ms_str( |
| base::CommandLine::ForCurrentProcess()->GetSwitchValueASCII( |
| switches::kVideoUnderflowThresholdMs)); |
| if (base::StringToInt(threshold_ms_str, &threshold_ms) && threshold_ms > 0) { |
| video_underflow_threshold_ = |
| base::TimeDelta::FromMilliseconds(threshold_ms); |
| } |
| } |
| |
| RendererImpl::~RendererImpl() { |
| DVLOG(1) << __func__; |
| DCHECK(task_runner_->BelongsToCurrentThread()); |
| |
| // RendererImpl is being destroyed, so invalidate weak pointers right away to |
| // avoid getting callbacks which might try to access fields that has been |
| // destroyed, e.g. audio_renderer_/video_renderer_ below (crbug.com/668963). |
| weak_factory_.InvalidateWeakPtrs(); |
| |
| // Tear down in opposite order of construction as |video_renderer_| can still |
| // need |time_source_| (which can be |audio_renderer_|) to be alive. |
| video_renderer_.reset(); |
| audio_renderer_.reset(); |
| |
| if (!init_cb_.is_null()) { |
| FinishInitialization(PIPELINE_ERROR_ABORT); |
| } else if (!flush_cb_.is_null()) { |
| base::ResetAndReturn(&flush_cb_).Run(); |
| } |
| } |
| |
| void RendererImpl::Initialize(MediaResource* media_resource, |
| RendererClient* client, |
| const PipelineStatusCB& init_cb) { |
| DVLOG(1) << __func__; |
| DCHECK(task_runner_->BelongsToCurrentThread()); |
| DCHECK_EQ(state_, STATE_UNINITIALIZED); |
| DCHECK(!init_cb.is_null()); |
| DCHECK(client); |
| |
| client_ = client; |
| media_resource_ = media_resource; |
| init_cb_ = init_cb; |
| |
| if (HasEncryptedStream() && !cdm_context_) { |
| DVLOG(1) << __func__ << ": Has encrypted stream but CDM is not set."; |
| state_ = STATE_INIT_PENDING_CDM; |
| return; |
| } |
| |
| state_ = STATE_INITIALIZING; |
| InitializeAudioRenderer(); |
| } |
| |
| void RendererImpl::SetCdm(CdmContext* cdm_context, |
| const CdmAttachedCB& cdm_attached_cb) { |
| DVLOG(1) << __func__; |
| DCHECK(task_runner_->BelongsToCurrentThread()); |
| DCHECK(cdm_context); |
| |
| if (cdm_context_) { |
| DVLOG(1) << "Switching CDM not supported."; |
| cdm_attached_cb.Run(false); |
| return; |
| } |
| |
| cdm_context_ = cdm_context; |
| |
| if (state_ != STATE_INIT_PENDING_CDM) { |
| cdm_attached_cb.Run(true); |
| return; |
| } |
| |
| DCHECK(!init_cb_.is_null()); |
| state_ = STATE_INITIALIZING; |
| // |cdm_attached_cb| will be fired after initialization finishes. |
| pending_cdm_attached_cb_ = cdm_attached_cb; |
| |
| InitializeAudioRenderer(); |
| } |
| |
| void RendererImpl::Flush(const base::Closure& flush_cb) { |
| DVLOG(1) << __func__; |
| DCHECK(task_runner_->BelongsToCurrentThread()); |
| DCHECK(flush_cb_.is_null()); |
| |
| if (state_ == STATE_FLUSHED) { |
| task_runner_->PostTask(FROM_HERE, flush_cb); |
| return; |
| } |
| |
| if (state_ != STATE_PLAYING) { |
| DCHECK_EQ(state_, STATE_ERROR); |
| return; |
| } |
| |
| flush_cb_ = flush_cb; |
| state_ = STATE_FLUSHING; |
| |
| // If we are currently handling a media stream status change, then postpone |
| // Flush until after that's done (because stream status changes also flush |
| // audio_renderer_/video_renderer_ and they need to be restarted before they |
| // can be flushed again). OnStreamRestartCompleted will resume Flush |
| // processing after audio/video restart has completed and there are no other |
| // pending stream status changes. |
| if (restarting_audio_ || restarting_video_) { |
| pending_actions_.push_back( |
| base::Bind(&RendererImpl::FlushInternal, weak_this_)); |
| return; |
| } |
| |
| FlushInternal(); |
| } |
| |
| void RendererImpl::StartPlayingFrom(base::TimeDelta time) { |
| DVLOG(1) << __func__; |
| DCHECK(task_runner_->BelongsToCurrentThread()); |
| |
| if (state_ != STATE_FLUSHED) { |
| DCHECK_EQ(state_, STATE_ERROR); |
| return; |
| } |
| |
| time_source_->SetMediaTime(time); |
| |
| state_ = STATE_PLAYING; |
| if (audio_renderer_) |
| audio_renderer_->StartPlaying(); |
| if (video_renderer_) |
| video_renderer_->StartPlayingFrom(time); |
| } |
| |
| void RendererImpl::SetPlaybackRate(double playback_rate) { |
| DVLOG(1) << __func__ << "(" << playback_rate << ")"; |
| DCHECK(task_runner_->BelongsToCurrentThread()); |
| |
| // Playback rate changes are only carried out while playing. |
| if (state_ != STATE_PLAYING && state_ != STATE_FLUSHED) |
| return; |
| |
| time_source_->SetPlaybackRate(playback_rate); |
| |
| const double old_rate = playback_rate_; |
| playback_rate_ = playback_rate; |
| if (!time_ticking_ || !video_renderer_) |
| return; |
| |
| if (old_rate == 0 && playback_rate > 0) |
| video_renderer_->OnTimeProgressing(); |
| else if (old_rate > 0 && playback_rate == 0) |
| video_renderer_->OnTimeStopped(); |
| } |
| |
| void RendererImpl::SetVolume(float volume) { |
| DVLOG(1) << __func__; |
| DCHECK(task_runner_->BelongsToCurrentThread()); |
| |
| if (audio_renderer_) |
| audio_renderer_->SetVolume(volume); |
| } |
| |
| base::TimeDelta RendererImpl::GetMediaTime() { |
| // No BelongsToCurrentThread() checking because this can be called from other |
| // threads. |
| return time_source_->CurrentMediaTime(); |
| } |
| |
| void RendererImpl::DisableUnderflowForTesting() { |
| DVLOG(1) << __func__; |
| DCHECK(task_runner_->BelongsToCurrentThread()); |
| DCHECK_EQ(state_, STATE_UNINITIALIZED); |
| |
| underflow_disabled_for_testing_ = true; |
| } |
| |
| void RendererImpl::EnableClocklessVideoPlaybackForTesting() { |
| DVLOG(1) << __func__; |
| DCHECK(task_runner_->BelongsToCurrentThread()); |
| DCHECK_EQ(state_, STATE_UNINITIALIZED); |
| DCHECK(underflow_disabled_for_testing_) |
| << "Underflow must be disabled for clockless video playback"; |
| |
| clockless_video_playback_enabled_for_testing_ = true; |
| } |
| |
| bool RendererImpl::GetWallClockTimes( |
| const std::vector<base::TimeDelta>& media_timestamps, |
| std::vector<base::TimeTicks>* wall_clock_times) { |
| // No BelongsToCurrentThread() checking because this can be called from other |
| // threads. |
| // |
| // TODO(scherkus): Currently called from VideoRendererImpl's internal thread, |
| // which should go away at some point http://crbug.com/110814 |
| if (clockless_video_playback_enabled_for_testing_) { |
| if (media_timestamps.empty()) { |
| *wall_clock_times = std::vector<base::TimeTicks>(1, |
| base::TimeTicks::Now()); |
| } else { |
| *wall_clock_times = std::vector<base::TimeTicks>(); |
| for (auto const &media_time : media_timestamps) { |
| wall_clock_times->push_back(base::TimeTicks() + media_time); |
| } |
| } |
| return true; |
| } |
| |
| return time_source_->GetWallClockTimes(media_timestamps, wall_clock_times); |
| } |
| |
| bool RendererImpl::HasEncryptedStream() { |
| std::vector<DemuxerStream*> demuxer_streams = |
| media_resource_->GetAllStreams(); |
| |
| for (auto* stream : demuxer_streams) { |
| if (stream->type() == DemuxerStream::AUDIO && |
| stream->audio_decoder_config().is_encrypted()) |
| return true; |
| if (stream->type() == DemuxerStream::VIDEO && |
| stream->video_decoder_config().is_encrypted()) |
| return true; |
| } |
| |
| return false; |
| } |
| |
| void RendererImpl::FinishInitialization(PipelineStatus status) { |
| DCHECK(!init_cb_.is_null()); |
| |
| if (!pending_cdm_attached_cb_.is_null()) |
| base::ResetAndReturn(&pending_cdm_attached_cb_).Run(status == PIPELINE_OK); |
| |
| base::ResetAndReturn(&init_cb_).Run(status); |
| } |
| |
| void RendererImpl::InitializeAudioRenderer() { |
| DVLOG(1) << __func__; |
| DCHECK(task_runner_->BelongsToCurrentThread()); |
| DCHECK_EQ(state_, STATE_INITIALIZING); |
| DCHECK(!init_cb_.is_null()); |
| |
| PipelineStatusCB done_cb = |
| base::Bind(&RendererImpl::OnAudioRendererInitializeDone, weak_this_); |
| |
| // TODO(servolk): Implement proper support for multiple streams. But for now |
| // pick the first enabled stream to preserve the existing behavior. |
| DemuxerStream* audio_stream = |
| media_resource_->GetFirstStream(DemuxerStream::AUDIO); |
| if (!audio_stream) { |
| audio_renderer_.reset(); |
| task_runner_->PostTask(FROM_HERE, base::Bind(done_cb, PIPELINE_OK)); |
| return; |
| } |
| |
| current_audio_stream_ = audio_stream; |
| |
| audio_renderer_client_.reset( |
| new RendererClientInternal(DemuxerStream::AUDIO, this)); |
| // Note: After the initialization of a renderer, error events from it may |
| // happen at any time and all future calls must guard against STATE_ERROR. |
| audio_renderer_->Initialize(audio_stream, cdm_context_, |
| audio_renderer_client_.get(), done_cb); |
| } |
| |
| void RendererImpl::OnAudioRendererInitializeDone(PipelineStatus status) { |
| DVLOG(1) << __func__ << ": " << status; |
| DCHECK(task_runner_->BelongsToCurrentThread()); |
| |
| // OnError() may be fired at any time by the renderers, even if they thought |
| // they initialized successfully (due to delayed output device setup). |
| if (state_ != STATE_INITIALIZING) { |
| DCHECK(init_cb_.is_null()); |
| audio_renderer_.reset(); |
| return; |
| } |
| |
| if (status != PIPELINE_OK) { |
| FinishInitialization(status); |
| return; |
| } |
| |
| DCHECK(!init_cb_.is_null()); |
| InitializeVideoRenderer(); |
| } |
| |
| void RendererImpl::InitializeVideoRenderer() { |
| DVLOG(1) << __func__; |
| DCHECK(task_runner_->BelongsToCurrentThread()); |
| DCHECK_EQ(state_, STATE_INITIALIZING); |
| DCHECK(!init_cb_.is_null()); |
| |
| PipelineStatusCB done_cb = |
| base::Bind(&RendererImpl::OnVideoRendererInitializeDone, weak_this_); |
| |
| // TODO(servolk): Implement proper support for multiple streams. But for now |
| // pick the first enabled stream to preserve the existing behavior. |
| DemuxerStream* video_stream = |
| media_resource_->GetFirstStream(DemuxerStream::VIDEO); |
| if (!video_stream) { |
| video_renderer_.reset(); |
| task_runner_->PostTask(FROM_HERE, base::Bind(done_cb, PIPELINE_OK)); |
| return; |
| } |
| |
| current_video_stream_ = video_stream; |
| |
| video_renderer_client_.reset( |
| new RendererClientInternal(DemuxerStream::VIDEO, this)); |
| video_renderer_->Initialize( |
| video_stream, cdm_context_, video_renderer_client_.get(), |
| base::Bind(&RendererImpl::GetWallClockTimes, base::Unretained(this)), |
| done_cb); |
| } |
| |
| void RendererImpl::OnVideoRendererInitializeDone(PipelineStatus status) { |
| DVLOG(1) << __func__ << ": " << status; |
| DCHECK(task_runner_->BelongsToCurrentThread()); |
| |
| // OnError() may be fired at any time by the renderers, even if they thought |
| // they initialized successfully (due to delayed output device setup). |
| if (state_ != STATE_INITIALIZING) { |
| DCHECK(init_cb_.is_null()); |
| audio_renderer_.reset(); |
| video_renderer_.reset(); |
| return; |
| } |
| |
| DCHECK(!init_cb_.is_null()); |
| |
| if (status != PIPELINE_OK) { |
| FinishInitialization(status); |
| return; |
| } |
| |
| media_resource_->SetStreamStatusChangeCB( |
| base::Bind(&RendererImpl::OnStreamStatusChanged, weak_this_)); |
| |
| if (audio_renderer_) { |
| time_source_ = audio_renderer_->GetTimeSource(); |
| } else if (!time_source_) { |
| wall_clock_time_source_.reset(new WallClockTimeSource()); |
| time_source_ = wall_clock_time_source_.get(); |
| } |
| |
| state_ = STATE_FLUSHED; |
| DCHECK(time_source_); |
| DCHECK(audio_renderer_ || video_renderer_); |
| |
| FinishInitialization(PIPELINE_OK); |
| } |
| |
| void RendererImpl::FlushInternal() { |
| DVLOG(1) << __func__; |
| DCHECK(task_runner_->BelongsToCurrentThread()); |
| DCHECK_EQ(state_, STATE_FLUSHING); |
| DCHECK(!flush_cb_.is_null()); |
| |
| if (time_ticking_) |
| PausePlayback(); |
| |
| FlushAudioRenderer(); |
| } |
| |
| void RendererImpl::FlushAudioRenderer() { |
| DVLOG(1) << __func__; |
| DCHECK(task_runner_->BelongsToCurrentThread()); |
| DCHECK_EQ(state_, STATE_FLUSHING); |
| DCHECK(!flush_cb_.is_null()); |
| |
| if (!audio_renderer_) { |
| OnAudioRendererFlushDone(); |
| return; |
| } |
| |
| audio_renderer_->Flush( |
| base::Bind(&RendererImpl::OnAudioRendererFlushDone, weak_this_)); |
| } |
| |
| void RendererImpl::OnAudioRendererFlushDone() { |
| DVLOG(1) << __func__; |
| DCHECK(task_runner_->BelongsToCurrentThread()); |
| |
| if (state_ == STATE_ERROR) { |
| DCHECK(flush_cb_.is_null()); |
| return; |
| } |
| |
| DCHECK_EQ(state_, STATE_FLUSHING); |
| DCHECK(!flush_cb_.is_null()); |
| |
| // If we had a deferred video renderer underflow prior to the flush, it should |
| // have been cleared by the audio renderer changing to BUFFERING_HAVE_NOTHING. |
| DCHECK(deferred_video_underflow_cb_.IsCancelled()); |
| |
| DCHECK_EQ(audio_buffering_state_, BUFFERING_HAVE_NOTHING); |
| audio_ended_ = false; |
| FlushVideoRenderer(); |
| } |
| |
| void RendererImpl::FlushVideoRenderer() { |
| DVLOG(1) << __func__; |
| DCHECK(task_runner_->BelongsToCurrentThread()); |
| DCHECK_EQ(state_, STATE_FLUSHING); |
| DCHECK(!flush_cb_.is_null()); |
| |
| if (!video_renderer_) { |
| OnVideoRendererFlushDone(); |
| return; |
| } |
| |
| video_renderer_->Flush( |
| base::Bind(&RendererImpl::OnVideoRendererFlushDone, weak_this_)); |
| } |
| |
| void RendererImpl::OnVideoRendererFlushDone() { |
| DVLOG(1) << __func__; |
| DCHECK(task_runner_->BelongsToCurrentThread()); |
| |
| if (state_ == STATE_ERROR) { |
| DCHECK(flush_cb_.is_null()); |
| return; |
| } |
| |
| DCHECK_EQ(state_, STATE_FLUSHING); |
| DCHECK(!flush_cb_.is_null()); |
| |
| DCHECK_EQ(video_buffering_state_, BUFFERING_HAVE_NOTHING); |
| video_ended_ = false; |
| state_ = STATE_FLUSHED; |
| base::ResetAndReturn(&flush_cb_).Run(); |
| |
| if (!pending_actions_.empty()) { |
| base::Closure closure = pending_actions_.front(); |
| pending_actions_.pop_front(); |
| closure.Run(); |
| } |
| } |
| |
| void RendererImpl::OnStreamStatusChanged(DemuxerStream* stream, |
| bool enabled, |
| base::TimeDelta time) { |
| DCHECK(task_runner_->BelongsToCurrentThread()); |
| DCHECK(stream); |
| bool video = (stream->type() == DemuxerStream::VIDEO); |
| DVLOG(1) << __func__ << (video ? " video" : " audio") << " stream=" << stream |
| << " enabled=" << enabled << " time=" << time.InSecondsF(); |
| |
| if ((state_ != STATE_PLAYING && state_ != STATE_FLUSHING && |
| state_ != STATE_FLUSHED) || |
| (audio_ended_ && video_ended_)) |
| return; |
| |
| if (restarting_audio_ || restarting_video_ || state_ == STATE_FLUSHING) { |
| DVLOG(3) << __func__ << ": postponed stream " << stream |
| << " status change handling."; |
| pending_actions_.push_back(base::Bind(&RendererImpl::OnStreamStatusChanged, |
| weak_this_, stream, enabled, time)); |
| return; |
| } |
| |
| DCHECK(state_ == STATE_PLAYING || state_ == STATE_FLUSHED); |
| if (stream->type() == DemuxerStream::VIDEO) { |
| DCHECK(video_renderer_); |
| restarting_video_ = true; |
| base::Closure handle_track_status_cb = |
| base::Bind(stream == current_video_stream_ |
| ? &RendererImpl::RestartVideoRenderer |
| : &RendererImpl::ReinitializeVideoRenderer, |
| weak_this_, stream, time); |
| if (state_ == STATE_FLUSHED) |
| handle_track_status_cb.Run(); |
| else |
| video_renderer_->Flush(handle_track_status_cb); |
| } else if (stream->type() == DemuxerStream::AUDIO) { |
| DCHECK(audio_renderer_); |
| DCHECK(time_source_); |
| restarting_audio_ = true; |
| base::Closure handle_track_status_cb = |
| base::Bind(stream == current_audio_stream_ |
| ? &RendererImpl::RestartAudioRenderer |
| : &RendererImpl::ReinitializeAudioRenderer, |
| weak_this_, stream, time); |
| if (state_ == STATE_FLUSHED) { |
| handle_track_status_cb.Run(); |
| return; |
| } |
| // Stop ticking (transition into paused state) in audio renderer before |
| // calling Flush, since after Flush we are going to restart playback by |
| // calling audio renderer StartPlaying which would fail in playing state. |
| if (time_ticking_) { |
| time_ticking_ = false; |
| time_source_->StopTicking(); |
| } |
| audio_renderer_->Flush(handle_track_status_cb); |
| } |
| } |
| |
| void RendererImpl::ReinitializeAudioRenderer(DemuxerStream* stream, |
| base::TimeDelta time) { |
| DVLOG(2) << __func__ << " stream=" << stream << " time=" << time.InSecondsF(); |
| DCHECK(task_runner_->BelongsToCurrentThread()); |
| DCHECK_NE(stream, current_audio_stream_); |
| |
| current_audio_stream_ = stream; |
| audio_renderer_->Initialize( |
| stream, cdm_context_, audio_renderer_client_.get(), |
| base::Bind(&RendererImpl::OnAudioRendererReinitialized, weak_this_, |
| stream, time)); |
| } |
| |
| void RendererImpl::OnAudioRendererReinitialized(DemuxerStream* stream, |
| base::TimeDelta time, |
| PipelineStatus status) { |
| DVLOG(2) << __func__ << ": status=" << status; |
| DCHECK_EQ(stream, current_audio_stream_); |
| |
| if (status != PIPELINE_OK) { |
| OnError(status); |
| return; |
| } |
| RestartAudioRenderer(stream, time); |
| } |
| |
| void RendererImpl::ReinitializeVideoRenderer(DemuxerStream* stream, |
| base::TimeDelta time) { |
| DVLOG(2) << __func__ << " stream=" << stream << " time=" << time.InSecondsF(); |
| DCHECK(task_runner_->BelongsToCurrentThread()); |
| DCHECK_NE(stream, current_video_stream_); |
| |
| current_video_stream_ = stream; |
| video_renderer_->OnTimeStopped(); |
| video_renderer_->Initialize( |
| stream, cdm_context_, video_renderer_client_.get(), |
| base::Bind(&RendererImpl::GetWallClockTimes, base::Unretained(this)), |
| base::Bind(&RendererImpl::OnVideoRendererReinitialized, weak_this_, |
| stream, time)); |
| } |
| |
| void RendererImpl::OnVideoRendererReinitialized(DemuxerStream* stream, |
| base::TimeDelta time, |
| PipelineStatus status) { |
| DVLOG(2) << __func__ << ": status=" << status; |
| DCHECK_EQ(stream, current_video_stream_); |
| |
| if (status != PIPELINE_OK) { |
| OnError(status); |
| return; |
| } |
| RestartVideoRenderer(stream, time); |
| } |
| |
| void RendererImpl::RestartAudioRenderer(DemuxerStream* stream, |
| base::TimeDelta time) { |
| DVLOG(2) << __func__ << " stream=" << stream << " time=" << time.InSecondsF(); |
| DCHECK(task_runner_->BelongsToCurrentThread()); |
| DCHECK(state_ == STATE_PLAYING || state_ == STATE_FLUSHED || |
| state_ == STATE_FLUSHING); |
| DCHECK(time_source_); |
| DCHECK(audio_renderer_); |
| DCHECK_EQ(stream, current_audio_stream_); |
| |
| audio_ended_ = false; |
| if (state_ == STATE_FLUSHED) { |
| // If we are in the FLUSHED state, then we are done. The audio renderer will |
| // be restarted by a subsequent RendererImpl::StartPlayingFrom call. |
| OnStreamRestartCompleted(); |
| } else { |
| // Stream restart will be completed when the audio renderer decodes enough |
| // data and reports HAVE_ENOUGH to HandleRestartedStreamBufferingChanges. |
| audio_renderer_->StartPlaying(); |
| } |
| } |
| |
| void RendererImpl::RestartVideoRenderer(DemuxerStream* stream, |
| base::TimeDelta time) { |
| DVLOG(2) << __func__ << " stream=" << stream << " time=" << time.InSecondsF(); |
| DCHECK(task_runner_->BelongsToCurrentThread()); |
| DCHECK(video_renderer_); |
| DCHECK(state_ == STATE_PLAYING || state_ == STATE_FLUSHED || |
| state_ == STATE_FLUSHING); |
| DCHECK_EQ(stream, current_video_stream_); |
| |
| video_ended_ = false; |
| if (state_ == STATE_FLUSHED) { |
| // If we are in the FLUSHED state, then we are done. The video renderer will |
| // be restarted by a subsequent RendererImpl::StartPlayingFrom call. |
| OnStreamRestartCompleted(); |
| } else { |
| // Stream restart will be completed when the video renderer decodes enough |
| // data and reports HAVE_ENOUGH to HandleRestartedStreamBufferingChanges. |
| video_renderer_->StartPlayingFrom(time); |
| } |
| } |
| |
| void RendererImpl::OnStatisticsUpdate(const PipelineStatistics& stats) { |
| DCHECK(task_runner_->BelongsToCurrentThread()); |
| client_->OnStatisticsUpdate(stats); |
| } |
| |
| bool RendererImpl::HandleRestartedStreamBufferingChanges( |
| DemuxerStream::Type type, |
| BufferingState new_buffering_state) { |
| DCHECK(task_runner_->BelongsToCurrentThread()); |
| // When restarting playback we want to defer the BUFFERING_HAVE_NOTHING for |
| // the stream being restarted, to allow continuing uninterrupted playback on |
| // the other stream. |
| if (type == DemuxerStream::VIDEO && restarting_video_) { |
| if (new_buffering_state == BUFFERING_HAVE_ENOUGH) { |
| DVLOG(1) << __func__ << " Got BUFFERING_HAVE_ENOUGH for video stream," |
| " resuming playback."; |
| task_runner_->PostTask( |
| FROM_HERE, |
| base::Bind(&RendererImpl::OnStreamRestartCompleted, weak_this_)); |
| if (state_ == STATE_PLAYING && |
| !deferred_video_underflow_cb_.IsCancelled()) { |
| // If deferred_video_underflow_cb_ wasn't triggered, then audio should |
| // still be playing, we only need to unpause the video stream. |
| DVLOG(4) << "deferred_video_underflow_cb_.Cancel()"; |
| deferred_video_underflow_cb_.Cancel(); |
| video_buffering_state_ = new_buffering_state; |
| if (playback_rate_ > 0) |
| video_renderer_->OnTimeProgressing(); |
| return true; |
| } |
| } |
| // We don't handle the BUFFERING_HAVE_NOTHING case explicitly here, since |
| // the existing logic for deferring video underflow reporting in |
| // OnBufferingStateChange is exactly what we need. So fall through to the |
| // regular video underflow handling path in OnBufferingStateChange. |
| } |
| |
| if (type == DemuxerStream::AUDIO && restarting_audio_) { |
| if (new_buffering_state == BUFFERING_HAVE_NOTHING) { |
| if (deferred_video_underflow_cb_.IsCancelled() && |
| deferred_audio_restart_underflow_cb_.IsCancelled()) { |
| DVLOG(1) << __func__ << " Deferring BUFFERING_HAVE_NOTHING for " |
| "audio stream which is being restarted."; |
| audio_buffering_state_ = new_buffering_state; |
| deferred_audio_restart_underflow_cb_.Reset( |
| base::Bind(&RendererImpl::OnBufferingStateChange, weak_this_, type, |
| new_buffering_state)); |
| task_runner_->PostDelayedTask( |
| FROM_HERE, deferred_audio_restart_underflow_cb_.callback(), |
| base::TimeDelta::FromMilliseconds( |
| kAudioRestartUnderflowThresholdMs)); |
| return true; |
| } |
| // Cancel the deferred callback and report the underflow immediately. |
| DVLOG(4) << "deferred_audio_restart_underflow_cb_.Cancel()"; |
| deferred_audio_restart_underflow_cb_.Cancel(); |
| } else if (new_buffering_state == BUFFERING_HAVE_ENOUGH) { |
| DVLOG(1) << __func__ << " Got BUFFERING_HAVE_ENOUGH for audio stream," |
| " resuming playback."; |
| deferred_audio_restart_underflow_cb_.Cancel(); |
| // Now that we have decoded enough audio, pause playback momentarily to |
| // ensure video renderer is synchronised with audio. |
| PausePlayback(); |
| task_runner_->PostTask( |
| FROM_HERE, |
| base::Bind(&RendererImpl::OnStreamRestartCompleted, weak_this_)); |
| } |
| } |
| return false; |
| } |
| |
| void RendererImpl::OnStreamRestartCompleted() { |
| DVLOG(3) << __func__ << " restarting_audio_=" << restarting_audio_ |
| << " restarting_video_=" << restarting_video_; |
| DCHECK(task_runner_->BelongsToCurrentThread()); |
| DCHECK(restarting_audio_ || restarting_video_); |
| restarting_audio_ = false; |
| restarting_video_ = false; |
| if (!pending_actions_.empty()) { |
| base::Closure closure = pending_actions_.front(); |
| pending_actions_.pop_front(); |
| closure.Run(); |
| } |
| } |
| |
| void RendererImpl::OnBufferingStateChange(DemuxerStream::Type type, |
| BufferingState new_buffering_state) { |
| DCHECK((type == DemuxerStream::AUDIO) || (type == DemuxerStream::VIDEO)); |
| BufferingState* buffering_state = type == DemuxerStream::AUDIO |
| ? &audio_buffering_state_ |
| : &video_buffering_state_; |
| |
| DVLOG(1) << __func__ << (type == DemuxerStream::AUDIO ? " audio " : " video ") |
| << MediaLog::BufferingStateToString(*buffering_state) << " -> " |
| << MediaLog::BufferingStateToString(new_buffering_state); |
| DCHECK(task_runner_->BelongsToCurrentThread()); |
| |
| bool was_waiting_for_enough_data = WaitingForEnoughData(); |
| |
| if (restarting_audio_ || restarting_video_) { |
| if (HandleRestartedStreamBufferingChanges(type, new_buffering_state)) |
| return; |
| } |
| |
| // When audio is present and has enough data, defer video underflow callbacks |
| // for some time to avoid unnecessary glitches in audio; see |
| // http://crbug.com/144683#c53. |
| if (audio_renderer_ && type == DemuxerStream::VIDEO && |
| state_ == STATE_PLAYING) { |
| if (video_buffering_state_ == BUFFERING_HAVE_ENOUGH && |
| audio_buffering_state_ == BUFFERING_HAVE_ENOUGH && |
| new_buffering_state == BUFFERING_HAVE_NOTHING && |
| deferred_video_underflow_cb_.IsCancelled()) { |
| DVLOG(4) << __func__ << " Deferring HAVE_NOTHING for video stream."; |
| deferred_video_underflow_cb_.Reset( |
| base::Bind(&RendererImpl::OnBufferingStateChange, weak_this_, type, |
| new_buffering_state)); |
| task_runner_->PostDelayedTask(FROM_HERE, |
| deferred_video_underflow_cb_.callback(), |
| video_underflow_threshold_); |
| return; |
| } |
| |
| DVLOG(4) << "deferred_video_underflow_cb_.Cancel()"; |
| deferred_video_underflow_cb_.Cancel(); |
| } else if (!deferred_video_underflow_cb_.IsCancelled() && |
| type == DemuxerStream::AUDIO && |
| new_buffering_state == BUFFERING_HAVE_NOTHING) { |
| // If audio underflows while we have a deferred video underflow in progress |
| // we want to mark video as underflowed immediately and cancel the deferral. |
| deferred_video_underflow_cb_.Cancel(); |
| video_buffering_state_ = BUFFERING_HAVE_NOTHING; |
| } |
| |
| *buffering_state = new_buffering_state; |
| |
| // Disable underflow by ignoring updates that renderers have ran out of data. |
| if (state_ == STATE_PLAYING && underflow_disabled_for_testing_ && |
| time_ticking_) { |
| DVLOG(1) << "Update ignored because underflow is disabled for testing."; |
| return; |
| } |
| |
| // Renderer underflowed. |
| if (!was_waiting_for_enough_data && WaitingForEnoughData()) { |
| PausePlayback(); |
| client_->OnBufferingStateChange(BUFFERING_HAVE_NOTHING); |
| return; |
| } |
| |
| // Renderer prerolled. |
| if (was_waiting_for_enough_data && !WaitingForEnoughData()) { |
| StartPlayback(); |
| client_->OnBufferingStateChange(BUFFERING_HAVE_ENOUGH); |
| return; |
| } |
| } |
| |
| bool RendererImpl::WaitingForEnoughData() const { |
| DCHECK(task_runner_->BelongsToCurrentThread()); |
| if (state_ != STATE_PLAYING) |
| return false; |
| if (audio_renderer_ && audio_buffering_state_ != BUFFERING_HAVE_ENOUGH) |
| return true; |
| if (video_renderer_ && video_buffering_state_ != BUFFERING_HAVE_ENOUGH) |
| return true; |
| return false; |
| } |
| |
| void RendererImpl::PausePlayback() { |
| DVLOG(1) << __func__; |
| DCHECK(task_runner_->BelongsToCurrentThread()); |
| switch (state_) { |
| case STATE_PLAYING: |
| DCHECK(PlaybackHasEnded() || WaitingForEnoughData() || restarting_audio_) |
| << "Playback should only pause due to ending or underflowing or" |
| " when restarting audio stream"; |
| |
| break; |
| |
| case STATE_FLUSHING: |
| case STATE_FLUSHED: |
| // It's OK to pause playback when flushing. |
| break; |
| |
| case STATE_UNINITIALIZED: |
| case STATE_INIT_PENDING_CDM: |
| case STATE_INITIALIZING: |
| NOTREACHED() << "Invalid state: " << state_; |
| break; |
| |
| case STATE_ERROR: |
| // An error state may occur at any time. |
| break; |
| } |
| |
| if (time_ticking_) { |
| time_ticking_ = false; |
| time_source_->StopTicking(); |
| } |
| if (playback_rate_ > 0 && video_renderer_) |
| video_renderer_->OnTimeStopped(); |
| } |
| |
| void RendererImpl::StartPlayback() { |
| DVLOG(1) << __func__; |
| DCHECK(task_runner_->BelongsToCurrentThread()); |
| DCHECK_EQ(state_, STATE_PLAYING); |
| DCHECK(!time_ticking_); |
| DCHECK(!WaitingForEnoughData()); |
| |
| time_ticking_ = true; |
| time_source_->StartTicking(); |
| if (playback_rate_ > 0 && video_renderer_) |
| video_renderer_->OnTimeProgressing(); |
| } |
| |
| void RendererImpl::OnRendererEnded(DemuxerStream::Type type) { |
| DVLOG(1) << __func__ << (type == DemuxerStream::AUDIO ? " audio" : " video"); |
| DCHECK(task_runner_->BelongsToCurrentThread()); |
| DCHECK((type == DemuxerStream::AUDIO) || (type == DemuxerStream::VIDEO)); |
| |
| if (state_ != STATE_PLAYING) |
| return; |
| |
| if (type == DemuxerStream::AUDIO) { |
| DCHECK(!audio_ended_); |
| audio_ended_ = true; |
| } else { |
| DCHECK(!video_ended_); |
| video_ended_ = true; |
| DCHECK(video_renderer_); |
| video_renderer_->OnTimeStopped(); |
| } |
| |
| RunEndedCallbackIfNeeded(); |
| } |
| |
| bool RendererImpl::PlaybackHasEnded() const { |
| DCHECK(task_runner_->BelongsToCurrentThread()); |
| |
| if (audio_renderer_ && !audio_ended_) |
| return false; |
| |
| if (video_renderer_ && !video_ended_) |
| return false; |
| |
| return true; |
| } |
| |
| void RendererImpl::RunEndedCallbackIfNeeded() { |
| DVLOG(1) << __func__; |
| DCHECK(task_runner_->BelongsToCurrentThread()); |
| |
| if (!PlaybackHasEnded()) |
| return; |
| |
| if (time_ticking_) |
| PausePlayback(); |
| |
| client_->OnEnded(); |
| } |
| |
| void RendererImpl::OnError(PipelineStatus error) { |
| DVLOG(1) << __func__ << "(" << error << ")"; |
| DCHECK(task_runner_->BelongsToCurrentThread()); |
| DCHECK_NE(PIPELINE_OK, error) << "PIPELINE_OK isn't an error!"; |
| |
| // An error has already been delivered. |
| if (state_ == STATE_ERROR) |
| return; |
| |
| const State old_state = state_; |
| state_ = STATE_ERROR; |
| |
| if (!init_cb_.is_null()) { |
| DCHECK(old_state == STATE_INITIALIZING || |
| old_state == STATE_INIT_PENDING_CDM); |
| FinishInitialization(error); |
| return; |
| } |
| |
| // After OnError() returns, the pipeline may destroy |this|. |
| client_->OnError(error); |
| |
| if (!flush_cb_.is_null()) |
| base::ResetAndReturn(&flush_cb_).Run(); |
| } |
| |
| void RendererImpl::OnWaitingForDecryptionKey() { |
| DCHECK(task_runner_->BelongsToCurrentThread()); |
| client_->OnWaitingForDecryptionKey(); |
| } |
| |
| void RendererImpl::OnVideoNaturalSizeChange(const gfx::Size& size) { |
| DCHECK(task_runner_->BelongsToCurrentThread()); |
| client_->OnVideoNaturalSizeChange(size); |
| } |
| |
| void RendererImpl::OnVideoOpacityChange(bool opaque) { |
| DCHECK(task_runner_->BelongsToCurrentThread()); |
| client_->OnVideoOpacityChange(opaque); |
| } |
| |
| } // namespace media |