blob: 0be8b4983d385c1a12d23ebfaf1f9e1b725eef38 [file] [log] [blame]
// Copyright 2014 The Chromium Authors. All rights reserved.
// Use of this source code is governed by a BSD-style license that can be
// found in the LICENSE file.
#include "remoting/client/software_video_renderer.h"
#include <list>
#include "base/bind.h"
#include "base/callback.h"
#include "base/callback_helpers.h"
#include "base/location.h"
#include "base/logging.h"
#include "base/single_thread_task_runner.h"
#include "remoting/base/util.h"
#include "remoting/client/frame_consumer.h"
#include "remoting/codec/video_decoder.h"
#include "remoting/codec/video_decoder_verbatim.h"
#include "remoting/codec/video_decoder_vpx.h"
#include "remoting/protocol/session_config.h"
#include "third_party/libyuv/include/libyuv/convert_argb.h"
#include "third_party/webrtc/modules/desktop_capture/desktop_frame.h"
using base::Passed;
using remoting::protocol::ChannelConfig;
using remoting::protocol::SessionConfig;
namespace remoting {
// This class wraps a VideoDecoder and byte-swaps the pixels for compatibility
// with the android.graphics.Bitmap class.
// TODO(lambroslambrou): Refactor so that the VideoDecoder produces data
// in the right byte-order, instead of swapping it here.
class RgbToBgrVideoDecoderFilter : public VideoDecoder {
public:
RgbToBgrVideoDecoderFilter(scoped_ptr<VideoDecoder> parent)
: parent_(parent.Pass()) {
}
void Initialize(const webrtc::DesktopSize& screen_size) override {
parent_->Initialize(screen_size);
}
bool DecodePacket(const VideoPacket& packet) override {
return parent_->DecodePacket(packet);
}
void Invalidate(const webrtc::DesktopSize& view_size,
const webrtc::DesktopRegion& region) override {
return parent_->Invalidate(view_size, region);
}
void RenderFrame(const webrtc::DesktopSize& view_size,
const webrtc::DesktopRect& clip_area,
uint8* image_buffer,
int image_stride,
webrtc::DesktopRegion* output_region) override {
parent_->RenderFrame(view_size, clip_area, image_buffer, image_stride,
output_region);
for (webrtc::DesktopRegion::Iterator i(*output_region); !i.IsAtEnd();
i.Advance()) {
webrtc::DesktopRect rect = i.rect();
uint8* pixels = image_buffer + (rect.top() * image_stride) +
(rect.left() * kBytesPerPixel);
libyuv::ABGRToARGB(pixels, image_stride, pixels, image_stride,
rect.width(), rect.height());
}
}
const webrtc::DesktopRegion* GetImageShape() override {
return parent_->GetImageShape();
}
private:
scoped_ptr<VideoDecoder> parent_;
};
class SoftwareVideoRenderer::Core {
public:
Core(scoped_refptr<base::SingleThreadTaskRunner> main_task_runner,
scoped_refptr<base::SingleThreadTaskRunner> decode_task_runner,
scoped_refptr<FrameConsumerProxy> consumer);
~Core();
void OnSessionConfig(const protocol::SessionConfig& config);
void DrawBuffer(webrtc::DesktopFrame* buffer);
void InvalidateRegion(const webrtc::DesktopRegion& region);
void RequestReturnBuffers(const base::Closure& done);
void SetOutputSizeAndClip(
const webrtc::DesktopSize& view_size,
const webrtc::DesktopRect& clip_area);
// Decodes the contents of |packet|. DecodePacket may keep a reference to
// |packet| so the |packet| must remain alive and valid until |done| is
// executed.
void DecodePacket(scoped_ptr<VideoPacket> packet, const base::Closure& done);
private:
// Paints the invalidated region to the next available buffer and returns it
// to the consumer.
void SchedulePaint();
void DoPaint();
scoped_refptr<base::SingleThreadTaskRunner> main_task_runner_;
scoped_refptr<base::SingleThreadTaskRunner> decode_task_runner_;
scoped_refptr<FrameConsumerProxy> consumer_;
scoped_ptr<VideoDecoder> decoder_;
// Remote screen size in pixels.
webrtc::DesktopSize source_size_;
// Vertical and horizontal DPI of the remote screen.
webrtc::DesktopVector source_dpi_;
// The current dimensions of the frame consumer view.
webrtc::DesktopSize view_size_;
webrtc::DesktopRect clip_area_;
// The drawing buffers supplied by the frame consumer.
std::list<webrtc::DesktopFrame*> buffers_;
// Flag used to coalesce runs of SchedulePaint()s into a single DoPaint().
bool paint_scheduled_;
base::WeakPtrFactory<Core> weak_factory_;
};
SoftwareVideoRenderer::Core::Core(
scoped_refptr<base::SingleThreadTaskRunner> main_task_runner,
scoped_refptr<base::SingleThreadTaskRunner> decode_task_runner,
scoped_refptr<FrameConsumerProxy> consumer)
: main_task_runner_(main_task_runner),
decode_task_runner_(decode_task_runner),
consumer_(consumer),
paint_scheduled_(false),
weak_factory_(this) {
}
SoftwareVideoRenderer::Core::~Core() {
}
void SoftwareVideoRenderer::Core::OnSessionConfig(const SessionConfig& config) {
DCHECK(decode_task_runner_->BelongsToCurrentThread());
// Initialize decoder based on the selected codec.
ChannelConfig::Codec codec = config.video_config().codec;
if (codec == ChannelConfig::CODEC_VERBATIM) {
decoder_.reset(new VideoDecoderVerbatim());
} else if (codec == ChannelConfig::CODEC_VP8) {
decoder_ = VideoDecoderVpx::CreateForVP8();
} else if (codec == ChannelConfig::CODEC_VP9) {
decoder_ = VideoDecoderVpx::CreateForVP9();
} else {
NOTREACHED() << "Invalid Encoding found: " << codec;
}
if (consumer_->GetPixelFormat() == FrameConsumer::FORMAT_RGBA) {
scoped_ptr<VideoDecoder> wrapper(
new RgbToBgrVideoDecoderFilter(decoder_.Pass()));
decoder_ = wrapper.Pass();
}
}
void SoftwareVideoRenderer::Core::DecodePacket(scoped_ptr<VideoPacket> packet,
const base::Closure& done) {
DCHECK(decode_task_runner_->BelongsToCurrentThread());
bool decoder_needs_reset = false;
bool notify_size_or_dpi_change = false;
// If the packet includes screen size or DPI information, store them.
if (packet->format().has_screen_width() &&
packet->format().has_screen_height()) {
webrtc::DesktopSize source_size(packet->format().screen_width(),
packet->format().screen_height());
if (!source_size_.equals(source_size)) {
source_size_ = source_size;
decoder_needs_reset = true;
notify_size_or_dpi_change = true;
}
}
if (packet->format().has_x_dpi() && packet->format().has_y_dpi()) {
webrtc::DesktopVector source_dpi(packet->format().x_dpi(),
packet->format().y_dpi());
if (!source_dpi.equals(source_dpi_)) {
source_dpi_ = source_dpi;
notify_size_or_dpi_change = true;
}
}
// If we've never seen a screen size, ignore the packet.
if (source_size_.is_empty()) {
main_task_runner_->PostTask(FROM_HERE, base::Bind(done));
return;
}
if (decoder_needs_reset)
decoder_->Initialize(source_size_);
if (notify_size_or_dpi_change)
consumer_->SetSourceSize(source_size_, source_dpi_);
if (decoder_->DecodePacket(*packet.get())) {
SchedulePaint();
} else {
LOG(ERROR) << "DecodePacket() failed.";
}
main_task_runner_->PostTask(FROM_HERE, base::Bind(done));
}
void SoftwareVideoRenderer::Core::SchedulePaint() {
DCHECK(decode_task_runner_->BelongsToCurrentThread());
if (paint_scheduled_)
return;
paint_scheduled_ = true;
decode_task_runner_->PostTask(
FROM_HERE, base::Bind(&SoftwareVideoRenderer::Core::DoPaint,
weak_factory_.GetWeakPtr()));
}
void SoftwareVideoRenderer::Core::DoPaint() {
DCHECK(decode_task_runner_->BelongsToCurrentThread());
DCHECK(paint_scheduled_);
paint_scheduled_ = false;
// If the view size is empty or we have no output buffers ready, return.
if (buffers_.empty() || view_size_.is_empty())
return;
// If no Decoder is initialized, or the host dimensions are empty, return.
if (!decoder_.get() || source_size_.is_empty())
return;
// Draw the invalidated region to the buffer.
webrtc::DesktopFrame* buffer = buffers_.front();
webrtc::DesktopRegion output_region;
decoder_->RenderFrame(view_size_, clip_area_,
buffer->data(), buffer->stride(), &output_region);
// Notify the consumer that painting is done.
if (!output_region.is_empty()) {
buffers_.pop_front();
consumer_->ApplyBuffer(view_size_, clip_area_, buffer, output_region,
*decoder_->GetImageShape());
}
}
void SoftwareVideoRenderer::Core::RequestReturnBuffers(
const base::Closure& done) {
DCHECK(decode_task_runner_->BelongsToCurrentThread());
while (!buffers_.empty()) {
consumer_->ReturnBuffer(buffers_.front());
buffers_.pop_front();
}
if (!done.is_null())
done.Run();
}
void SoftwareVideoRenderer::Core::DrawBuffer(webrtc::DesktopFrame* buffer) {
DCHECK(decode_task_runner_->BelongsToCurrentThread());
DCHECK(clip_area_.width() <= buffer->size().width() &&
clip_area_.height() <= buffer->size().height());
buffers_.push_back(buffer);
SchedulePaint();
}
void SoftwareVideoRenderer::Core::InvalidateRegion(
const webrtc::DesktopRegion& region) {
DCHECK(decode_task_runner_->BelongsToCurrentThread());
if (decoder_.get()) {
decoder_->Invalidate(view_size_, region);
SchedulePaint();
}
}
void SoftwareVideoRenderer::Core::SetOutputSizeAndClip(
const webrtc::DesktopSize& view_size,
const webrtc::DesktopRect& clip_area) {
DCHECK(decode_task_runner_->BelongsToCurrentThread());
// The whole frame needs to be repainted if the scaling factor has changed.
if (!view_size_.equals(view_size) && decoder_.get()) {
webrtc::DesktopRegion region;
region.AddRect(webrtc::DesktopRect::MakeSize(view_size));
decoder_->Invalidate(view_size, region);
}
if (!view_size_.equals(view_size) ||
!clip_area_.equals(clip_area)) {
view_size_ = view_size;
clip_area_ = clip_area;
// Return buffers that are smaller than needed to the consumer for
// reuse/reallocation.
std::list<webrtc::DesktopFrame*>::iterator i = buffers_.begin();
while (i != buffers_.end()) {
if ((*i)->size().width() < clip_area_.width() ||
(*i)->size().height() < clip_area_.height()) {
consumer_->ReturnBuffer(*i);
i = buffers_.erase(i);
} else {
++i;
}
}
SchedulePaint();
}
}
SoftwareVideoRenderer::SoftwareVideoRenderer(
scoped_refptr<base::SingleThreadTaskRunner> main_task_runner,
scoped_refptr<base::SingleThreadTaskRunner> decode_task_runner,
scoped_refptr<FrameConsumerProxy> consumer)
: decode_task_runner_(decode_task_runner),
core_(new Core(main_task_runner, decode_task_runner, consumer)),
latest_event_timestamp_(0),
weak_factory_(this) {
DCHECK(CalledOnValidThread());
}
SoftwareVideoRenderer::~SoftwareVideoRenderer() {
DCHECK(CalledOnValidThread());
decode_task_runner_->DeleteSoon(FROM_HERE, core_.release());
}
void SoftwareVideoRenderer::OnSessionConfig(
const protocol::SessionConfig& config) {
DCHECK(CalledOnValidThread());
decode_task_runner_->PostTask(
FROM_HERE, base::Bind(&SoftwareVideoRenderer::Core::OnSessionConfig,
base::Unretained(core_.get()), config));
}
ChromotingStats* SoftwareVideoRenderer::GetStats() {
DCHECK(CalledOnValidThread());
return &stats_;
}
protocol::VideoStub* SoftwareVideoRenderer::GetVideoStub() {
return this;
}
void SoftwareVideoRenderer::ProcessVideoPacket(scoped_ptr<VideoPacket> packet,
const base::Closure& done) {
DCHECK(CalledOnValidThread());
// If the video packet is empty then drop it. Empty packets are used to
// maintain activity on the network.
if (!packet->has_data() || packet->data().size() == 0) {
done.Run();
return;
}
// Add one frame to the counter.
stats_.video_frame_rate()->Record(1);
// Record other statistics received from host.
stats_.video_bandwidth()->Record(packet->data().size());
if (packet->has_capture_time_ms())
stats_.video_capture_ms()->Record(packet->capture_time_ms());
if (packet->has_encode_time_ms())
stats_.video_encode_ms()->Record(packet->encode_time_ms());
if (packet->has_latest_event_timestamp() &&
packet->latest_event_timestamp() > latest_event_timestamp_) {
latest_event_timestamp_ = packet->latest_event_timestamp();
base::TimeDelta round_trip_latency =
base::Time::Now() -
base::Time::FromInternalValue(packet->latest_event_timestamp());
stats_.round_trip_ms()->Record(round_trip_latency.InMilliseconds());
}
// Measure the latency between the last packet being received and presented.
base::Time decode_start = base::Time::Now();
base::Closure decode_done = base::Bind(&SoftwareVideoRenderer::OnPacketDone,
weak_factory_.GetWeakPtr(),
decode_start, done);
decode_task_runner_->PostTask(FROM_HERE, base::Bind(
&SoftwareVideoRenderer::Core::DecodePacket,
base::Unretained(core_.get()), base::Passed(&packet), decode_done));
}
void SoftwareVideoRenderer::DrawBuffer(webrtc::DesktopFrame* buffer) {
decode_task_runner_->PostTask(
FROM_HERE, base::Bind(&SoftwareVideoRenderer::Core::DrawBuffer,
base::Unretained(core_.get()), buffer));
}
void SoftwareVideoRenderer::InvalidateRegion(
const webrtc::DesktopRegion& region) {
decode_task_runner_->PostTask(
FROM_HERE, base::Bind(&SoftwareVideoRenderer::Core::InvalidateRegion,
base::Unretained(core_.get()), region));
}
void SoftwareVideoRenderer::RequestReturnBuffers(const base::Closure& done) {
decode_task_runner_->PostTask(
FROM_HERE,
base::Bind(&SoftwareVideoRenderer::Core::RequestReturnBuffers,
base::Unretained(core_.get()), done));
}
void SoftwareVideoRenderer::SetOutputSizeAndClip(
const webrtc::DesktopSize& view_size,
const webrtc::DesktopRect& clip_area) {
decode_task_runner_->PostTask(
FROM_HERE,
base::Bind(&SoftwareVideoRenderer::Core::SetOutputSizeAndClip,
base::Unretained(core_.get()), view_size, clip_area));
}
void SoftwareVideoRenderer::OnPacketDone(base::Time decode_start,
const base::Closure& done) {
DCHECK(CalledOnValidThread());
// Record the latency between the packet being received and presented.
stats_.video_decode_ms()->Record(
(base::Time::Now() - decode_start).InMilliseconds());
done.Run();
}
} // namespace remoting