blob: 771eedc557a019fe0c853c687e17696977c0e1d0 [file] [log] [blame]
// Copyright (c) 2012 The Chromium Authors. All rights reserved.
// Use of this source code is governed by a BSD-style license that can be
// found in the LICENSE file.
#include <algorithm>
#include <limits>
#include <memory>
#include "base/bind.h"
#include "base/callback_helpers.h"
#include "base/feature_list.h"
#include "base/logging.h"
#include "base/numerics/safe_conversions.h"
#include "base/stl_util.h"
#include "media/base/media_switches.h"
#include "media/gpu/h264_decoder.h"
#include "media/video/h264_level_limits.h"
#include "third_party/abseil-cpp/absl/types/optional.h"
namespace media {
namespace {
bool ParseBitDepth(const H264SPS& sps, uint8_t& bit_depth) {
// Spec 7.4.2.1.1
if (sps.bit_depth_luma_minus8 != sps.bit_depth_chroma_minus8) {
DVLOG(1) << "H264Decoder doesn't support different bit depths between luma"
<< "and chroma, bit_depth_luma_minus8="
<< sps.bit_depth_luma_minus8
<< ", bit_depth_chroma_minus8=" << sps.bit_depth_chroma_minus8;
return false;
}
DCHECK_GE(sps.bit_depth_luma_minus8, 0);
DCHECK_LE(sps.bit_depth_luma_minus8, 6);
switch (sps.bit_depth_luma_minus8) {
case 0:
bit_depth = 8u;
break;
case 2:
bit_depth = 10u;
break;
case 4:
bit_depth = 12u;
break;
case 6:
bit_depth = 14u;
break;
default:
DVLOG(1) << "Invalid bit depth: "
<< base::checked_cast<int>(sps.bit_depth_luma_minus8 + 8);
return false;
}
return true;
}
bool IsValidBitDepth(uint8_t bit_depth, VideoCodecProfile profile) {
// Spec A.2.
switch (profile) {
case H264PROFILE_BASELINE:
case H264PROFILE_MAIN:
case H264PROFILE_EXTENDED:
case H264PROFILE_HIGH:
return bit_depth == 8u;
case H264PROFILE_HIGH10PROFILE:
case H264PROFILE_HIGH422PROFILE:
return bit_depth == 8u || bit_depth == 10u;
case H264PROFILE_HIGH444PREDICTIVEPROFILE:
return bit_depth == 8u || bit_depth == 10u || bit_depth == 12u ||
bit_depth == 14u;
case H264PROFILE_SCALABLEBASELINE:
case H264PROFILE_SCALABLEHIGH:
// Spec G.10.1.
return bit_depth == 8u;
case H264PROFILE_STEREOHIGH:
case H264PROFILE_MULTIVIEWHIGH:
// Spec H.10.1.1 and H.10.1.2.
return bit_depth == 8u;
default:
NOTREACHED();
return false;
}
}
bool IsYUV420Sequence(const H264SPS& sps) {
// Spec 6.2
return sps.chroma_format_idc == 1;
}
} // namespace
H264Decoder::H264Accelerator::H264Accelerator() = default;
H264Decoder::H264Accelerator::~H264Accelerator() = default;
H264Decoder::H264Accelerator::Status H264Decoder::H264Accelerator::SetStream(
base::span<const uint8_t> stream,
const DecryptConfig* decrypt_config) {
return H264Decoder::H264Accelerator::Status::kNotSupported;
}
H264Decoder::H264Accelerator::Status
H264Decoder::H264Accelerator::ParseEncryptedSliceHeader(
const std::vector<base::span<const uint8_t>>& data,
const std::vector<SubsampleEntry>& subsamples,
const std::vector<uint8_t>& sps_nalu_data,
const std::vector<uint8_t>& pps_nalu_data,
H264SliceHeader* slice_header_out) {
return H264Decoder::H264Accelerator::Status::kNotSupported;
}
H264Decoder::H264Decoder(std::unique_ptr<H264Accelerator> accelerator,
VideoCodecProfile profile,
const VideoColorSpace& container_color_space)
: state_(kNeedStreamMetadata),
container_color_space_(container_color_space),
max_frame_num_(0),
max_pic_num_(0),
max_long_term_frame_idx_(0),
max_num_reorder_frames_(0),
// TODO(hiroh): Set profile to UNKNOWN.
profile_(profile),
accelerator_(std::move(accelerator)) {
DCHECK(accelerator_);
Reset();
}
H264Decoder::~H264Decoder() = default;
void H264Decoder::Reset() {
curr_pic_ = nullptr;
curr_nalu_ = nullptr;
curr_slice_hdr_ = nullptr;
curr_sps_id_ = -1;
curr_pps_id_ = -1;
prev_frame_num_ = -1;
prev_ref_frame_num_ = -1;
prev_frame_num_offset_ = -1;
prev_has_memmgmnt5_ = false;
prev_ref_has_memmgmnt5_ = false;
prev_ref_top_field_order_cnt_ = -1;
prev_ref_pic_order_cnt_msb_ = -1;
prev_ref_pic_order_cnt_lsb_ = -1;
prev_ref_field_ = H264Picture::FIELD_NONE;
ref_pic_list_p0_.clear();
ref_pic_list_b0_.clear();
ref_pic_list_b1_.clear();
dpb_.Clear();
parser_.Reset();
accelerator_->Reset();
last_output_poc_ = std::numeric_limits<int>::min();
encrypted_sei_nalus_.clear();
sei_subsamples_.clear();
recovery_frame_num_.reset();
recovery_frame_cnt_.reset();
// If we are in kDecoding, we can resume without processing an SPS.
// The state becomes kDecoding again, (1) at the first IDR slice or (2) at
// the first slice after the recovery point SEI.
if (state_ == kDecoding)
state_ = kAfterReset;
}
void H264Decoder::PrepareRefPicLists() {
ConstructReferencePicListsP();
ConstructReferencePicListsB();
}
bool H264Decoder::ModifyReferencePicLists(const H264SliceHeader* slice_hdr,
H264Picture::Vector* ref_pic_list0,
H264Picture::Vector* ref_pic_list1) {
ref_pic_list0->clear();
ref_pic_list1->clear();
// Fill reference picture lists for B and S/SP slices.
if (slice_hdr->IsPSlice() || slice_hdr->IsSPSlice()) {
*ref_pic_list0 = ref_pic_list_p0_;
return ModifyReferencePicList(slice_hdr, 0, ref_pic_list0);
} else if (slice_hdr->IsBSlice()) {
*ref_pic_list0 = ref_pic_list_b0_;
*ref_pic_list1 = ref_pic_list_b1_;
return ModifyReferencePicList(slice_hdr, 0, ref_pic_list0) &&
ModifyReferencePicList(slice_hdr, 1, ref_pic_list1);
}
return true;
}
H264Decoder::H264Accelerator::Status H264Decoder::DecodePicture() {
DCHECK(curr_pic_.get());
return accelerator_->SubmitDecode(curr_pic_);
}
bool H264Decoder::InitNonexistingPicture(scoped_refptr<H264Picture> pic,
int frame_num) {
pic->nonexisting = true;
pic->nal_ref_idc = 1;
pic->frame_num = pic->pic_num = frame_num;
pic->adaptive_ref_pic_marking_mode_flag = false;
pic->ref = true;
pic->long_term_reference_flag = false;
pic->field = H264Picture::FIELD_NONE;
return CalculatePicOrderCounts(pic);
}
bool H264Decoder::InitCurrPicture(const H264SliceHeader* slice_hdr) {
if (!FillH264PictureFromSliceHeader(parser_.GetSPS(curr_sps_id_), *slice_hdr,
curr_pic_.get())) {
return false;
}
if (!CalculatePicOrderCounts(curr_pic_))
return false;
curr_pic_->long_term_reference_flag = slice_hdr->long_term_reference_flag;
curr_pic_->adaptive_ref_pic_marking_mode_flag =
slice_hdr->adaptive_ref_pic_marking_mode_flag;
// If the slice header indicates we will have to perform reference marking
// process after this picture is decoded, store required data for that
// purpose.
if (slice_hdr->adaptive_ref_pic_marking_mode_flag) {
static_assert(sizeof(curr_pic_->ref_pic_marking) ==
sizeof(slice_hdr->ref_pic_marking),
"Array sizes of ref pic marking do not match.");
memcpy(curr_pic_->ref_pic_marking, slice_hdr->ref_pic_marking,
sizeof(curr_pic_->ref_pic_marking));
}
curr_pic_->set_visible_rect(visible_rect_);
curr_pic_->set_bitstream_id(stream_id_);
return true;
}
bool H264Decoder::CalculatePicOrderCounts(scoped_refptr<H264Picture> pic) {
const H264SPS* sps = parser_.GetSPS(curr_sps_id_);
if (!sps)
return false;
switch (pic->pic_order_cnt_type) {
case 0: {
// See spec 8.2.1.1.
int prev_pic_order_cnt_msb, prev_pic_order_cnt_lsb;
if (pic->idr) {
prev_pic_order_cnt_msb = prev_pic_order_cnt_lsb = 0;
} else {
if (prev_ref_has_memmgmnt5_) {
if (prev_ref_field_ != H264Picture::FIELD_BOTTOM) {
prev_pic_order_cnt_msb = 0;
prev_pic_order_cnt_lsb = prev_ref_top_field_order_cnt_;
} else {
prev_pic_order_cnt_msb = 0;
prev_pic_order_cnt_lsb = 0;
}
} else {
prev_pic_order_cnt_msb = prev_ref_pic_order_cnt_msb_;
prev_pic_order_cnt_lsb = prev_ref_pic_order_cnt_lsb_;
}
}
int max_pic_order_cnt_lsb =
1 << (sps->log2_max_pic_order_cnt_lsb_minus4 + 4);
DCHECK_NE(max_pic_order_cnt_lsb, 0);
if ((pic->pic_order_cnt_lsb < prev_pic_order_cnt_lsb) &&
(prev_pic_order_cnt_lsb - pic->pic_order_cnt_lsb >=
max_pic_order_cnt_lsb / 2)) {
pic->pic_order_cnt_msb = prev_pic_order_cnt_msb + max_pic_order_cnt_lsb;
} else if ((pic->pic_order_cnt_lsb > prev_pic_order_cnt_lsb) &&
(pic->pic_order_cnt_lsb - prev_pic_order_cnt_lsb >
max_pic_order_cnt_lsb / 2)) {
pic->pic_order_cnt_msb = prev_pic_order_cnt_msb - max_pic_order_cnt_lsb;
} else {
pic->pic_order_cnt_msb = prev_pic_order_cnt_msb;
}
if (pic->field != H264Picture::FIELD_BOTTOM) {
pic->top_field_order_cnt =
pic->pic_order_cnt_msb + pic->pic_order_cnt_lsb;
}
if (pic->field != H264Picture::FIELD_TOP) {
if (pic->field == H264Picture::FIELD_NONE) {
pic->bottom_field_order_cnt =
pic->top_field_order_cnt + pic->delta_pic_order_cnt_bottom;
} else {
pic->bottom_field_order_cnt =
pic->pic_order_cnt_msb + pic->pic_order_cnt_lsb;
}
}
break;
}
case 1: {
// See spec 8.2.1.2.
if (prev_has_memmgmnt5_)
prev_frame_num_offset_ = 0;
if (pic->idr)
pic->frame_num_offset = 0;
else if (prev_frame_num_ > pic->frame_num)
pic->frame_num_offset = prev_frame_num_offset_ + max_frame_num_;
else
pic->frame_num_offset = prev_frame_num_offset_;
int abs_frame_num = 0;
if (sps->num_ref_frames_in_pic_order_cnt_cycle != 0)
abs_frame_num = pic->frame_num_offset + pic->frame_num;
else
abs_frame_num = 0;
if (pic->nal_ref_idc == 0 && abs_frame_num > 0)
--abs_frame_num;
int expected_pic_order_cnt = 0;
if (abs_frame_num > 0) {
if (sps->num_ref_frames_in_pic_order_cnt_cycle == 0) {
DVLOG(1) << "Invalid num_ref_frames_in_pic_order_cnt_cycle "
<< "in stream";
return false;
}
int pic_order_cnt_cycle_cnt =
(abs_frame_num - 1) / sps->num_ref_frames_in_pic_order_cnt_cycle;
int frame_num_in_pic_order_cnt_cycle =
(abs_frame_num - 1) % sps->num_ref_frames_in_pic_order_cnt_cycle;
expected_pic_order_cnt = pic_order_cnt_cycle_cnt *
sps->expected_delta_per_pic_order_cnt_cycle;
// frame_num_in_pic_order_cnt_cycle is verified < 255 in parser
for (int i = 0; i <= frame_num_in_pic_order_cnt_cycle; ++i)
expected_pic_order_cnt += sps->offset_for_ref_frame[i];
}
if (!pic->nal_ref_idc)
expected_pic_order_cnt += sps->offset_for_non_ref_pic;
if (pic->field == H264Picture::FIELD_NONE) {
pic->top_field_order_cnt =
expected_pic_order_cnt + pic->delta_pic_order_cnt0;
pic->bottom_field_order_cnt = pic->top_field_order_cnt +
sps->offset_for_top_to_bottom_field +
pic->delta_pic_order_cnt1;
} else if (pic->field != H264Picture::FIELD_BOTTOM) {
pic->top_field_order_cnt =
expected_pic_order_cnt + pic->delta_pic_order_cnt0;
} else {
pic->bottom_field_order_cnt = expected_pic_order_cnt +
sps->offset_for_top_to_bottom_field +
pic->delta_pic_order_cnt0;
}
break;
}
case 2: {
// See spec 8.2.1.3.
if (prev_has_memmgmnt5_)
prev_frame_num_offset_ = 0;
if (pic->idr)
pic->frame_num_offset = 0;
else if (prev_frame_num_ > pic->frame_num)
pic->frame_num_offset = prev_frame_num_offset_ + max_frame_num_;
else
pic->frame_num_offset = prev_frame_num_offset_;
int temp_pic_order_cnt;
if (pic->idr) {
temp_pic_order_cnt = 0;
} else if (!pic->nal_ref_idc) {
temp_pic_order_cnt = 2 * (pic->frame_num_offset + pic->frame_num) - 1;
} else {
temp_pic_order_cnt = 2 * (pic->frame_num_offset + pic->frame_num);
}
if (pic->field == H264Picture::FIELD_NONE) {
pic->top_field_order_cnt = temp_pic_order_cnt;
pic->bottom_field_order_cnt = temp_pic_order_cnt;
} else if (pic->field == H264Picture::FIELD_BOTTOM) {
pic->bottom_field_order_cnt = temp_pic_order_cnt;
} else {
pic->top_field_order_cnt = temp_pic_order_cnt;
}
break;
}
default:
DVLOG(1) << "Invalid pic_order_cnt_type: " << sps->pic_order_cnt_type;
return false;
}
switch (pic->field) {
case H264Picture::FIELD_NONE:
pic->pic_order_cnt =
std::min(pic->top_field_order_cnt, pic->bottom_field_order_cnt);
break;
case H264Picture::FIELD_TOP:
pic->pic_order_cnt = pic->top_field_order_cnt;
break;
case H264Picture::FIELD_BOTTOM:
pic->pic_order_cnt = pic->bottom_field_order_cnt;
break;
}
return true;
}
void H264Decoder::UpdatePicNums(int frame_num) {
for (auto& pic : dpb_) {
if (!pic->ref)
continue;
// 8.2.4.1. Assumes non-interlaced stream.
DCHECK_EQ(pic->field, H264Picture::FIELD_NONE);
if (pic->long_term) {
pic->long_term_pic_num = pic->long_term_frame_idx;
} else {
if (pic->frame_num > frame_num)
pic->frame_num_wrap = pic->frame_num - max_frame_num_;
else
pic->frame_num_wrap = pic->frame_num;
pic->pic_num = pic->frame_num_wrap;
}
}
}
struct PicNumDescCompare {
bool operator()(const scoped_refptr<H264Picture>& a,
const scoped_refptr<H264Picture>& b) const {
return a->pic_num > b->pic_num;
}
};
struct LongTermPicNumAscCompare {
bool operator()(const scoped_refptr<H264Picture>& a,
const scoped_refptr<H264Picture>& b) const {
return a->long_term_pic_num < b->long_term_pic_num;
}
};
void H264Decoder::ConstructReferencePicListsP() {
// RefPicList0 (8.2.4.2.1) [[1] [2]], where:
// [1] shortterm ref pics sorted by descending pic_num,
// [2] longterm ref pics by ascending long_term_pic_num.
ref_pic_list_p0_.clear();
// First get the short ref pics...
dpb_.GetShortTermRefPicsAppending(&ref_pic_list_p0_);
size_t num_short_refs = ref_pic_list_p0_.size();
// and sort them to get [1].
std::sort(ref_pic_list_p0_.begin(), ref_pic_list_p0_.end(),
PicNumDescCompare());
// Now get long term pics and sort them by long_term_pic_num to get [2].
dpb_.GetLongTermRefPicsAppending(&ref_pic_list_p0_);
std::sort(ref_pic_list_p0_.begin() + num_short_refs, ref_pic_list_p0_.end(),
LongTermPicNumAscCompare());
}
struct POCAscCompare {
bool operator()(const scoped_refptr<H264Picture>& a,
const scoped_refptr<H264Picture>& b) const {
return a->pic_order_cnt < b->pic_order_cnt;
}
};
struct POCDescCompare {
bool operator()(const scoped_refptr<H264Picture>& a,
const scoped_refptr<H264Picture>& b) const {
return a->pic_order_cnt > b->pic_order_cnt;
}
};
void H264Decoder::ConstructReferencePicListsB() {
// RefPicList0 (8.2.4.2.3) [[1] [2] [3]], where:
// [1] shortterm ref pics with POC < curr_pic's POC sorted by descending POC,
// [2] shortterm ref pics with POC > curr_pic's POC by ascending POC,
// [3] longterm ref pics by ascending long_term_pic_num.
ref_pic_list_b0_.clear();
ref_pic_list_b1_.clear();
dpb_.GetShortTermRefPicsAppending(&ref_pic_list_b0_);
size_t num_short_refs = ref_pic_list_b0_.size();
// First sort ascending, this will put [1] in right place and finish [2].
std::sort(ref_pic_list_b0_.begin(), ref_pic_list_b0_.end(), POCAscCompare());
// Find first with POC > curr_pic's POC to get first element in [2]...
H264Picture::Vector::iterator iter;
iter = std::upper_bound(ref_pic_list_b0_.begin(), ref_pic_list_b0_.end(),
curr_pic_.get(), POCAscCompare());
// and sort [1] descending, thus finishing sequence [1] [2].
std::sort(ref_pic_list_b0_.begin(), iter, POCDescCompare());
// Now add [3] and sort by ascending long_term_pic_num.
dpb_.GetLongTermRefPicsAppending(&ref_pic_list_b0_);
std::sort(ref_pic_list_b0_.begin() + num_short_refs, ref_pic_list_b0_.end(),
LongTermPicNumAscCompare());
// RefPicList1 (8.2.4.2.4) [[1] [2] [3]], where:
// [1] shortterm ref pics with POC > curr_pic's POC sorted by ascending POC,
// [2] shortterm ref pics with POC < curr_pic's POC by descending POC,
// [3] longterm ref pics by ascending long_term_pic_num.
dpb_.GetShortTermRefPicsAppending(&ref_pic_list_b1_);
num_short_refs = ref_pic_list_b1_.size();
// First sort by descending POC.
std::sort(ref_pic_list_b1_.begin(), ref_pic_list_b1_.end(), POCDescCompare());
// Find first with POC < curr_pic's POC to get first element in [2]...
iter = std::upper_bound(ref_pic_list_b1_.begin(), ref_pic_list_b1_.end(),
curr_pic_.get(), POCDescCompare());
// and sort [1] ascending.
std::sort(ref_pic_list_b1_.begin(), iter, POCAscCompare());
// Now add [3] and sort by ascending long_term_pic_num
dpb_.GetLongTermRefPicsAppending(&ref_pic_list_b1_);
std::sort(ref_pic_list_b1_.begin() + num_short_refs, ref_pic_list_b1_.end(),
LongTermPicNumAscCompare());
// If lists identical, swap first two entries in RefPicList1 (spec 8.2.4.2.3)
if (ref_pic_list_b1_.size() > 1 &&
std::equal(ref_pic_list_b0_.begin(), ref_pic_list_b0_.end(),
ref_pic_list_b1_.begin()))
std::swap(ref_pic_list_b1_[0], ref_pic_list_b1_[1]);
}
// See 8.2.4
int H264Decoder::PicNumF(const H264Picture& pic) {
if (!pic.long_term)
return pic.pic_num;
else
return max_pic_num_;
}
// See 8.2.4
int H264Decoder::LongTermPicNumF(const H264Picture& pic) {
if (pic.ref && pic.long_term)
return pic.long_term_pic_num;
else
return 2 * (max_long_term_frame_idx_ + 1);
}
// Shift elements on the |v| starting from |from| to |to|, inclusive,
// one position to the right and insert pic at |from|.
static void ShiftRightAndInsert(H264Picture::Vector* v,
int from,
int to,
scoped_refptr<H264Picture> pic) {
// Security checks, do not disable in Debug mode.
CHECK(from <= to);
CHECK(to <= std::numeric_limits<int>::max() - 2);
// Additional checks. Debug mode ok.
DCHECK(v);
DCHECK(pic);
DCHECK((to + 1 == static_cast<int>(v->size())) ||
(to + 2 == static_cast<int>(v->size())));
v->resize(to + 2);
for (int i = to + 1; i > from; --i)
(*v)[i] = (*v)[i - 1];
(*v)[from] = std::move(pic);
}
bool H264Decoder::ModifyReferencePicList(const H264SliceHeader* slice_hdr,
int list,
H264Picture::Vector* ref_pic_listx) {
bool ref_pic_list_modification_flag_lX;
int num_ref_idx_lX_active_minus1;
const H264ModificationOfPicNum* list_mod;
// This can process either ref_pic_list0 or ref_pic_list1, depending on
// the list argument. Set up pointers to proper list to be processed here.
if (list == 0) {
ref_pic_list_modification_flag_lX =
slice_hdr->ref_pic_list_modification_flag_l0;
num_ref_idx_lX_active_minus1 = slice_hdr->num_ref_idx_l0_active_minus1;
list_mod = slice_hdr->ref_list_l0_modifications;
} else {
ref_pic_list_modification_flag_lX =
slice_hdr->ref_pic_list_modification_flag_l1;
num_ref_idx_lX_active_minus1 = slice_hdr->num_ref_idx_l1_active_minus1;
list_mod = slice_hdr->ref_list_l1_modifications;
}
// Resize the list to the size requested in the slice header.
// Note that per 8.2.4.2 it's possible for num_ref_idx_lX_active_minus1 to
// indicate there should be more ref pics on list than we constructed.
// Those superfluous ones should be treated as non-reference and will be
// initialized to nullptr, which must be handled by clients.
DCHECK_GE(num_ref_idx_lX_active_minus1, 0);
ref_pic_listx->resize(num_ref_idx_lX_active_minus1 + 1);
if (!ref_pic_list_modification_flag_lX)
return true;
// Spec 8.2.4.3:
// Reorder pictures on the list in a way specified in the stream.
int pic_num_lx_pred = curr_pic_->pic_num;
int ref_idx_lx = 0;
int pic_num_lx_no_wrap;
int pic_num_lx;
bool done = false;
scoped_refptr<H264Picture> pic;
for (int i = 0; i < H264SliceHeader::kRefListModSize && !done; ++i) {
switch (list_mod->modification_of_pic_nums_idc) {
case 0:
case 1:
// Modify short reference picture position.
if (list_mod->modification_of_pic_nums_idc == 0) {
// Subtract given value from predicted PicNum.
pic_num_lx_no_wrap =
pic_num_lx_pred -
(static_cast<int>(list_mod->abs_diff_pic_num_minus1) + 1);
// Wrap around max_pic_num_ if it becomes < 0 as result
// of subtraction.
if (pic_num_lx_no_wrap < 0)
pic_num_lx_no_wrap += max_pic_num_;
} else {
// Add given value to predicted PicNum.
pic_num_lx_no_wrap =
pic_num_lx_pred +
(static_cast<int>(list_mod->abs_diff_pic_num_minus1) + 1);
// Wrap around max_pic_num_ if it becomes >= max_pic_num_ as result
// of the addition.
if (pic_num_lx_no_wrap >= max_pic_num_)
pic_num_lx_no_wrap -= max_pic_num_;
}
// For use in next iteration.
pic_num_lx_pred = pic_num_lx_no_wrap;
if (pic_num_lx_no_wrap > curr_pic_->pic_num)
pic_num_lx = pic_num_lx_no_wrap - max_pic_num_;
else
pic_num_lx = pic_num_lx_no_wrap;
DCHECK_LT(num_ref_idx_lX_active_minus1 + 1,
H264SliceHeader::kRefListModSize);
pic = dpb_.GetShortRefPicByPicNum(pic_num_lx);
if (!pic) {
DVLOG(1) << "Malformed stream, no pic num " << pic_num_lx;
return false;
}
if (ref_idx_lx > num_ref_idx_lX_active_minus1) {
DVLOG(1) << "Bounds mismatch: expected " << ref_idx_lx
<< " <= " << num_ref_idx_lX_active_minus1;
return false;
}
ShiftRightAndInsert(ref_pic_listx, ref_idx_lx,
num_ref_idx_lX_active_minus1, pic);
ref_idx_lx++;
for (int src = ref_idx_lx, dst = ref_idx_lx;
src <= num_ref_idx_lX_active_minus1 + 1; ++src) {
auto* src_pic = (*ref_pic_listx)[src].get();
int src_pic_num_lx = src_pic ? PicNumF(*src_pic) : -1;
if (src_pic_num_lx != pic_num_lx)
(*ref_pic_listx)[dst++] = (*ref_pic_listx)[src];
}
break;
case 2:
// Modify long term reference picture position.
DCHECK_LT(num_ref_idx_lX_active_minus1 + 1,
H264SliceHeader::kRefListModSize);
pic = dpb_.GetLongRefPicByLongTermPicNum(list_mod->long_term_pic_num);
if (!pic) {
DVLOG(1) << "Malformed stream, no pic num "
<< list_mod->long_term_pic_num;
return false;
}
ShiftRightAndInsert(ref_pic_listx, ref_idx_lx,
num_ref_idx_lX_active_minus1, pic);
ref_idx_lx++;
for (int src = ref_idx_lx, dst = ref_idx_lx;
src <= num_ref_idx_lX_active_minus1 + 1; ++src) {
if (LongTermPicNumF(*(*ref_pic_listx)[src]) !=
static_cast<int>(list_mod->long_term_pic_num))
(*ref_pic_listx)[dst++] = (*ref_pic_listx)[src];
}
break;
case 3:
// End of modification list.
done = true;
break;
default:
// May be recoverable.
DVLOG(1) << "Invalid modification_of_pic_nums_idc="
<< list_mod->modification_of_pic_nums_idc << " in position "
<< i;
break;
}
++list_mod;
}
// Per NOTE 2 in 8.2.4.3.2, the ref_pic_listx size in the above loop is
// temporarily made one element longer than the required final list.
// Resize the list back to its required size.
ref_pic_listx->resize(num_ref_idx_lX_active_minus1 + 1);
return true;
}
bool H264Decoder::OutputPic(scoped_refptr<H264Picture> pic) {
DCHECK(!pic->outputted);
pic->outputted = true;
VideoColorSpace colorspace_for_frame = container_color_space_;
const H264SPS* sps = parser_.GetSPS(curr_sps_id_);
if (sps && sps->GetColorSpace().IsSpecified())
colorspace_for_frame = sps->GetColorSpace();
pic->set_colorspace(colorspace_for_frame);
if (pic->nonexisting) {
DVLOG(4) << "Skipping output, non-existing frame_num: " << pic->frame_num;
return true;
}
DVLOG_IF(1, pic->pic_order_cnt < last_output_poc_)
<< "Outputting out of order, likely a broken stream: " << last_output_poc_
<< " -> " << pic->pic_order_cnt;
last_output_poc_ = pic->pic_order_cnt;
DVLOG(4) << "Posting output task for POC: " << pic->pic_order_cnt;
return accelerator_->OutputPicture(pic);
}
void H264Decoder::ClearDPB() {
// Clear DPB contents, marking the pictures as unused first.
dpb_.Clear();
last_output_poc_ = std::numeric_limits<int>::min();
}
bool H264Decoder::OutputAllRemainingPics() {
// Output all pictures that are waiting to be outputted.
if (FinishPrevFrameIfPresent() != H264Accelerator::Status::kOk)
return false;
H264Picture::Vector to_output;
dpb_.GetNotOutputtedPicsAppending(&to_output);
// Sort them by ascending POC to output in order.
std::sort(to_output.begin(), to_output.end(), POCAscCompare());
for (auto& pic : to_output) {
if (!OutputPic(pic))
return false;
}
return true;
}
bool H264Decoder::Flush() {
DVLOG(2) << "Decoder flush";
if (!OutputAllRemainingPics())
return false;
ClearDPB();
DVLOG(2) << "Decoder flush finished";
return true;
}
H264Decoder::H264Accelerator::Status H264Decoder::StartNewFrame(
const H264SliceHeader* slice_hdr) {
// TODO posciak: add handling of max_num_ref_frames per spec.
CHECK(curr_pic_.get());
DCHECK(slice_hdr);
curr_pps_id_ = slice_hdr->pic_parameter_set_id;
const H264PPS* pps = parser_.GetPPS(curr_pps_id_);
if (!pps)
return H264Accelerator::Status::kFail;
curr_sps_id_ = pps->seq_parameter_set_id;
const H264SPS* sps = parser_.GetSPS(curr_sps_id_);
if (!sps)
return H264Accelerator::Status::kFail;
max_frame_num_ = 1 << (sps->log2_max_frame_num_minus4 + 4);
int frame_num = slice_hdr->frame_num;
if (slice_hdr->idr_pic_flag)
prev_ref_frame_num_ = 0;
// 7.4.3
if (frame_num != prev_ref_frame_num_ &&
frame_num != (prev_ref_frame_num_ + 1) % max_frame_num_) {
if (!HandleFrameNumGap(frame_num))
return H264Accelerator::Status::kFail;
}
if (!InitCurrPicture(slice_hdr))
return H264Accelerator::Status::kFail;
UpdatePicNums(frame_num);
PrepareRefPicLists();
return accelerator_->SubmitFrameMetadata(sps, pps, dpb_, ref_pic_list_p0_,
ref_pic_list_b0_, ref_pic_list_b1_,
curr_pic_.get());
}
bool H264Decoder::HandleMemoryManagementOps(scoped_refptr<H264Picture> pic) {
// 8.2.5.4
for (size_t i = 0; i < base::size(pic->ref_pic_marking); ++i) {
// Code below does not support interlaced stream (per-field pictures).
H264DecRefPicMarking* ref_pic_marking = &pic->ref_pic_marking[i];
scoped_refptr<H264Picture> to_mark;
int pic_num_x;
switch (ref_pic_marking->memory_mgmnt_control_operation) {
case 0:
// Normal end of operations' specification.
return true;
case 1:
// Mark a short term reference picture as unused so it can be removed
// if outputted.
pic_num_x =
pic->pic_num - (ref_pic_marking->difference_of_pic_nums_minus1 + 1);
to_mark = dpb_.GetShortRefPicByPicNum(pic_num_x);
if (to_mark) {
to_mark->ref = false;
} else {
DVLOG(1) << "Invalid short ref pic num to unmark";
return false;
}
break;
case 2:
// Mark a long term reference picture as unused so it can be removed
// if outputted.
to_mark = dpb_.GetLongRefPicByLongTermPicNum(
ref_pic_marking->long_term_pic_num);
if (to_mark) {
to_mark->ref = false;
} else {
DVLOG(1) << "Invalid long term ref pic num to unmark";
return false;
}
break;
case 3:
// Mark a short term reference picture as long term reference.
pic_num_x =
pic->pic_num - (ref_pic_marking->difference_of_pic_nums_minus1 + 1);
to_mark = dpb_.GetShortRefPicByPicNum(pic_num_x);
if (to_mark) {
DCHECK(to_mark->ref && !to_mark->long_term);
to_mark->long_term = true;
to_mark->long_term_frame_idx = ref_pic_marking->long_term_frame_idx;
} else {
DVLOG(1) << "Invalid short term ref pic num to mark as long ref";
return false;
}
break;
case 4: {
// Unmark all reference pictures with long_term_frame_idx over new max.
max_long_term_frame_idx_ =
ref_pic_marking->max_long_term_frame_idx_plus1 - 1;
H264Picture::Vector long_terms;
dpb_.GetLongTermRefPicsAppending(&long_terms);
for (size_t i = 0; i < long_terms.size(); ++i) {
scoped_refptr<H264Picture>& long_term_pic = long_terms[i];
DCHECK(long_term_pic->ref && long_term_pic->long_term);
// Ok to cast, max_long_term_frame_idx is much smaller than 16bit.
if (long_term_pic->long_term_frame_idx >
static_cast<int>(max_long_term_frame_idx_))
long_term_pic->ref = false;
}
break;
}
case 5:
// Unmark all reference pictures.
dpb_.MarkAllUnusedForRef();
max_long_term_frame_idx_ = -1;
pic->mem_mgmt_5 = true;
break;
case 6: {
// Replace long term reference pictures with current picture.
// First unmark if any existing with this long_term_frame_idx...
H264Picture::Vector long_terms;
dpb_.GetLongTermRefPicsAppending(&long_terms);
for (size_t i = 0; i < long_terms.size(); ++i) {
scoped_refptr<H264Picture>& long_term_pic = long_terms[i];
DCHECK(long_term_pic->ref && long_term_pic->long_term);
// Ok to cast, long_term_frame_idx is much smaller than 16bit.
if (long_term_pic->long_term_frame_idx ==
static_cast<int>(ref_pic_marking->long_term_frame_idx))
long_term_pic->ref = false;
}
// and mark the current one instead.
pic->ref = true;
pic->long_term = true;
pic->long_term_frame_idx = ref_pic_marking->long_term_frame_idx;
break;
}
default:
// Would indicate a bug in parser.
NOTREACHED();
}
}
return true;
}
// This method ensures that DPB does not overflow, either by removing
// reference pictures as specified in the stream, or using a sliding window
// procedure to remove the oldest one.
// It also performs marking and unmarking pictures as reference.
// See spac 8.2.5.1.
bool H264Decoder::ReferencePictureMarking(scoped_refptr<H264Picture> pic) {
// If the current picture is an IDR, all reference pictures are unmarked.
if (pic->idr) {
dpb_.MarkAllUnusedForRef();
if (pic->long_term_reference_flag) {
pic->long_term = true;
pic->long_term_frame_idx = 0;
max_long_term_frame_idx_ = 0;
} else {
pic->long_term = false;
max_long_term_frame_idx_ = -1;
}
return true;
}
// Not an IDR. If the stream contains instructions on how to discard pictures
// from DPB and how to mark/unmark existing reference pictures, do so.
// Otherwise, fall back to default sliding window process.
if (pic->adaptive_ref_pic_marking_mode_flag) {
DCHECK(!pic->nonexisting);
return HandleMemoryManagementOps(pic);
} else {
return SlidingWindowPictureMarking();
}
}
bool H264Decoder::SlidingWindowPictureMarking() {
const H264SPS* sps = parser_.GetSPS(curr_sps_id_);
if (!sps)
return false;
// 8.2.5.3. Ensure the DPB doesn't overflow by discarding the oldest picture.
int num_ref_pics = dpb_.CountRefPics();
DCHECK_LE(num_ref_pics, std::max<int>(sps->max_num_ref_frames, 1));
if (num_ref_pics == std::max<int>(sps->max_num_ref_frames, 1)) {
// Max number of reference pics reached, need to remove one of the short
// term ones. Find smallest frame_num_wrap short reference picture and mark
// it as unused.
scoped_refptr<H264Picture> to_unmark =
dpb_.GetLowestFrameNumWrapShortRefPic();
if (!to_unmark) {
DVLOG(1) << "Couldn't find a short ref picture to unmark";
return false;
}
to_unmark->ref = false;
}
return true;
}
bool H264Decoder::FinishPicture(scoped_refptr<H264Picture> pic) {
// Finish processing the picture.
// Start by storing previous picture data for later use.
if (pic->ref) {
ReferencePictureMarking(pic);
prev_ref_has_memmgmnt5_ = pic->mem_mgmt_5;
prev_ref_top_field_order_cnt_ = pic->top_field_order_cnt;
prev_ref_pic_order_cnt_msb_ = pic->pic_order_cnt_msb;
prev_ref_pic_order_cnt_lsb_ = pic->pic_order_cnt_lsb;
prev_ref_field_ = pic->field;
prev_ref_frame_num_ = pic->frame_num;
}
prev_frame_num_ = pic->frame_num;
prev_has_memmgmnt5_ = pic->mem_mgmt_5;
prev_frame_num_offset_ = pic->frame_num_offset;
// Remove unused (for reference or later output) pictures from DPB, marking
// them as such.
dpb_.DeleteUnused();
DVLOG(4) << "Finishing picture frame_num: " << pic->frame_num
<< ", entries in DPB: " << dpb_.size();
if (recovery_frame_cnt_) {
// This is the first picture after the recovery point SEI message. Computes
// the frame_num of the frame that should be output from (Spec D.2.8).
recovery_frame_num_ =
(*recovery_frame_cnt_ + pic->frame_num) % max_frame_num_;
DVLOG(3) << "recovery_frame_num_" << *recovery_frame_num_;
recovery_frame_cnt_.reset();
}
// The ownership of pic will either be transferred to DPB - if the picture is
// still needed (for output and/or reference) - or we will release it
// immediately if we manage to output it here and won't have to store it for
// future reference.
// Get all pictures that haven't been outputted yet.
H264Picture::Vector not_outputted;
dpb_.GetNotOutputtedPicsAppending(&not_outputted);
// Include the one we've just decoded.
not_outputted.push_back(pic);
// Sort in output order.
std::sort(not_outputted.begin(), not_outputted.end(), POCAscCompare());
// Try to output as many pictures as we can. A picture can be output,
// if the number of decoded and not yet outputted pictures that would remain
// in DPB afterwards would at least be equal to max_num_reorder_frames.
// If the outputted picture is not a reference picture, it doesn't have
// to remain in the DPB and can be removed.
auto output_candidate = not_outputted.begin();
size_t num_remaining = not_outputted.size();
while (num_remaining > max_num_reorder_frames_ ||
// If the condition below is used, this is an invalid stream. We should
// not be forced to output beyond max_num_reorder_frames in order to
// make room in DPB to store the current picture (if we need to do so).
// However, if this happens, ignore max_num_reorder_frames and try
// to output more. This may cause out-of-order output, but is not
// fatal, and better than failing instead.
((dpb_.IsFull() && (!pic->outputted || pic->ref)) && num_remaining)) {
DVLOG_IF(1, num_remaining <= max_num_reorder_frames_)
<< "Invalid stream: max_num_reorder_frames not preserved";
if (!recovery_frame_num_ ||
// If we are decoding ahead to reach a SEI recovery point, skip
// outputting all pictures before it, to avoid outputting corrupted
// frames.
(*output_candidate)->frame_num == *recovery_frame_num_) {
recovery_frame_num_ = absl::nullopt;
if (!OutputPic(*output_candidate))
return false;
}
if (!(*output_candidate)->ref) {
// Current picture hasn't been inserted into DPB yet, so don't remove it
// if we managed to output it immediately.
int outputted_poc = (*output_candidate)->pic_order_cnt;
if (outputted_poc != pic->pic_order_cnt)
dpb_.DeleteByPOC(outputted_poc);
}
++output_candidate;
--num_remaining;
}
// If we haven't managed to output the picture that we just decoded, or if
// it's a reference picture, we have to store it in DPB.
if (!pic->outputted || pic->ref) {
if (dpb_.IsFull()) {
// If we haven't managed to output anything to free up space in DPB
// to store this picture, it's an error in the stream.
DVLOG(1) << "Could not free up space in DPB!";
return false;
}
dpb_.StorePic(std::move(pic));
}
return true;
}
bool H264Decoder::UpdateMaxNumReorderFrames(const H264SPS* sps) {
if (sps->vui_parameters_present_flag && sps->bitstream_restriction_flag) {
max_num_reorder_frames_ =
base::checked_cast<size_t>(sps->max_num_reorder_frames);
if (max_num_reorder_frames_ > dpb_.max_num_pics()) {
DVLOG(1)
<< "max_num_reorder_frames present, but larger than MaxDpbFrames ("
<< max_num_reorder_frames_ << " > " << dpb_.max_num_pics() << ")";
max_num_reorder_frames_ = 0;
return false;
}
return true;
}
// max_num_reorder_frames not present, infer from profile/constraints
// (see VUI semantics in spec).
if (sps->constraint_set3_flag) {
switch (sps->profile_idc) {
case 44:
case 86:
case 100:
case 110:
case 122:
case 244:
max_num_reorder_frames_ = 0;
break;
default:
max_num_reorder_frames_ = dpb_.max_num_pics();
break;
}
} else {
max_num_reorder_frames_ = dpb_.max_num_pics();
}
return true;
}
bool H264Decoder::ProcessSPS(int sps_id, bool* need_new_buffers) {
DVLOG(4) << "Processing SPS id:" << sps_id;
const H264SPS* sps = parser_.GetSPS(sps_id);
if (!sps)
return false;
*need_new_buffers = false;
if (sps->frame_mbs_only_flag == 0) {
DVLOG(1) << "frame_mbs_only_flag != 1 not supported";
return false;
}
gfx::Size new_pic_size = sps->GetCodedSize().value_or(gfx::Size());
if (new_pic_size.IsEmpty()) {
DVLOG(1) << "Invalid picture size";
return false;
}
int width_mb = new_pic_size.width() / 16;
int height_mb = new_pic_size.height() / 16;
// Verify that the values are not too large before multiplying.
if (std::numeric_limits<int>::max() / width_mb < height_mb) {
DVLOG(1) << "Picture size is too big: " << new_pic_size.ToString();
return false;
}
// Spec A.3.1 and A.3.2
// For Baseline, Constrained Baseline and Main profile, the indicated level is
// Level 1b if level_idc is equal to 11 and constraint_set3_flag is equal to 1
uint8_t level = base::checked_cast<uint8_t>(sps->level_idc);
if ((sps->profile_idc == H264SPS::kProfileIDCBaseline ||
sps->profile_idc == H264SPS::kProfileIDCConstrainedBaseline ||
sps->profile_idc == H264SPS::kProfileIDCMain) &&
level == 11 && sps->constraint_set3_flag) {
level = 9; // Level 1b
}
int max_dpb_mbs = base::checked_cast<int>(H264LevelToMaxDpbMbs(level));
if (max_dpb_mbs == 0)
return false;
// MaxDpbFrames from level limits per spec.
size_t max_dpb_frames = std::min(max_dpb_mbs / (width_mb * height_mb),
static_cast<int>(H264DPB::kDPBMaxSize));
DVLOG(1) << "MaxDpbFrames: " << max_dpb_frames
<< ", max_num_ref_frames: " << sps->max_num_ref_frames
<< ", max_dec_frame_buffering: " << sps->max_dec_frame_buffering;
// Set DPB size to at least the level limit, or what the stream requires.
size_t max_dpb_size =
std::max(static_cast<int>(max_dpb_frames),
std::max(sps->max_num_ref_frames, sps->max_dec_frame_buffering));
// Some non-conforming streams specify more frames are needed than the current
// level limit. Allow this, but only up to the maximum number of reference
// frames allowed per spec.
DVLOG_IF(1, max_dpb_size > max_dpb_frames)
<< "Invalid stream, DPB size > MaxDpbFrames";
if (max_dpb_size == 0 || max_dpb_size > H264DPB::kDPBMaxSize) {
DVLOG(1) << "Invalid DPB size: " << max_dpb_size;
return false;
}
if (!IsYUV420Sequence(*sps)) {
DVLOG(1) << "Only YUV 4:2:0 is supported";
return false;
}
VideoCodecProfile new_profile =
H264Parser::ProfileIDCToVideoCodecProfile(sps->profile_idc);
uint8_t new_bit_depth = 0;
if (!ParseBitDepth(*sps, new_bit_depth))
return false;
if (!IsValidBitDepth(new_bit_depth, new_profile)) {
DVLOG(1) << "Invalid bit depth=" << base::strict_cast<int>(new_bit_depth)
<< ", profile=" << GetProfileName(new_profile);
return false;
}
if (pic_size_ != new_pic_size || dpb_.max_num_pics() != max_dpb_size ||
profile_ != new_profile || bit_depth_ != new_bit_depth) {
if (!Flush())
return false;
DVLOG(1) << "Codec profile: " << GetProfileName(new_profile)
<< ", level: " << level << ", DPB size: " << max_dpb_size
<< ", Picture size: " << new_pic_size.ToString()
<< ", bit depth: " << base::strict_cast<int>(new_bit_depth);
*need_new_buffers = true;
profile_ = new_profile;
bit_depth_ = new_bit_depth;
pic_size_ = new_pic_size;
dpb_.set_max_num_pics(max_dpb_size);
}
gfx::Rect new_visible_rect = sps->GetVisibleRect().value_or(gfx::Rect());
if (visible_rect_ != new_visible_rect) {
DVLOG(2) << "New visible rect: " << new_visible_rect.ToString();
visible_rect_ = new_visible_rect;
}
if (!UpdateMaxNumReorderFrames(sps))
return false;
DVLOG(1) << "max_num_reorder_frames: " << max_num_reorder_frames_;
return true;
}
H264Decoder::H264Accelerator::Status H264Decoder::FinishPrevFrameIfPresent() {
// If we already have a frame waiting to be decoded, decode it and finish.
if (curr_pic_) {
H264Accelerator::Status result = DecodePicture();
if (result != H264Accelerator::Status::kOk)
return result;
scoped_refptr<H264Picture> pic = curr_pic_;
curr_pic_ = nullptr;
if (!FinishPicture(pic))
return H264Accelerator::Status::kFail;
}
return H264Accelerator::Status::kOk;
}
bool H264Decoder::HandleFrameNumGap(int frame_num) {
const H264SPS* sps = parser_.GetSPS(curr_sps_id_);
if (!sps)
return false;
if (!sps->gaps_in_frame_num_value_allowed_flag) {
DVLOG(1) << "Invalid frame_num: " << frame_num;
// TODO(b:129119729, b:146914440): Youtube android app sometimes sends an
// invalid frame number after a seek. The sequence goes like:
// Seek, SPS, PPS, IDR-frame, non-IDR, ... non-IDR with invalid number.
// The only way to work around this reliably is to ignore this error.
// Video playback is not affected, no artefacts are visible.
// return false;
}
DVLOG(2) << "Handling frame_num gap: " << prev_ref_frame_num_ << "->"
<< frame_num;
// 7.4.3/7-23
int unused_short_term_frame_num = (prev_ref_frame_num_ + 1) % max_frame_num_;
while (unused_short_term_frame_num != frame_num) {
scoped_refptr<H264Picture> pic = new H264Picture();
if (!InitNonexistingPicture(pic, unused_short_term_frame_num))
return false;
UpdatePicNums(unused_short_term_frame_num);
if (!FinishPicture(pic))
return false;
unused_short_term_frame_num++;
unused_short_term_frame_num %= max_frame_num_;
}
return true;
}
H264Decoder::H264Accelerator::Status H264Decoder::ProcessEncryptedSliceHeader(
const std::vector<SubsampleEntry>& subsamples) {
DCHECK(curr_nalu_);
DCHECK(curr_slice_hdr_);
std::vector<base::span<const uint8_t>> spans(encrypted_sei_nalus_.size() + 1);
spans.assign(encrypted_sei_nalus_.begin(), encrypted_sei_nalus_.end());
spans.emplace_back(curr_nalu_->data, curr_nalu_->size);
std::vector<SubsampleEntry> all_subsamples(sei_subsamples_.size() + 1);
all_subsamples.assign(sei_subsamples_.begin(), sei_subsamples_.end());
all_subsamples.insert(all_subsamples.end(), subsamples.begin(),
subsamples.end());
return accelerator_->ParseEncryptedSliceHeader(spans, all_subsamples,
last_sps_nalu_, last_pps_nalu_,
curr_slice_hdr_.get());
}
H264Decoder::H264Accelerator::Status H264Decoder::PreprocessCurrentSlice() {
const H264SliceHeader* slice_hdr = curr_slice_hdr_.get();
DCHECK(slice_hdr);
if (IsNewPrimaryCodedPicture(curr_pic_.get(), curr_pps_id_,
parser_.GetSPS(curr_sps_id_), *slice_hdr)) {
// New picture, so first finish the previous one before processing it.
H264Accelerator::Status result = FinishPrevFrameIfPresent();
if (result != H264Accelerator::Status::kOk)
return result;
DCHECK(!curr_pic_);
if (slice_hdr->first_mb_in_slice != 0) {
DVLOG(1) << "ASO/invalid stream, first_mb_in_slice: "
<< slice_hdr->first_mb_in_slice;
return H264Accelerator::Status::kFail;
}
// If the new picture is an IDR, flush DPB.
if (slice_hdr->idr_pic_flag) {
// Output all remaining pictures, unless we are explicitly instructed
// not to do so.
if (!slice_hdr->no_output_of_prior_pics_flag) {
if (!Flush())
return H264Accelerator::Status::kFail;
}
dpb_.Clear();
last_output_poc_ = std::numeric_limits<int>::min();
}
}
return H264Accelerator::Status::kOk;
}
H264Decoder::H264Accelerator::Status H264Decoder::ProcessCurrentSlice() {
DCHECK(curr_pic_);
const H264SliceHeader* slice_hdr = curr_slice_hdr_.get();
DCHECK(slice_hdr);
if (slice_hdr->field_pic_flag == 0)
max_pic_num_ = max_frame_num_;
else
max_pic_num_ = 2 * max_frame_num_;
H264Picture::Vector ref_pic_list0, ref_pic_list1;
// If we are using full sample encryption then we do not have the information
// we need to update the ref pic lists here, but that's OK because the
// accelerator doesn't actually need to submit them in this case.
if (!slice_hdr->full_sample_encryption &&
!ModifyReferencePicLists(slice_hdr, &ref_pic_list0, &ref_pic_list1)) {
return H264Accelerator::Status::kFail;
}
const H264PPS* pps = parser_.GetPPS(curr_pps_id_);
if (!pps)
return H264Accelerator::Status::kFail;
return accelerator_->SubmitSlice(pps, slice_hdr, ref_pic_list0, ref_pic_list1,
curr_pic_.get(), slice_hdr->nalu_data,
slice_hdr->nalu_size,
parser_.GetCurrentSubsamples());
}
#define SET_ERROR_AND_RETURN() \
do { \
DVLOG(1) << "Error during decode"; \
state_ = kError; \
return H264Decoder::kDecodeError; \
} while (0)
#define CHECK_ACCELERATOR_RESULT(func) \
do { \
H264Accelerator::Status result = (func); \
switch (result) { \
case H264Accelerator::Status::kOk: \
break; \
case H264Accelerator::Status::kTryAgain: \
DVLOG(1) << #func " needs to try again"; \
return H264Decoder::kTryAgain; \
case H264Accelerator::Status::kFail: \
case H264Accelerator::Status::kNotSupported: \
SET_ERROR_AND_RETURN(); \
} \
} while (0)
void H264Decoder::SetStream(int32_t id, const DecoderBuffer& decoder_buffer) {
const uint8_t* ptr = decoder_buffer.data();
const size_t size = decoder_buffer.data_size();
const DecryptConfig* decrypt_config = decoder_buffer.decrypt_config();
DCHECK(ptr);
DCHECK(size);
DVLOG(4) << "New input stream id: " << id << " at: " << (void*)ptr
<< " size: " << size;
stream_id_ = id;
current_stream_ = ptr;
current_stream_size_ = size;
current_stream_has_been_changed_ = true;
encrypted_sei_nalus_.clear();
sei_subsamples_.clear();
if (decrypt_config) {
parser_.SetEncryptedStream(ptr, size, decrypt_config->subsamples());
current_decrypt_config_ = decrypt_config->Clone();
} else {
parser_.SetStream(ptr, size);
current_decrypt_config_ = nullptr;
}
}
H264Decoder::DecodeResult H264Decoder::Decode() {
if (state_ == kError) {
DVLOG(1) << "Decoder in error state";
return kDecodeError;
}
if (current_stream_has_been_changed_) {
// Calling H264Accelerator::SetStream() here instead of when the stream is
// originally set in case the accelerator needs to return kTryAgain.
H264Accelerator::Status result = accelerator_->SetStream(
base::span<const uint8_t>(current_stream_, current_stream_size_),
current_decrypt_config_.get());
switch (result) {
case H264Accelerator::Status::kOk:
case H264Accelerator::Status::kNotSupported:
// kNotSupported means the accelerator can't handle this stream,
// so everything will be done through the parser.
break;
case H264Accelerator::Status::kTryAgain:
DVLOG(1) << "SetStream() needs to try again";
return H264Decoder::kTryAgain;
case H264Accelerator::Status::kFail:
SET_ERROR_AND_RETURN();
}
// Reset the flag so that this is only called again next time SetStream()
// is called.
current_stream_has_been_changed_ = false;
}
while (1) {
H264Parser::Result par_res;
if (!curr_nalu_) {
curr_nalu_ = std::make_unique<H264NALU>();
par_res = parser_.AdvanceToNextNALU(curr_nalu_.get());
if (par_res == H264Parser::kEOStream) {
CHECK_ACCELERATOR_RESULT(FinishPrevFrameIfPresent());
return kRanOutOfStreamData;
} else if (par_res != H264Parser::kOk) {
SET_ERROR_AND_RETURN();
}
DVLOG(4) << "New NALU: " << static_cast<int>(curr_nalu_->nal_unit_type);
}
switch (curr_nalu_->nal_unit_type) {
case H264NALU::kNonIDRSlice:
// We can't resume from a non-IDR slice unless recovery point SEI
// process is going.
if (state_ == kError || (state_ == kAfterReset && !recovery_frame_cnt_))
break;
FALLTHROUGH;
case H264NALU::kIDRSlice: {
// TODO(posciak): the IDR may require an SPS that we don't have
// available. For now we'd fail if that happens, but ideally we'd like
// to keep going until the next SPS in the stream.
if (state_ == kNeedStreamMetadata) {
// We need an SPS, skip this IDR and keep looking.
break;
}
// If after reset or waiting for a key, we should be able to recover
// from an IDR. |state_|, |curr_slice_hdr_|, and |curr_pic_| are used
// to keep track of what has previously been attempted, so that after
// a retryable result is returned, subsequent calls to Decode() retry
// the call that failed previously. If it succeeds (it may not if no
// additional key has been provided, for example), then the remaining
// steps will be executed.
if (!curr_slice_hdr_) {
curr_slice_hdr_ = std::make_unique<H264SliceHeader>();
state_ = kParseSliceHeader;
}
if (state_ == kParseSliceHeader) {
// Check if the slice header is encrypted.
bool parsed_header = false;
if (current_decrypt_config_) {
const std::vector<SubsampleEntry>& subsamples =
parser_.GetCurrentSubsamples();
// There is only a single clear byte for the NALU information for
// full sample encryption, and the rest is encrypted.
if (!subsamples.empty() && subsamples[0].clear_bytes == 1) {
CHECK_ACCELERATOR_RESULT(ProcessEncryptedSliceHeader(subsamples));
parsed_header = true;
curr_slice_hdr_->pic_parameter_set_id = last_parsed_pps_id_;
encrypted_sei_nalus_.clear();
sei_subsamples_.clear();
}
}
if (!parsed_header) {
par_res =
parser_.ParseSliceHeader(*curr_nalu_, curr_slice_hdr_.get());
if (par_res != H264Parser::kOk)
SET_ERROR_AND_RETURN();
}
state_ = kTryPreprocessCurrentSlice;
}
if (state_ == kTryPreprocessCurrentSlice) {
CHECK_ACCELERATOR_RESULT(PreprocessCurrentSlice());
state_ = kEnsurePicture;
}
if (state_ == kEnsurePicture) {
if (curr_pic_) {
// |curr_pic_| already exists, so skip to ProcessCurrentSlice().
state_ = kTryCurrentSlice;
} else {
// New picture/finished previous one, try to start a new one
// or tell the client we need more surfaces.
curr_pic_ = accelerator_->CreateH264Picture();
if (!curr_pic_)
return kRanOutOfSurfaces;
if (current_decrypt_config_)
curr_pic_->set_decrypt_config(current_decrypt_config_->Clone());
state_ = kTryNewFrame;
}
}
if (state_ == kTryNewFrame) {
CHECK_ACCELERATOR_RESULT(StartNewFrame(curr_slice_hdr_.get()));
state_ = kTryCurrentSlice;
}
DCHECK_EQ(state_, kTryCurrentSlice);
CHECK_ACCELERATOR_RESULT(ProcessCurrentSlice());
curr_slice_hdr_.reset();
state_ = kDecoding;
break;
}
case H264NALU::kSPS: {
int sps_id;
CHECK_ACCELERATOR_RESULT(FinishPrevFrameIfPresent());
par_res = parser_.ParseSPS(&sps_id);
if (par_res != H264Parser::kOk)
SET_ERROR_AND_RETURN();
bool need_new_buffers = false;
if (!ProcessSPS(sps_id, &need_new_buffers))
SET_ERROR_AND_RETURN();
last_sps_nalu_.assign(curr_nalu_->data,
curr_nalu_->data + curr_nalu_->size);
if (state_ == kNeedStreamMetadata)
state_ = kAfterReset;
if (need_new_buffers) {
curr_pic_ = nullptr;
curr_nalu_ = nullptr;
ref_pic_list_p0_.clear();
ref_pic_list_b0_.clear();
ref_pic_list_b1_.clear();
return kConfigChange;
}
break;
}
case H264NALU::kPPS: {
CHECK_ACCELERATOR_RESULT(FinishPrevFrameIfPresent());
par_res = parser_.ParsePPS(&last_parsed_pps_id_);
if (par_res != H264Parser::kOk)
SET_ERROR_AND_RETURN();
last_pps_nalu_.assign(curr_nalu_->data,
curr_nalu_->data + curr_nalu_->size);
break;
}
case H264NALU::kAUD:
case H264NALU::kEOSeq:
case H264NALU::kEOStream:
if (state_ != kDecoding)
break;
CHECK_ACCELERATOR_RESULT(FinishPrevFrameIfPresent());
break;
case H264NALU::kSEIMessage:
if (current_decrypt_config_) {
// If there are encrypted SEI NALUs as part of CENCv1, then we also
// need to save those so we can send them into the accelerator so it
// can decrypt the sample properly (otherwise it would be starting
// partway into a block).
const std::vector<SubsampleEntry>& subsamples =
parser_.GetCurrentSubsamples();
if (!subsamples.empty()) {
encrypted_sei_nalus_.emplace_back(curr_nalu_->data,
curr_nalu_->size);
DCHECK_EQ(1u, subsamples.size());
sei_subsamples_.push_back(subsamples[0]);
}
}
if (state_ == kAfterReset && !recovery_frame_cnt_ &&
!recovery_frame_num_) {
// If we are after reset, we can also resume from a SEI recovery point
// (spec D.2.8) if one is present. However, if we are already in the
// process of handling one, skip any subsequent ones until we are done
// processing.
H264SEIMessage sei{};
if (parser_.ParseSEI(&sei) != H264Parser::kOk)
SET_ERROR_AND_RETURN();
if (sei.type == H264SEIMessage::kSEIRecoveryPoint) {
recovery_frame_cnt_ = sei.recovery_point.recovery_frame_cnt;
if (0 > *recovery_frame_cnt_ ||
*recovery_frame_cnt_ >= max_frame_num_) {
DVLOG(1) << "Invalid recovery_frame_cnt=" << *recovery_frame_cnt_
<< " (it must be [0, max_frame_num_-1="
<< max_frame_num_ - 1 << "])";
SET_ERROR_AND_RETURN();
}
DVLOG(3) << "Recovery point SEI is found, recovery_frame_cnt_="
<< *recovery_frame_cnt_;
break;
}
}
FALLTHROUGH;
default:
DVLOG(4) << "Skipping NALU type: " << curr_nalu_->nal_unit_type;
break;
}
DVLOG(4) << "NALU done";
curr_nalu_.reset();
}
}
gfx::Size H264Decoder::GetPicSize() const {
return pic_size_;
}
gfx::Rect H264Decoder::GetVisibleRect() const {
return visible_rect_;
}
VideoCodecProfile H264Decoder::GetProfile() const {
return profile_;
}
uint8_t H264Decoder::GetBitDepth() const {
return bit_depth_;
}
size_t H264Decoder::GetRequiredNumOfPictures() const {
constexpr size_t kPicsInPipeline = limits::kMaxVideoFrames + 1;
return GetNumReferenceFrames() + kPicsInPipeline;
}
size_t H264Decoder::GetNumReferenceFrames() const {
// Use the maximum number of pictures in the Decoded Picture Buffer.
return dpb_.max_num_pics();
}
// static
bool H264Decoder::FillH264PictureFromSliceHeader(
const H264SPS* sps,
const H264SliceHeader& slice_hdr,
H264Picture* pic) {
DCHECK(pic);
pic->idr = slice_hdr.idr_pic_flag;
if (pic->idr)
pic->idr_pic_id = slice_hdr.idr_pic_id;
if (slice_hdr.field_pic_flag) {
pic->field = slice_hdr.bottom_field_flag ? H264Picture::FIELD_BOTTOM
: H264Picture::FIELD_TOP;
} else {
pic->field = H264Picture::FIELD_NONE;
}
if (pic->field != H264Picture::FIELD_NONE) {
DVLOG(1) << "Interlaced video not supported.";
return false;
}
pic->nal_ref_idc = slice_hdr.nal_ref_idc;
pic->ref = slice_hdr.nal_ref_idc != 0;
// This assumes non-interlaced stream.
pic->frame_num = pic->pic_num = slice_hdr.frame_num;
if (!sps)
return false;
pic->pic_order_cnt_type = sps->pic_order_cnt_type;
switch (pic->pic_order_cnt_type) {
case 0:
pic->pic_order_cnt_lsb = slice_hdr.pic_order_cnt_lsb;
pic->delta_pic_order_cnt_bottom = slice_hdr.delta_pic_order_cnt_bottom;
break;
case 1:
pic->delta_pic_order_cnt0 = slice_hdr.delta_pic_order_cnt0;
pic->delta_pic_order_cnt1 = slice_hdr.delta_pic_order_cnt1;
break;
case 2:
break;
default:
NOTREACHED();
return false;
}
return true;
}
// static
bool H264Decoder::IsNewPrimaryCodedPicture(const H264Picture* curr_pic,
int curr_pps_id,
const H264SPS* sps,
const H264SliceHeader& slice_hdr) {
if (!curr_pic)
return true;
// 7.4.1.2.4, assumes non-interlaced.
if (slice_hdr.frame_num != curr_pic->frame_num ||
slice_hdr.pic_parameter_set_id != curr_pps_id ||
slice_hdr.nal_ref_idc != curr_pic->nal_ref_idc ||
slice_hdr.idr_pic_flag != curr_pic->idr ||
(slice_hdr.idr_pic_flag &&
(slice_hdr.idr_pic_id != curr_pic->idr_pic_id ||
// If we have two consecutive IDR slices, and the second one has
// first_mb_in_slice == 0, treat it as a new picture.
// Per spec, idr_pic_id should not be equal in this case (and we should
// have hit the condition above instead, see spec 7.4.3 on idr_pic_id),
// but some encoders neglect changing idr_pic_id for two consecutive
// IDRs. Work around this by checking if the next slice contains the
// zeroth macroblock, i.e. data that belongs to the next picture.
slice_hdr.first_mb_in_slice == 0)))
return true;
if (!sps)
return false;
if (sps->pic_order_cnt_type == curr_pic->pic_order_cnt_type) {
if (curr_pic->pic_order_cnt_type == 0) {
if (slice_hdr.pic_order_cnt_lsb != curr_pic->pic_order_cnt_lsb ||
slice_hdr.delta_pic_order_cnt_bottom !=
curr_pic->delta_pic_order_cnt_bottom)
return true;
} else if (curr_pic->pic_order_cnt_type == 1) {
if (slice_hdr.delta_pic_order_cnt0 != curr_pic->delta_pic_order_cnt0 ||
slice_hdr.delta_pic_order_cnt1 != curr_pic->delta_pic_order_cnt1)
return true;
}
}
return false;
}
} // namespace media