blob: 5cea679a42838ed84f92449b3fe1125a4d019356 [file]
// Copyright 2021 The Chromium Authors
// Use of this source code is governed by a BSD-style license that can be
// found in the LICENSE file.
#include "media/gpu/vaapi/test/h264_decoder.h"
#include "base/notreached.h"
#include "base/ranges/algorithm.h"
#include "media/base/subsample_entry.h"
#include "media/gpu/macros.h"
#include "media/gpu/vaapi/test/h264_dpb.h"
#include "media/gpu/vaapi/test/video_decoder.h"
#include "media/video/h264_parser.h"
#include <va/va.h>
namespace media::vaapi_test {
namespace {
struct PicNumDescCompare {
bool operator()(const scoped_refptr<H264Picture>& a,
const scoped_refptr<H264Picture>& b) const {
return a->pic_num > b->pic_num;
}
};
struct LongTermPicNumAscCompare {
bool operator()(const scoped_refptr<H264Picture>& a,
const scoped_refptr<H264Picture>& b) const {
return a->long_term_pic_num < b->long_term_pic_num;
}
};
struct POCAscCompare {
bool operator()(const scoped_refptr<H264Picture>& a,
const scoped_refptr<H264Picture>& b) const {
return a->pic_order_cnt < b->pic_order_cnt;
}
};
struct POCDescCompare {
bool operator()(const scoped_refptr<H264Picture>& a,
const scoped_refptr<H264Picture>& b) const {
return a->pic_order_cnt > b->pic_order_cnt;
}
};
bool FillH264PictureFromSliceHeader(const H264SPS* sps,
const H264SliceHeader& slice_hdr,
H264Picture* pic) {
DCHECK(pic);
pic->idr = slice_hdr.idr_pic_flag;
if (pic->idr)
pic->idr_pic_id = slice_hdr.idr_pic_id;
if (slice_hdr.field_pic_flag) {
pic->field = slice_hdr.bottom_field_flag ? H264Picture::FIELD_BOTTOM
: H264Picture::FIELD_TOP;
} else {
pic->field = H264Picture::FIELD_NONE;
}
if (pic->field != H264Picture::FIELD_NONE) {
VLOG(1) << "Interlaced video not supported.";
return false;
}
pic->nal_ref_idc = slice_hdr.nal_ref_idc;
pic->ref = slice_hdr.nal_ref_idc != 0;
// This assumes non-interlaced stream.
pic->frame_num = pic->pic_num = slice_hdr.frame_num;
if (!sps)
return false;
pic->pic_order_cnt_type = sps->pic_order_cnt_type;
switch (pic->pic_order_cnt_type) {
case 0:
pic->pic_order_cnt_lsb = slice_hdr.pic_order_cnt_lsb;
pic->delta_pic_order_cnt_bottom = slice_hdr.delta_pic_order_cnt_bottom;
break;
case 1:
pic->delta_pic_order_cnt0 = slice_hdr.delta_pic_order_cnt0;
pic->delta_pic_order_cnt1 = slice_hdr.delta_pic_order_cnt1;
break;
case 2:
break;
default:
NOTREACHED();
return false;
}
return true;
}
} // namespace
H264Decoder::H264Decoder(const uint8_t* stream_data,
size_t stream_len,
const VaapiDevice& va_device,
SharedVASurface::FetchPolicy fetch_policy)
: VideoDecoder::VideoDecoder(va_device, fetch_policy),
parser_(std::make_unique<H264Parser>()),
curr_sps_id_(-1),
curr_pps_id_(-1),
curr_slice_hdr_(nullptr),
curr_nalu_(nullptr),
curr_picture_(nullptr),
is_stream_over_(false),
va_wrapper_(va_device) {
parser_->SetStream(stream_data, stream_len);
LOG_ASSERT(GetStreamMetadata()) << "Stream contains no Sequence Parameters!";
ExtractSliceHeader();
}
H264Decoder::~H264Decoder() {
dpb_.Clear();
last_decoded_surface_.reset();
}
VideoDecoder::Result H264Decoder::DecodeNextFrame() {
while (!is_stream_over_ && output_queue.empty())
DecodeNextFrameInStream();
if (is_stream_over_ && output_queue.empty())
return VideoDecoder::kEOStream;
last_decoded_surface_ = output_queue.front()->surface;
VLOG(4) << "Outputting frame poc: " << output_queue.front()->pic_order_cnt;
last_decoded_frame_visible_ = !output_queue.front()->nonexisting;
output_queue.pop();
return VideoDecoder::kOk;
}
// H264 does not guarantee frames appear in the order they need to be shown in.
// This method merely decodes the next frame in the stream, it's up to the
// caller to actually figure out how to order them.
void H264Decoder::DecodeNextFrameInStream() {
StartNewFrame();
ProcessSlice();
while (true) {
curr_nalu_ = std::make_unique<H264NALU>();
H264Parser::Result parser_result =
parser_->AdvanceToNextNALU(curr_nalu_.get());
if (parser_result == H264Parser::kEOStream) {
is_stream_over_ = true;
break;
}
if (curr_nalu_->nal_unit_type == H264NALU::kNonIDRSlice ||
curr_nalu_->nal_unit_type == H264NALU::kIDRSlice) {
ExtractSliceHeader();
if (IsNewFrame())
break;
ProcessSlice();
} else if (curr_nalu_->nal_unit_type == H264NALU::kSPS) {
UpdateSequenceParams();
} else if (curr_nalu_->nal_unit_type == H264NALU::kPPS) {
UpdatePictureParams();
} else {
VLOG(4) << "Skipping NALU type " << curr_nalu_->nal_unit_type;
}
}
DecodeFrame();
FinishPicture(curr_picture_);
if (is_stream_over_) {
FlushDPB();
}
}
void H264Decoder::ProcessSlice() {
if (!curr_slice_hdr_->field_pic_flag) {
max_pic_num_ = max_frame_num_;
} else {
max_pic_num_ = 2 * max_frame_num_;
}
H264Picture::Vector ref_pic_list0, ref_pic_list1;
LOG_ASSERT(ModifyReferencePicLists(curr_slice_hdr_.get(), &ref_pic_list0,
&ref_pic_list1))
<< "Error modify reference pic lists!";
const H264PPS* pic_params = parser_->GetPPS(curr_pps_id_);
LOG_ASSERT(pic_params) << "No picture params for ID " << curr_pps_id_;
va_wrapper_.SubmitSlice(
pic_params, curr_slice_hdr_.get(), ref_pic_list0, ref_pic_list1,
curr_picture_.get(), curr_slice_hdr_->nalu_data,
curr_slice_hdr_->nalu_size, parser_->GetCurrentSubsamples());
}
void H264Decoder::UpdateSequenceParams() {
LOG_ASSERT(parser_->ParseSPS(&curr_sps_id_) == H264Parser::kOk)
<< "Error parsing SPS NALU";
const H264SPS* sps = parser_->GetSPS(curr_sps_id_);
LOG_ASSERT(sps) << "SPS not present for ID " << curr_sps_id_;
const gfx::Rect new_visible_rect =
sps->GetVisibleRect().value_or(gfx::Rect());
if (visible_rect_ != new_visible_rect) {
VLOG(2) << "New visible rect: " << new_visible_rect.ToString();
visible_rect_ = new_visible_rect;
}
const gfx::Size new_pic_size = sps->GetCodedSize().value_or(gfx::Size());
LOG_ASSERT(!new_pic_size.IsEmpty()) << "Invalid picture size";
const int width_mb = base::checked_cast<int>(new_pic_size.width()) / 16;
const int height_mb = base::checked_cast<int>(new_pic_size.height()) / 16;
LOG_ASSERT(std::numeric_limits<int>::max() / width_mb > height_mb)
<< "Picture size is too big: " << new_pic_size.ToString();
// Spec A.3.1 and A.3.2
// For Baseline, Constrained Baseline and Main profile, the indicated level is
// Level 1b if level_idc is equal to 11 and constraint_set3_flag is equal to 1
uint8_t level = base::checked_cast<uint8_t>(sps->level_idc);
if ((sps->profile_idc == H264SPS::kProfileIDCBaseline ||
sps->profile_idc == H264SPS::kProfileIDCConstrainedBaseline ||
sps->profile_idc == H264SPS::kProfileIDCMain) &&
level == 11 && sps->constraint_set3_flag) {
level = 9; // Level 1b
}
int max_dpb_mbs = base::checked_cast<int>(H264LevelToMaxDpbMbs(level));
LOG_ASSERT(max_dpb_mbs) << "Invalid profile level " << level;
// MaxDpbFrames from level limits per spec.
size_t max_dpb_frames = std::min(max_dpb_mbs / (width_mb * height_mb),
static_cast<int>(H264DPB::kDPBMaxSize));
VLOG(1) << "MaxDpbFrames: " << max_dpb_frames
<< ", max_num_ref_frames: " << sps->max_num_ref_frames
<< ", max_dec_frame_buffering: " << sps->max_dec_frame_buffering;
// Set DPB size to at least the level limit, or what the stream requires.
size_t max_dpb_size =
std::max(static_cast<int>(max_dpb_frames),
std::max(sps->max_num_ref_frames, sps->max_dec_frame_buffering));
// Some non-conforming streams specify more frames are needed than the current
// level limit. Allow this, but only up to the maximum number of reference
// frames allowed per spec.
LOG_ASSERT(max_dpb_size <= max_dpb_frames)
<< "Invalid stream, DPB size > MaxDpbFrames";
LOG_ASSERT(max_dpb_size != 0 && max_dpb_size <= H264DPB::kDPBMaxSize)
<< "Invalid DPB size: " << max_dpb_size;
LOG_ASSERT(sps->chroma_format_idc == 1) << "Only YUV 4:2:0 is supported";
dpb_.set_max_num_pics(max_dpb_size);
UpdateMaxNumReorderFrames(sps);
}
void H264Decoder::UpdatePictureParams() {
LOG_ASSERT(parser_->ParsePPS(&curr_pps_id_) == H264Parser::kOk)
<< "Error parsing PPS NALU";
}
void H264Decoder::DecodeFrame() {
va_wrapper_.SubmitDecode(curr_picture_);
}
void H264Decoder::FinishPicture(scoped_refptr<H264Picture> pic) {
// Finish processing the picture.
// Start by storing previous picture data for later use.
if (pic->ref) {
ReferencePictureMarking(pic);
prev_ref_has_memmgmnt5_ = pic->mem_mgmt_5;
prev_ref_top_field_order_cnt_ = pic->top_field_order_cnt;
prev_ref_pic_order_cnt_msb_ = pic->pic_order_cnt_msb;
prev_ref_pic_order_cnt_lsb_ = pic->pic_order_cnt_lsb;
prev_ref_field_ = pic->field;
prev_ref_frame_num_ = pic->frame_num;
}
prev_frame_num_ = pic->frame_num;
prev_has_memmgmnt5_ = pic->mem_mgmt_5;
prev_frame_num_offset_ = pic->frame_num_offset;
dpb_.DeleteUnused();
VLOG(4) << "Finishing picture frame_num: " << pic->frame_num
<< ", poc: " << pic->pic_order_cnt
<< ", entries in DPB: " << dpb_.size();
if (recovery_frame_cnt_) {
// This is the first picture after the recovery point SEI message. Computes
// the frame_num of the frame that should be output from (Spec D.2.8).
recovery_frame_num_ =
(*recovery_frame_cnt_ + pic->frame_num) % max_frame_num_;
VLOG(3) << "recovery_frame_num_" << *recovery_frame_num_;
recovery_frame_cnt_.reset();
}
// Get all pictures that haven't been outputted yet.
H264Picture::Vector not_outputted;
dpb_.GetNotOutputtedPicsAppending(&not_outputted);
// Include the one we've just decoded.
not_outputted.push_back(pic);
// Sort in output order.
std::sort(not_outputted.begin(), not_outputted.end(), POCAscCompare());
// Try to output as many pictures as we can. A picture can be output,
// if the number of decoded and not yet outputted pictures that would remain
// in DPB afterwards would at least be equal to max_num_reorder_frames.
// If the outputted picture is not a reference picture, it doesn't have
// to remain in the DPB and can be removed.
auto output_candidate = not_outputted.begin();
size_t num_remaining = not_outputted.size();
while (num_remaining > max_num_reorder_frames_ ||
// If the condition below is used, this is an invalid stream. We should
// not be forced to output beyond max_num_reorder_frames in order to
// make room in DPB to store the current picture (if we need to do so).
// However, if this happens, ignore max_num_reorder_frames and try
// to output more. This may cause out-of-order output, but is not
// fatal, and better than failing instead.
((dpb_.IsFull() && (!pic->outputted || pic->ref)) && num_remaining)) {
VLOG_IF(1, num_remaining <= max_num_reorder_frames_)
<< "Invalid stream: max_num_reorder_frames not preserved";
if (!recovery_frame_num_ ||
// If we are decoding ahead to reach a SEI recovery point, skip
// outputting all pictures before it, to avoid outputting corrupted
// frames.
(*output_candidate)->frame_num == *recovery_frame_num_) {
recovery_frame_num_ = absl::nullopt;
output_queue.push(*output_candidate);
(*output_candidate)->outputted = true;
}
if (!(*output_candidate)->ref) {
// Current picture hasn't been inserted into DPB yet, so don't remove it
// if we managed to output it immediately.
int outputted_poc = (*output_candidate)->pic_order_cnt;
if (outputted_poc != pic->pic_order_cnt) {
dpb_.DeleteByPOC(outputted_poc);
}
}
++output_candidate;
--num_remaining;
}
// If we haven't managed to output the picture that we just decoded, or if
// it's a reference picture, we have to store it in DPB.
if (!pic->outputted || pic->ref) {
if (dpb_.IsFull()) {
// If we haven't managed to output anything to free up space in DPB
// to store this picture, it's an error in the stream.
VLOG(1) << "Could not free up space in DPB!";
}
dpb_.StorePic(std::move(pic));
}
}
bool H264Decoder::GetStreamMetadata() {
bool found_sps = false;
while (true) {
curr_nalu_ = std::make_unique<H264NALU>();
H264Parser::Result parser_result =
parser_->AdvanceToNextNALU(curr_nalu_.get());
if (parser_result == H264Parser::kEOStream)
return found_sps;
if (curr_nalu_->nal_unit_type == H264NALU::kNonIDRSlice ||
curr_nalu_->nal_unit_type == H264NALU::kIDRSlice) {
return found_sps;
} else if (curr_nalu_->nal_unit_type == H264NALU::kSPS) {
found_sps = true;
UpdateSequenceParams();
} else if (curr_nalu_->nal_unit_type == H264NALU::kPPS) {
UpdatePictureParams();
} else {
VLOG(4) << "Skipping NALU type " << curr_nalu_->nal_unit_type;
}
}
}
bool H264Decoder::IsNewFrame() {
if (curr_slice_hdr_->frame_num != curr_picture_->frame_num ||
curr_slice_hdr_->pic_parameter_set_id != curr_pps_id_ ||
curr_slice_hdr_->nal_ref_idc != curr_picture_->nal_ref_idc ||
curr_slice_hdr_->idr_pic_flag != curr_picture_->idr ||
(curr_slice_hdr_->idr_pic_flag &&
(curr_slice_hdr_->idr_pic_id != curr_picture_->idr_pic_id ||
curr_slice_hdr_->first_mb_in_slice == 0))) {
return true;
}
const H264SPS* sequence_params = parser_->GetSPS(curr_sps_id_);
if (!sequence_params)
return false;
if (sequence_params->pic_order_cnt_type ==
curr_picture_->pic_order_cnt_type) {
if (curr_picture_->pic_order_cnt_type == 0) {
if (curr_slice_hdr_->pic_order_cnt_lsb !=
curr_picture_->pic_order_cnt_lsb ||
curr_slice_hdr_->delta_pic_order_cnt_bottom !=
curr_picture_->delta_pic_order_cnt_bottom) {
return true;
}
} else if (curr_picture_->pic_order_cnt_type == 1) {
if (curr_slice_hdr_->delta_pic_order_cnt0 !=
curr_picture_->delta_pic_order_cnt0 ||
curr_slice_hdr_->delta_pic_order_cnt1 !=
curr_picture_->delta_pic_order_cnt1) {
return true;
}
}
}
return false;
}
void H264Decoder::ExtractSliceHeader() {
curr_slice_hdr_ = std::make_unique<H264SliceHeader>();
H264Parser::Result result =
parser_->ParseSliceHeader(*curr_nalu_, curr_slice_hdr_.get());
LOG_ASSERT(result == H264Parser::kOk)
<< "Error parsing slice header! Parser returned " << result;
}
void H264Decoder::StartNewFrame() {
if (curr_slice_hdr_->idr_pic_flag) {
if (!curr_slice_hdr_->no_output_of_prior_pics_flag)
FlushDPB();
dpb_.Clear();
}
curr_pps_id_ = curr_slice_hdr_->pic_parameter_set_id;
const H264PPS* picture_params = parser_->GetPPS(curr_pps_id_);
LOG_ASSERT(picture_params)
<< "Error extracting picture parameters for ID " << curr_pps_id_;
curr_sps_id_ = picture_params->seq_parameter_set_id;
const H264SPS* sequence_params = parser_->GetSPS(curr_sps_id_);
LOG_ASSERT(sequence_params)
<< "Error extracting sequence parameters for ID" << curr_sps_id_;
curr_picture_ = va_wrapper_.CreatePicture(sequence_params);
max_frame_num_ = 1 << (sequence_params->log2_max_frame_num_minus4 + 4);
int frame_num = curr_slice_hdr_->frame_num;
if (curr_slice_hdr_->idr_pic_flag)
prev_ref_frame_num_ = 0;
if (frame_num != prev_ref_frame_num_ &&
frame_num != (prev_ref_frame_num_ + 1) % max_frame_num_) {
LOG_ASSERT(HandleFrameNumGap(frame_num)) << "Error handling frame num gap";
}
LOG_ASSERT(InitCurrPicture(curr_slice_hdr_.get())) << "Error initializing"
<< " picture.";
UpdatePicNums(frame_num);
ConstructReferencePicListsP();
ConstructReferencePicListsB();
va_wrapper_.SubmitFrameMetadata(sequence_params, picture_params, dpb_,
ref_pic_list_p0_, ref_pic_list_b0_,
ref_pic_list_b1_, curr_picture_);
}
bool H264Decoder::InitCurrPicture(const H264SliceHeader* slice_hdr) {
if (!FillH264PictureFromSliceHeader(parser_->GetSPS(curr_sps_id_), *slice_hdr,
curr_picture_.get())) {
return false;
}
if (!CalculatePicOrderCounts(curr_picture_))
return false;
curr_picture_->long_term_reference_flag = slice_hdr->long_term_reference_flag;
curr_picture_->adaptive_ref_pic_marking_mode_flag =
slice_hdr->adaptive_ref_pic_marking_mode_flag;
// If the slice header indicates we will have to perform reference marking
// process after this picture is decoded, store required data for that
// purpose.
if (slice_hdr->adaptive_ref_pic_marking_mode_flag) {
static_assert(sizeof(curr_picture_->ref_pic_marking) ==
sizeof(slice_hdr->ref_pic_marking),
"Array sizes of ref pic marking do not match.");
memcpy(curr_picture_->ref_pic_marking, slice_hdr->ref_pic_marking,
sizeof(curr_picture_->ref_pic_marking));
}
curr_picture_->visible_rect = visible_rect_;
return true;
}
void H264Decoder::ConstructReferencePicListsP() {
// RefPicList0 (8.2.4.2.1) [[1] [2]], where:
// [1] shortterm ref pics sorted by descending pic_num,
// [2] longterm ref pics by ascending long_term_pic_num.
ref_pic_list_p0_.clear();
// First get the short ref pics...
dpb_.GetShortTermRefPicsAppending(&ref_pic_list_p0_);
size_t num_short_refs = ref_pic_list_p0_.size();
// and sort them to get [1].
std::sort(ref_pic_list_p0_.begin(), ref_pic_list_p0_.end(),
PicNumDescCompare());
// Now get long term pics and sort them by long_term_pic_num to get [2].
dpb_.GetLongTermRefPicsAppending(&ref_pic_list_p0_);
std::sort(ref_pic_list_p0_.begin() + num_short_refs, ref_pic_list_p0_.end(),
LongTermPicNumAscCompare());
}
void H264Decoder::ConstructReferencePicListsB() {
// RefPicList0 (8.2.4.2.3) [[1] [2] [3]], where:
// [1] shortterm ref pics with POC < curr_pic's POC sorted by descending POC,
// [2] shortterm ref pics with POC > curr_pic's POC by ascending POC,
// [3] longterm ref pics by ascending long_term_pic_num.
ref_pic_list_b0_.clear();
ref_pic_list_b1_.clear();
dpb_.GetShortTermRefPicsAppending(&ref_pic_list_b0_);
size_t num_short_refs = ref_pic_list_b0_.size();
// First sort ascending, this will put [1] in right place and finish [2].
std::sort(ref_pic_list_b0_.begin(), ref_pic_list_b0_.end(), POCAscCompare());
// Find first with POC > curr_pic's POC to get first element in [2]...
H264Picture::Vector::iterator iter;
iter = std::upper_bound(ref_pic_list_b0_.begin(), ref_pic_list_b0_.end(),
curr_picture_.get(), POCAscCompare());
// and sort [1] descending, thus finishing sequence [1] [2].
std::sort(ref_pic_list_b0_.begin(), iter, POCDescCompare());
// Now add [3] and sort by ascending long_term_pic_num.
dpb_.GetLongTermRefPicsAppending(&ref_pic_list_b0_);
std::sort(ref_pic_list_b0_.begin() + num_short_refs, ref_pic_list_b0_.end(),
LongTermPicNumAscCompare());
// RefPicList1 (8.2.4.2.4) [[1] [2] [3]], where:
// [1] shortterm ref pics with POC > curr_pic's POC sorted by ascending POC,
// [2] shortterm ref pics with POC < curr_pic's POC by descending POC,
// [3] longterm ref pics by ascending long_term_pic_num.
dpb_.GetShortTermRefPicsAppending(&ref_pic_list_b1_);
num_short_refs = ref_pic_list_b1_.size();
// First sort by descending POC.
std::sort(ref_pic_list_b1_.begin(), ref_pic_list_b1_.end(), POCDescCompare());
// Find first with POC < curr_pic's POC to get first element in [2]...
iter = std::upper_bound(ref_pic_list_b1_.begin(), ref_pic_list_b1_.end(),
curr_picture_.get(), POCDescCompare());
// and sort [1] ascending.
std::sort(ref_pic_list_b1_.begin(), iter, POCAscCompare());
// Now add [3] and sort by ascending long_term_pic_num
dpb_.GetLongTermRefPicsAppending(&ref_pic_list_b1_);
std::sort(ref_pic_list_b1_.begin() + num_short_refs, ref_pic_list_b1_.end(),
LongTermPicNumAscCompare());
// If lists identical, swap first two entries in RefPicList1 (spec 8.2.4.2.3)
if (ref_pic_list_b1_.size() > 1 &&
base::ranges::equal(ref_pic_list_b0_, ref_pic_list_b1_))
std::swap(ref_pic_list_b1_[0], ref_pic_list_b1_[1]);
}
void H264Decoder::UpdatePicNums(int frame_num) {
for (auto& pic : dpb_) {
if (!pic->ref)
continue;
// 8.2.4.1. Assumes non-interlaced stream.
DCHECK_EQ(pic->field, H264Picture::FIELD_NONE);
if (pic->long_term) {
pic->long_term_pic_num = pic->long_term_frame_idx;
} else {
if (pic->frame_num > frame_num)
pic->frame_num_wrap = pic->frame_num - max_frame_num_;
else
pic->frame_num_wrap = pic->frame_num;
pic->pic_num = pic->frame_num_wrap;
}
}
}
bool H264Decoder::InitNonexistingPicture(scoped_refptr<H264Picture> pic,
int frame_num,
bool ref) {
pic->nonexisting = true;
pic->nal_ref_idc = 1;
pic->frame_num = pic->pic_num = frame_num;
pic->adaptive_ref_pic_marking_mode_flag = false;
pic->ref = ref;
pic->long_term_reference_flag = false;
pic->field = H264Picture::FIELD_NONE;
return CalculatePicOrderCounts(pic);
}
bool H264Decoder::HandleFrameNumGap(int frame_num) {
const H264SPS* sps = parser_->GetSPS(curr_sps_id_);
if (!sps)
return false;
if (!sps->gaps_in_frame_num_value_allowed_flag) {
VLOG(1) << "Invalid frame_num: " << frame_num;
// TODO(b:129119729, b:146914440): Youtube android app sometimes sends an
// invalid frame number after a seek. The sequence goes like:
// Seek, SPS, PPS, IDR-frame, non-IDR, ... non-IDR with invalid number.
// The only way to work around this reliably is to ignore this error.
// Video playback is not affected, no artefacts are visible.
// return false;
}
VLOG(2) << "Handling frame_num gap: " << prev_ref_frame_num_ << "->"
<< frame_num;
// 7.4.3/7-23
int unused_short_term_frame_num = (prev_ref_frame_num_ + 1) % max_frame_num_;
while (unused_short_term_frame_num != frame_num) {
scoped_refptr<H264Picture> pic = va_wrapper_.CreatePicture(sps);
if (!InitNonexistingPicture(pic, unused_short_term_frame_num, true))
return false;
UpdatePicNums(unused_short_term_frame_num);
FinishPicture(pic);
unused_short_term_frame_num++;
unused_short_term_frame_num %= max_frame_num_;
}
return true;
}
bool H264Decoder::CalculatePicOrderCounts(scoped_refptr<H264Picture> pic) {
const H264SPS* sps = parser_->GetSPS(curr_sps_id_);
if (!sps)
return false;
switch (pic->pic_order_cnt_type) {
case 0: {
// See spec 8.2.1.1.
int prev_pic_order_cnt_msb, prev_pic_order_cnt_lsb;
if (pic->idr) {
prev_pic_order_cnt_msb = prev_pic_order_cnt_lsb = 0;
} else {
if (prev_ref_has_memmgmnt5_) {
if (prev_ref_field_ != H264Picture::FIELD_BOTTOM) {
prev_pic_order_cnt_msb = 0;
prev_pic_order_cnt_lsb = prev_ref_top_field_order_cnt_;
} else {
prev_pic_order_cnt_msb = 0;
prev_pic_order_cnt_lsb = 0;
}
} else {
prev_pic_order_cnt_msb = prev_ref_pic_order_cnt_msb_;
prev_pic_order_cnt_lsb = prev_ref_pic_order_cnt_lsb_;
}
}
int max_pic_order_cnt_lsb =
1 << (sps->log2_max_pic_order_cnt_lsb_minus4 + 4);
DCHECK_NE(max_pic_order_cnt_lsb, 0);
if ((pic->pic_order_cnt_lsb < prev_pic_order_cnt_lsb) &&
(prev_pic_order_cnt_lsb - pic->pic_order_cnt_lsb >=
max_pic_order_cnt_lsb / 2)) {
pic->pic_order_cnt_msb = prev_pic_order_cnt_msb + max_pic_order_cnt_lsb;
} else if ((pic->pic_order_cnt_lsb > prev_pic_order_cnt_lsb) &&
(pic->pic_order_cnt_lsb - prev_pic_order_cnt_lsb >
max_pic_order_cnt_lsb / 2)) {
pic->pic_order_cnt_msb = prev_pic_order_cnt_msb - max_pic_order_cnt_lsb;
} else {
pic->pic_order_cnt_msb = prev_pic_order_cnt_msb;
}
if (pic->field != H264Picture::FIELD_BOTTOM) {
pic->top_field_order_cnt =
pic->pic_order_cnt_msb + pic->pic_order_cnt_lsb;
}
if (pic->field != H264Picture::FIELD_TOP) {
if (pic->field == H264Picture::FIELD_NONE) {
pic->bottom_field_order_cnt =
pic->top_field_order_cnt + pic->delta_pic_order_cnt_bottom;
} else {
pic->bottom_field_order_cnt =
pic->pic_order_cnt_msb + pic->pic_order_cnt_lsb;
}
}
break;
}
case 1: {
// See spec 8.2.1.2.
if (prev_has_memmgmnt5_) {
prev_frame_num_offset_ = 0;
}
if (pic->idr) {
pic->frame_num_offset = 0;
} else if (prev_frame_num_ > pic->frame_num) {
pic->frame_num_offset = prev_frame_num_offset_ + max_frame_num_;
} else {
pic->frame_num_offset = prev_frame_num_offset_;
}
int abs_frame_num = 0;
if (sps->num_ref_frames_in_pic_order_cnt_cycle != 0) {
abs_frame_num = pic->frame_num_offset + pic->frame_num;
} else {
abs_frame_num = 0;
}
if (pic->nal_ref_idc == 0 && abs_frame_num > 0) {
--abs_frame_num;
}
int expected_pic_order_cnt = 0;
if (abs_frame_num > 0) {
if (sps->num_ref_frames_in_pic_order_cnt_cycle == 0) {
VLOG(1) << "Invalid num_ref_frames_in_pic_order_cnt_cycle "
<< "in stream";
return false;
}
int pic_order_cnt_cycle_cnt =
(abs_frame_num - 1) / sps->num_ref_frames_in_pic_order_cnt_cycle;
int frame_num_in_pic_order_cnt_cycle =
(abs_frame_num - 1) % sps->num_ref_frames_in_pic_order_cnt_cycle;
expected_pic_order_cnt = pic_order_cnt_cycle_cnt *
sps->expected_delta_per_pic_order_cnt_cycle;
// frame_num_in_pic_order_cnt_cycle is verified < 255 in parser
for (int i = 0; i <= frame_num_in_pic_order_cnt_cycle; ++i) {
expected_pic_order_cnt += sps->offset_for_ref_frame[i];
}
}
if (!pic->nal_ref_idc) {
expected_pic_order_cnt += sps->offset_for_non_ref_pic;
}
if (pic->field == H264Picture::FIELD_NONE) {
pic->top_field_order_cnt =
expected_pic_order_cnt + pic->delta_pic_order_cnt0;
pic->bottom_field_order_cnt = pic->top_field_order_cnt +
sps->offset_for_top_to_bottom_field +
pic->delta_pic_order_cnt1;
} else if (pic->field != H264Picture::FIELD_BOTTOM) {
pic->top_field_order_cnt =
expected_pic_order_cnt + pic->delta_pic_order_cnt0;
} else {
pic->bottom_field_order_cnt = expected_pic_order_cnt +
sps->offset_for_top_to_bottom_field +
pic->delta_pic_order_cnt0;
}
break;
}
case 2: {
// See spec 8.2.1.3.
if (prev_has_memmgmnt5_) {
prev_frame_num_offset_ = 0;
}
if (pic->idr) {
pic->frame_num_offset = 0;
} else if (prev_frame_num_ > pic->frame_num) {
pic->frame_num_offset = prev_frame_num_offset_ + max_frame_num_;
} else {
pic->frame_num_offset = prev_frame_num_offset_;
}
int temp_pic_order_cnt;
if (pic->idr) {
temp_pic_order_cnt = 0;
} else if (!pic->nal_ref_idc) {
temp_pic_order_cnt = 2 * (pic->frame_num_offset + pic->frame_num) - 1;
} else {
temp_pic_order_cnt = 2 * (pic->frame_num_offset + pic->frame_num);
}
if (pic->field == H264Picture::FIELD_NONE) {
pic->top_field_order_cnt = temp_pic_order_cnt;
pic->bottom_field_order_cnt = temp_pic_order_cnt;
} else if (pic->field == H264Picture::FIELD_BOTTOM) {
pic->bottom_field_order_cnt = temp_pic_order_cnt;
} else {
pic->top_field_order_cnt = temp_pic_order_cnt;
}
break;
}
default:
VLOG(1) << "Invalid pic_order_cnt_type: " << sps->pic_order_cnt_type;
return false;
}
switch (pic->field) {
case H264Picture::FIELD_NONE:
pic->pic_order_cnt =
std::min(pic->top_field_order_cnt, pic->bottom_field_order_cnt);
break;
case H264Picture::FIELD_TOP:
pic->pic_order_cnt = pic->top_field_order_cnt;
break;
case H264Picture::FIELD_BOTTOM:
pic->pic_order_cnt = pic->bottom_field_order_cnt;
break;
}
return true;
}
void H264Decoder::UpdateMaxNumReorderFrames(const H264SPS* sps) {
if (sps->vui_parameters_present_flag && sps->bitstream_restriction_flag) {
max_num_reorder_frames_ =
base::checked_cast<size_t>(sps->max_num_reorder_frames);
} else if (sps->constraint_set3_flag) {
// max_num_reorder_frames not present, infer from profile/constraints
// (see VUI semantics in spec).
switch (sps->profile_idc) {
case 44:
case 86:
case 100:
case 110:
case 122:
case 244:
max_num_reorder_frames_ = 0;
break;
default:
max_num_reorder_frames_ = dpb_.max_num_pics();
break;
}
} else {
max_num_reorder_frames_ = dpb_.max_num_pics();
}
}
bool H264Decoder::ModifyReferencePicLists(const H264SliceHeader* slice_hdr,
H264Picture::Vector* ref_pic_list0,
H264Picture::Vector* ref_pic_list1) {
ref_pic_list0->clear();
ref_pic_list1->clear();
// Fill reference picture lists for B and S/SP slices.
if (slice_hdr->IsPSlice() || slice_hdr->IsSPSlice()) {
*ref_pic_list0 = ref_pic_list_p0_;
return ModifyReferencePicList(slice_hdr, 0, ref_pic_list0);
} else if (slice_hdr->IsBSlice()) {
*ref_pic_list0 = ref_pic_list_b0_;
*ref_pic_list1 = ref_pic_list_b1_;
return ModifyReferencePicList(slice_hdr, 0, ref_pic_list0) &&
ModifyReferencePicList(slice_hdr, 1, ref_pic_list1);
}
return true;
}
bool H264Decoder::ModifyReferencePicList(const H264SliceHeader* slice_hdr,
int list,
H264Picture::Vector* ref_pic_listx) {
bool ref_pic_list_modification_flag_lX;
int num_ref_idx_lX_active_minus1;
const H264ModificationOfPicNum* list_mod;
// This can process either ref_pic_list0 or ref_pic_list1, depending on
// the list argument. Set up pointers to proper list to be processed here.
if (list == 0) {
ref_pic_list_modification_flag_lX =
slice_hdr->ref_pic_list_modification_flag_l0;
num_ref_idx_lX_active_minus1 = slice_hdr->num_ref_idx_l0_active_minus1;
list_mod = slice_hdr->ref_list_l0_modifications;
} else {
ref_pic_list_modification_flag_lX =
slice_hdr->ref_pic_list_modification_flag_l1;
num_ref_idx_lX_active_minus1 = slice_hdr->num_ref_idx_l1_active_minus1;
list_mod = slice_hdr->ref_list_l1_modifications;
}
// Resize the list to the size requested in the slice header.
// Note that per 8.2.4.2 it's possible for num_ref_idx_lX_active_minus1 to
// indicate there should be more ref pics on list than we constructed.
// Those superfluous ones should be treated as non-reference and will be
// initialized to nullptr, which must be handled by clients.
DCHECK_GE(num_ref_idx_lX_active_minus1, 0);
size_t original_size = ref_pic_listx->size();
ref_pic_listx->resize(num_ref_idx_lX_active_minus1 + 1);
for (int i = original_size; i < num_ref_idx_lX_active_minus1 + 1; i++) {
scoped_refptr<H264Picture> nonref_pic =
base::WrapRefCounted(new H264Picture(nullptr));
LOG_ASSERT(InitNonexistingPicture(nonref_pic, 0, false));
(*ref_pic_listx)[i] = nonref_pic;
}
if (!ref_pic_list_modification_flag_lX)
return true;
// Spec 8.2.4.3:
// Reorder pictures on the list in a way specified in the stream.
int pic_num_lx_pred = curr_picture_->pic_num;
int ref_idx_lx = 0;
int pic_num_lx_no_wrap;
int pic_num_lx;
bool done = false;
scoped_refptr<H264Picture> pic;
for (int i = 0; i < H264SliceHeader::kRefListModSize && !done; ++i) {
switch (list_mod->modification_of_pic_nums_idc) {
case 0:
case 1:
// Modify short reference picture position.
if (list_mod->modification_of_pic_nums_idc == 0) {
// Subtract given value from predicted PicNum.
pic_num_lx_no_wrap =
pic_num_lx_pred -
(static_cast<int>(list_mod->abs_diff_pic_num_minus1) + 1);
// Wrap around max_pic_num_ if it becomes < 0 as result
// of subtraction.
if (pic_num_lx_no_wrap < 0)
pic_num_lx_no_wrap += max_pic_num_;
} else {
// Add given value to predicted PicNum.
pic_num_lx_no_wrap =
pic_num_lx_pred +
(static_cast<int>(list_mod->abs_diff_pic_num_minus1) + 1);
// Wrap around max_pic_num_ if it becomes >= max_pic_num_ as result
// of the addition.
if (pic_num_lx_no_wrap >= max_pic_num_)
pic_num_lx_no_wrap -= max_pic_num_;
}
// For use in next iteration.
pic_num_lx_pred = pic_num_lx_no_wrap;
if (pic_num_lx_no_wrap > curr_picture_->pic_num)
pic_num_lx = pic_num_lx_no_wrap - max_pic_num_;
else
pic_num_lx = pic_num_lx_no_wrap;
DCHECK_LT(num_ref_idx_lX_active_minus1 + 1,
H264SliceHeader::kRefListModSize);
pic = dpb_.GetShortRefPicByPicNum(pic_num_lx);
if (!pic) {
VLOG(1) << "Malformed stream, no pic num " << pic_num_lx;
return false;
}
if (ref_idx_lx > num_ref_idx_lX_active_minus1) {
VLOG(1) << "Bounds mismatch: expected " << ref_idx_lx
<< " <= " << num_ref_idx_lX_active_minus1;
return false;
}
ShiftRightAndInsert(ref_pic_listx, ref_idx_lx,
num_ref_idx_lX_active_minus1, pic);
ref_idx_lx++;
for (int src = ref_idx_lx, dst = ref_idx_lx;
src <= num_ref_idx_lX_active_minus1 + 1; ++src) {
auto* src_pic = (*ref_pic_listx)[src].get();
int src_pic_num_lx = src_pic ? PicNumF(*src_pic) : -1;
if (src_pic_num_lx != pic_num_lx)
(*ref_pic_listx)[dst++] = (*ref_pic_listx)[src];
}
break;
case 2:
// Modify long term reference picture position.
DCHECK_LT(num_ref_idx_lX_active_minus1 + 1,
H264SliceHeader::kRefListModSize);
pic = dpb_.GetLongRefPicByLongTermPicNum(list_mod->long_term_pic_num);
if (!pic) {
VLOG(1) << "Malformed stream, no pic num "
<< list_mod->long_term_pic_num;
return false;
}
ShiftRightAndInsert(ref_pic_listx, ref_idx_lx,
num_ref_idx_lX_active_minus1, pic);
ref_idx_lx++;
for (int src = ref_idx_lx, dst = ref_idx_lx;
src <= num_ref_idx_lX_active_minus1 + 1; ++src) {
if ((*ref_pic_listx)[src] &&
LongTermPicNumF(*(*ref_pic_listx)[src]) !=
static_cast<int>(list_mod->long_term_pic_num))
(*ref_pic_listx)[dst++] = (*ref_pic_listx)[src];
}
break;
case 3:
// End of modification list.
done = true;
break;
default:
// May be recoverable.
VLOG(1) << "Invalid modification_of_pic_nums_idc="
<< list_mod->modification_of_pic_nums_idc << " in position "
<< i;
break;
}
++list_mod;
}
// Per NOTE 2 in 8.2.4.3.2, the ref_pic_listx size in the above loop is
// temporarily made one element longer than the required final list.
// Resize the list back to its required size.
ref_pic_listx->resize(num_ref_idx_lX_active_minus1 + 1);
return true;
}
// This method ensures that DPB does not overflow, either by removing
// reference pictures as specified in the stream, or using a sliding window
// procedure to remove the oldest one.
// It also performs marking and unmarking pictures as reference.
// See spac 8.2.5.1.
bool H264Decoder::ReferencePictureMarking(scoped_refptr<H264Picture> pic) {
// If the current picture is an IDR, all reference pictures are unmarked.
if (pic->idr) {
dpb_.MarkAllUnusedForRef();
if (pic->long_term_reference_flag) {
pic->long_term = true;
pic->long_term_frame_idx = 0;
max_long_term_frame_idx_ = 0;
} else {
pic->long_term = false;
max_long_term_frame_idx_ = -1;
}
return true;
}
// Not an IDR. If the stream contains instructions on how to discard pictures
// from DPB and how to mark/unmark existing reference pictures, do so.
// Otherwise, fall back to default sliding window process.
if (pic->adaptive_ref_pic_marking_mode_flag) {
DCHECK(!pic->nonexisting);
return HandleMemoryManagementOps(pic);
} else {
return SlidingWindowPictureMarking();
}
}
bool H264Decoder::HandleMemoryManagementOps(scoped_refptr<H264Picture> pic) {
// 8.2.5.4
for (size_t i = 0; i < std::size(pic->ref_pic_marking); ++i) {
// Code below does not support interlaced stream (per-field pictures).
H264DecRefPicMarking* ref_pic_marking = &pic->ref_pic_marking[i];
scoped_refptr<H264Picture> to_mark;
int pic_num_x;
switch (ref_pic_marking->memory_mgmnt_control_operation) {
case 0:
// Normal end of operations' specification.
return true;
case 1:
// Mark a short term reference picture as unused so it can be removed
// if outputted.
pic_num_x =
pic->pic_num - (ref_pic_marking->difference_of_pic_nums_minus1 + 1);
to_mark = dpb_.GetShortRefPicByPicNum(pic_num_x);
if (to_mark) {
to_mark->ref = false;
} else {
VLOG(1) << "Invalid short ref pic num to unmark";
return false;
}
break;
case 2:
// Mark a long term reference picture as unused so it can be removed
// if outputted.
to_mark = dpb_.GetLongRefPicByLongTermPicNum(
ref_pic_marking->long_term_pic_num);
if (to_mark) {
to_mark->ref = false;
} else {
VLOG(1) << "Invalid long term ref pic num to unmark";
return false;
}
break;
case 3:
// Mark a short term reference picture as long term reference.
pic_num_x =
pic->pic_num - (ref_pic_marking->difference_of_pic_nums_minus1 + 1);
to_mark = dpb_.GetShortRefPicByPicNum(pic_num_x);
if (to_mark) {
DCHECK(to_mark->ref && !to_mark->long_term);
scoped_refptr<H264Picture> long_term_mark =
dpb_.GetLongRefPicByLongTermIdx(
ref_pic_marking->long_term_frame_idx);
if (long_term_mark) {
long_term_mark->ref = false;
}
to_mark->long_term = true;
to_mark->long_term_frame_idx = ref_pic_marking->long_term_frame_idx;
} else {
VLOG(1) << "Invalid short term ref pic num to mark as long ref";
return false;
}
break;
case 4: {
// Unmark all reference pictures with long_term_frame_idx over new max.
max_long_term_frame_idx_ =
ref_pic_marking->max_long_term_frame_idx_plus1 - 1;
H264Picture::Vector long_terms;
dpb_.GetLongTermRefPicsAppending(&long_terms);
for (auto long_term_pic : long_terms) {
DCHECK(long_term_pic->ref && long_term_pic->long_term);
// Ok to cast, max_long_term_frame_idx is much smaller than 16bit.
if (long_term_pic->long_term_frame_idx >
static_cast<int>(max_long_term_frame_idx_))
long_term_pic->ref = false;
}
break;
}
case 5:
// Unmark all reference pictures.
dpb_.MarkAllUnusedForRef();
max_long_term_frame_idx_ = -1;
pic->mem_mgmt_5 = true;
break;
case 6: {
// Replace long term reference pictures with current picture.
// First unmark if any existing with this long_term_frame_idx...
H264Picture::Vector long_terms;
dpb_.GetLongTermRefPicsAppending(&long_terms);
for (auto long_term_pic : long_terms) {
DCHECK(long_term_pic->ref && long_term_pic->long_term);
// Ok to cast, long_term_frame_idx is much smaller than 16bit.
if (long_term_pic->long_term_frame_idx ==
static_cast<int>(ref_pic_marking->long_term_frame_idx))
long_term_pic->ref = false;
}
// and mark the current one instead.
pic->ref = true;
pic->long_term = true;
pic->long_term_frame_idx = ref_pic_marking->long_term_frame_idx;
break;
}
default:
// Would indicate a bug in parser.
NOTREACHED();
}
}
return true;
}
bool H264Decoder::SlidingWindowPictureMarking() {
const H264SPS* sps = parser_->GetSPS(curr_sps_id_);
if (!sps)
return false;
// 8.2.5.3. Ensure the DPB doesn't overflow by discarding the oldest picture.
int num_ref_pics = dpb_.CountRefPics();
DCHECK_LE(num_ref_pics, std::max<int>(sps->max_num_ref_frames, 1));
if (num_ref_pics == std::max<int>(sps->max_num_ref_frames, 1)) {
// Max number of reference pics reached, need to remove one of the short
// term ones. Find smallest frame_num_wrap short reference picture and mark
// it as unused.
scoped_refptr<H264Picture> to_unmark =
dpb_.GetLowestFrameNumWrapShortRefPic();
if (!to_unmark) {
VLOG(1) << "Couldn't find a short ref picture to unmark";
return false;
}
to_unmark->ref = false;
}
return true;
}
// See 8.2.4
int H264Decoder::PicNumF(const H264Picture& pic) const {
if (!pic.long_term)
return pic.pic_num;
else
return max_pic_num_;
}
// See 8.2.4
int H264Decoder::LongTermPicNumF(const H264Picture& pic) const {
if (pic.ref && pic.long_term)
return pic.long_term_pic_num;
else
return 2 * (max_long_term_frame_idx_ + 1);
}
// Shift elements on the |v| starting from |from| to |to|, inclusive,
// one position to the right and insert pic at |from|.
void H264Decoder::ShiftRightAndInsert(H264Picture::Vector* v,
int from,
int to,
scoped_refptr<H264Picture> pic) {
// Security checks, do not disable in Debug mode.
CHECK(from <= to);
CHECK(to <= std::numeric_limits<int>::max() - 2);
// Additional checks. Debug mode ok.
DCHECK(v);
DCHECK(pic);
DCHECK((to + 1 == static_cast<int>(v->size())) ||
(to + 2 == static_cast<int>(v->size())));
v->resize(to + 2);
for (int i = to + 1; i > from; --i)
(*v)[i] = (*v)[i - 1];
(*v)[from] = std::move(pic);
}
uint32_t H264Decoder::H264LevelToMaxDpbMbs(uint8_t level) {
switch (level) {
case H264SPS::kLevelIDC1p0:
case H264SPS::kLevelIDC1B:
return 396;
case H264SPS::kLevelIDC1p1:
return 900;
case H264SPS::kLevelIDC1p2:
case H264SPS::kLevelIDC1p3:
case H264SPS::kLevelIDC2p0:
return 2376;
case H264SPS::kLevelIDC2p1:
return 4752;
case H264SPS::kLevelIDC2p2:
case H264SPS::kLevelIDC3p0:
return 8100;
case H264SPS::kLevelIDC3p1:
return 18000;
case H264SPS::kLevelIDC3p2:
return 20480;
case H264SPS::kLevelIDC4p0:
case H264SPS::kLevelIDC4p1:
return 32768;
case H264SPS::kLevelIDC4p2:
return 34816;
case H264SPS::kLevelIDC5p0:
return 110400;
case H264SPS::kLevelIDC5p1:
case H264SPS::kLevelIDC5p2:
return 184320;
case H264SPS::kLevelIDC6p1:
case H264SPS::kLevelIDC6p2:
return 696320;
default:
return 0;
}
}
void H264Decoder::FlushDPB() {
H264Picture::Vector not_outputted_vec;
dpb_.GetNotOutputtedPicsAppending(&not_outputted_vec);
std::sort(not_outputted_vec.begin(), not_outputted_vec.end(),
POCDescCompare());
while (!not_outputted_vec.empty()) {
output_queue.push(not_outputted_vec.back());
not_outputted_vec.back()->outputted = true;
not_outputted_vec.pop_back();
}
dpb_.Clear();
}
} // namespace media::vaapi_test