| // Copyright 2015 The Chromium Authors. All rights reserved. |
| // Use of this source code is governed by a BSD-style license that can be |
| // found in the LICENSE file. |
| |
| #include "media/gpu/vp9_decoder.h" |
| |
| #include <memory> |
| |
| #include "base/bind.h" |
| #include "base/feature_list.h" |
| #include "base/logging.h" |
| #include "build/build_config.h" |
| #include "build/chromeos_buildflags.h" |
| #include "media/base/limits.h" |
| #include "media/base/media_switches.h" |
| #include "media/gpu/vp9_decoder.h" |
| |
| namespace media { |
| |
| namespace { |
| std::vector<uint32_t> GetSpatialLayerFrameSize( |
| const DecoderBuffer& decoder_buffer) { |
| #if defined(ARCH_CPU_X86_FAMILY) && BUILDFLAG(IS_CHROMEOS_ASH) |
| const uint32_t* cue_data = |
| reinterpret_cast<const uint32_t*>(decoder_buffer.side_data()); |
| if (!cue_data) { |
| return {}; |
| } |
| if (!base::FeatureList::IsEnabled(media::kVaapiVp9kSVCHWDecoding)) { |
| DLOG(ERROR) << "Vp9Parser doesn't support parsing SVC stream"; |
| return {}; |
| } |
| |
| size_t num_of_layers = decoder_buffer.side_data_size() / sizeof(uint32_t); |
| if (num_of_layers > 3u) { |
| DLOG(WARNING) << "The maximum number of spatial layers in VP9 is three"; |
| return {}; |
| } |
| return std::vector<uint32_t>(cue_data, cue_data + num_of_layers); |
| #else |
| return {}; |
| #endif // defined(ARCH_CPU_X86_FAMILY) && BUILDFLAG(IS_CHROMEOS_ASH) |
| } |
| |
| VideoCodecProfile VP9ProfileToVideoCodecProfile(uint8_t profile) { |
| switch (profile) { |
| case 0: |
| return VP9PROFILE_PROFILE0; |
| case 1: |
| return VP9PROFILE_PROFILE1; |
| case 2: |
| return VP9PROFILE_PROFILE2; |
| case 3: |
| return VP9PROFILE_PROFILE3; |
| default: |
| return VIDEO_CODEC_PROFILE_UNKNOWN; |
| } |
| } |
| |
| bool IsValidBitDepth(uint8_t bit_depth, VideoCodecProfile profile) { |
| // Spec 7.2. |
| switch (profile) { |
| case VP9PROFILE_PROFILE0: |
| case VP9PROFILE_PROFILE1: |
| return bit_depth == 8u; |
| case VP9PROFILE_PROFILE2: |
| case VP9PROFILE_PROFILE3: |
| return bit_depth == 10u || bit_depth == 12u; |
| default: |
| NOTREACHED(); |
| return false; |
| } |
| } |
| |
| bool IsYUV420Sequence(const Vp9FrameHeader& frame_header) { |
| // Spec 7.2.2 |
| return frame_header.subsampling_x == 1u && frame_header.subsampling_y == 1u; |
| } |
| } // namespace |
| |
| VP9Decoder::VP9Accelerator::VP9Accelerator() {} |
| |
| VP9Decoder::VP9Accelerator::~VP9Accelerator() {} |
| |
| VP9Decoder::VP9Decoder(std::unique_ptr<VP9Accelerator> accelerator, |
| VideoCodecProfile profile, |
| const VideoColorSpace& container_color_space) |
| : state_(kNeedStreamMetadata), |
| container_color_space_(container_color_space), |
| // TODO(hiroh): Set profile to UNKNOWN. |
| profile_(profile), |
| accelerator_(std::move(accelerator)), |
| parser_(accelerator_->IsFrameContextRequired()) {} |
| |
| VP9Decoder::~VP9Decoder() = default; |
| |
| void VP9Decoder::SetStream(int32_t id, const DecoderBuffer& decoder_buffer) { |
| const uint8_t* ptr = decoder_buffer.data(); |
| const size_t size = decoder_buffer.data_size(); |
| const DecryptConfig* decrypt_config = decoder_buffer.decrypt_config(); |
| |
| DCHECK(ptr); |
| DCHECK(size); |
| DVLOG(4) << "New input stream id: " << id << " at: " << (void*)ptr |
| << " size: " << size; |
| stream_id_ = id; |
| if (decrypt_config) { |
| parser_.SetStream(ptr, size, GetSpatialLayerFrameSize(decoder_buffer), |
| decrypt_config->Clone()); |
| } else { |
| parser_.SetStream(ptr, size, GetSpatialLayerFrameSize(decoder_buffer), |
| nullptr); |
| } |
| } |
| |
| bool VP9Decoder::Flush() { |
| DVLOG(2) << "Decoder flush"; |
| Reset(); |
| return true; |
| } |
| |
| void VP9Decoder::Reset() { |
| curr_frame_hdr_ = nullptr; |
| decrypt_config_.reset(); |
| pending_pic_.reset(); |
| |
| ref_frames_.Clear(); |
| |
| parser_.Reset(); |
| |
| if (state_ == kDecoding) { |
| state_ = kAfterReset; |
| } |
| } |
| |
| VP9Decoder::DecodeResult VP9Decoder::Decode() { |
| while (true) { |
| // If we have a pending picture to decode, try that first. |
| if (pending_pic_) { |
| VP9Accelerator::Status status = |
| DecodeAndOutputPicture(std::move(pending_pic_)); |
| if (status == VP9Accelerator::Status::kFail) { |
| SetError(); |
| return kDecodeError; |
| } |
| if (status == VP9Accelerator::Status::kTryAgain) |
| return kTryAgain; |
| } |
| |
| // Read a new frame header if one is not awaiting decoding already. |
| if (!curr_frame_hdr_) { |
| gfx::Size allocate_size; |
| std::unique_ptr<Vp9FrameHeader> hdr(new Vp9FrameHeader()); |
| Vp9Parser::Result res = |
| parser_.ParseNextFrame(hdr.get(), &allocate_size, &decrypt_config_); |
| switch (res) { |
| case Vp9Parser::kOk: |
| curr_frame_hdr_ = std::move(hdr); |
| curr_frame_size_ = allocate_size; |
| break; |
| |
| case Vp9Parser::kEOStream: |
| return kRanOutOfStreamData; |
| |
| case Vp9Parser::kInvalidStream: |
| DVLOG(1) << "Error parsing stream"; |
| SetError(); |
| return kDecodeError; |
| |
| case Vp9Parser::kAwaitingRefresh: |
| DVLOG(4) << "Awaiting context update"; |
| return kNeedContextUpdate; |
| } |
| } |
| |
| if (state_ != kDecoding) { |
| // Not kDecoding, so we need a resume point (a keyframe), as we are after |
| // reset or at the beginning of the stream. Drop anything that is not |
| // a keyframe in such case, and continue looking for a keyframe. |
| // Only exception is when the stream/sequence starts with an Intra only |
| // frame. |
| if (curr_frame_hdr_->IsKeyframe() || |
| (curr_frame_hdr_->IsIntra() && pic_size_.IsEmpty())) { |
| state_ = kDecoding; |
| } else { |
| curr_frame_hdr_.reset(); |
| decrypt_config_.reset(); |
| continue; |
| } |
| } |
| |
| if (curr_frame_hdr_->show_existing_frame) { |
| // This frame header only instructs us to display one of the |
| // previously-decoded frames, but has no frame data otherwise. Display |
| // and continue decoding subsequent frames. |
| size_t frame_to_show = curr_frame_hdr_->frame_to_show_map_idx; |
| if (frame_to_show >= kVp9NumRefFrames || |
| !ref_frames_.GetFrame(frame_to_show)) { |
| DVLOG(1) << "Request to show an invalid frame"; |
| SetError(); |
| return kDecodeError; |
| } |
| |
| // Duplicate the VP9Picture and set the current bitstream id to keep the |
| // correct timestamp. |
| scoped_refptr<VP9Picture> pic = |
| ref_frames_.GetFrame(frame_to_show)->Duplicate(); |
| if (pic == nullptr) { |
| DVLOG(1) << "Failed to duplicate the VP9Picture."; |
| SetError(); |
| return kDecodeError; |
| } |
| pic->set_bitstream_id(stream_id_); |
| if (!accelerator_->OutputPicture(std::move(pic))) { |
| SetError(); |
| return kDecodeError; |
| } |
| |
| curr_frame_hdr_.reset(); |
| decrypt_config_.reset(); |
| continue; |
| } |
| |
| gfx::Size new_pic_size = curr_frame_size_; |
| gfx::Rect new_render_rect(curr_frame_hdr_->render_width, |
| curr_frame_hdr_->render_height); |
| // For safety, check the validity of render size or leave it as pic size. |
| if (!gfx::Rect(new_pic_size).Contains(new_render_rect)) { |
| DVLOG(1) << "Render size exceeds picture size. render size: " |
| << new_render_rect.ToString() |
| << ", picture size: " << new_pic_size.ToString(); |
| new_render_rect = gfx::Rect(new_pic_size); |
| } |
| VideoCodecProfile new_profile = |
| VP9ProfileToVideoCodecProfile(curr_frame_hdr_->profile); |
| if (new_profile == VIDEO_CODEC_PROFILE_UNKNOWN) { |
| VLOG(1) << "Invalid profile: " << curr_frame_hdr_->profile; |
| return kDecodeError; |
| } |
| if (!IsValidBitDepth(curr_frame_hdr_->bit_depth, new_profile)) { |
| DVLOG(1) << "Invalid bit depth=" |
| << base::strict_cast<int>(curr_frame_hdr_->bit_depth) |
| << ", profile=" << GetProfileName(new_profile); |
| return kDecodeError; |
| } |
| if (!IsYUV420Sequence(*curr_frame_hdr_)) { |
| DVLOG(1) << "Only YUV 4:2:0 is supported"; |
| return kDecodeError; |
| } |
| |
| DCHECK(!new_pic_size.IsEmpty()); |
| if (new_pic_size != pic_size_ || new_profile != profile_ || |
| curr_frame_hdr_->bit_depth != bit_depth_) { |
| DVLOG(1) << "New profile: " << GetProfileName(new_profile) |
| << ", New resolution: " << new_pic_size.ToString() |
| << ", New bit depth: " |
| << base::strict_cast<int>(curr_frame_hdr_->bit_depth); |
| |
| if (!curr_frame_hdr_->IsKeyframe() && |
| !(curr_frame_hdr_->IsIntra() && pic_size_.IsEmpty())) { |
| // TODO(posciak): This is doable, but requires a few modifications to |
| // VDA implementations to allow multiple picture buffer sets in flight. |
| // http://crbug.com/832264 |
| DVLOG(1) << "Resolution change currently supported for keyframes and " |
| "sequence begins with Intra only when there is no prior " |
| "frames in the context"; |
| if (++size_change_failure_counter_ > kVPxMaxNumOfSizeChangeFailures) { |
| SetError(); |
| return kDecodeError; |
| } |
| |
| curr_frame_hdr_.reset(); |
| decrypt_config_.reset(); |
| return kRanOutOfStreamData; |
| } |
| |
| // TODO(posciak): This requires us to be on a keyframe (see above) and is |
| // required, because VDA clients expect all surfaces to be returned before |
| // they can cycle surface sets after receiving kConfigChange. |
| // This is only an implementation detail of VDAs and can be improved. |
| ref_frames_.Clear(); |
| |
| pic_size_ = new_pic_size; |
| visible_rect_ = new_render_rect; |
| profile_ = new_profile; |
| bit_depth_ = curr_frame_hdr_->bit_depth; |
| size_change_failure_counter_ = 0; |
| return kConfigChange; |
| } |
| |
| scoped_refptr<VP9Picture> pic = accelerator_->CreateVP9Picture(); |
| if (!pic) { |
| return kRanOutOfSurfaces; |
| } |
| DVLOG(2) << "Render resolution: " << new_render_rect.ToString(); |
| |
| pic->set_visible_rect(new_render_rect); |
| pic->set_bitstream_id(stream_id_); |
| |
| pic->set_decrypt_config(std::move(decrypt_config_)); |
| |
| // For VP9, container color spaces override video stream color spaces. |
| if (container_color_space_.IsSpecified()) |
| pic->set_colorspace(container_color_space_); |
| else if (curr_frame_hdr_) |
| pic->set_colorspace(curr_frame_hdr_->GetColorSpace()); |
| |
| pic->frame_hdr = std::move(curr_frame_hdr_); |
| |
| VP9Accelerator::Status status = DecodeAndOutputPicture(std::move(pic)); |
| if (status == VP9Accelerator::Status::kFail) { |
| SetError(); |
| return kDecodeError; |
| } |
| if (status == VP9Accelerator::Status::kTryAgain) |
| return kTryAgain; |
| } |
| } |
| |
| void VP9Decoder::UpdateFrameContext( |
| scoped_refptr<VP9Picture> pic, |
| Vp9Parser::ContextRefreshCallback context_refresh_cb) { |
| DCHECK(context_refresh_cb); |
| Vp9FrameContext frame_ctx; |
| memset(&frame_ctx, 0, sizeof(frame_ctx)); |
| |
| if (!accelerator_->GetFrameContext(std::move(pic), &frame_ctx)) { |
| SetError(); |
| return; |
| } |
| |
| std::move(context_refresh_cb).Run(frame_ctx); |
| } |
| |
| VP9Decoder::VP9Accelerator::Status VP9Decoder::DecodeAndOutputPicture( |
| scoped_refptr<VP9Picture> pic) { |
| DCHECK(!pic_size_.IsEmpty()); |
| DCHECK(pic->frame_hdr); |
| |
| base::OnceClosure done_cb; |
| Vp9Parser::ContextRefreshCallback context_refresh_cb = |
| parser_.GetContextRefreshCb(pic->frame_hdr->frame_context_idx); |
| if (context_refresh_cb) { |
| done_cb = |
| base::BindOnce(&VP9Decoder::UpdateFrameContext, base::Unretained(this), |
| pic, std::move(context_refresh_cb)); |
| } |
| |
| const Vp9Parser::Context& context = parser_.context(); |
| VP9Accelerator::Status status = accelerator_->SubmitDecode( |
| pic, context.segmentation(), context.loop_filter(), ref_frames_, |
| std::move(done_cb)); |
| if (status != VP9Accelerator::Status::kOk) { |
| if (status == VP9Accelerator::Status::kTryAgain) |
| pending_pic_ = std::move(pic); |
| return status; |
| } |
| |
| if (pic->frame_hdr->show_frame) { |
| if (!accelerator_->OutputPicture(pic)) |
| return VP9Accelerator::Status::kFail; |
| } |
| |
| ref_frames_.Refresh(std::move(pic)); |
| return status; |
| } |
| |
| void VP9Decoder::SetError() { |
| Reset(); |
| state_ = kError; |
| } |
| |
| gfx::Size VP9Decoder::GetPicSize() const { |
| return pic_size_; |
| } |
| |
| gfx::Rect VP9Decoder::GetVisibleRect() const { |
| return visible_rect_; |
| } |
| |
| VideoCodecProfile VP9Decoder::GetProfile() const { |
| return profile_; |
| } |
| |
| uint8_t VP9Decoder::GetBitDepth() const { |
| return bit_depth_; |
| } |
| |
| size_t VP9Decoder::GetRequiredNumOfPictures() const { |
| constexpr size_t kPicsInPipeline = limits::kMaxVideoFrames + 1; |
| return kPicsInPipeline + GetNumReferenceFrames(); |
| } |
| |
| size_t VP9Decoder::GetNumReferenceFrames() const { |
| // Maximum number of reference frames |
| return kVp9NumRefFrames; |
| } |
| |
| } // namespace media |