| // Copyright 2016 The Chromium Authors. All rights reserved. |
| // Use of this source code is governed by a BSD-style license that can be |
| // found in the LICENSE file. |
| |
| #include "media/remoting/proto_utils.h" |
| |
| #include <algorithm> |
| |
| #include "base/big_endian.h" |
| #include "base/logging.h" |
| #include "base/time/time.h" |
| #include "base/values.h" |
| #include "media/base/encryption_scheme.h" |
| #include "media/base/timestamp_constants.h" |
| #include "media/remoting/proto_enum_utils.h" |
| |
| namespace media { |
| namespace remoting { |
| |
| namespace { |
| |
| constexpr size_t kPayloadVersionFieldSize = sizeof(uint8_t); |
| constexpr size_t kProtoBufferHeaderSize = sizeof(uint16_t); |
| constexpr size_t kDataBufferHeaderSize = sizeof(uint32_t); |
| |
| scoped_refptr<DecoderBuffer> ConvertProtoToDecoderBuffer( |
| const openscreen::cast::DecoderBuffer& buffer_message, |
| scoped_refptr<DecoderBuffer> buffer) { |
| if (buffer_message.is_eos()) { |
| VLOG(1) << "EOS data"; |
| return DecoderBuffer::CreateEOSBuffer(); |
| } |
| |
| if (buffer_message.has_timestamp_usec()) { |
| buffer->set_timestamp(base::Microseconds(buffer_message.timestamp_usec())); |
| } |
| |
| if (buffer_message.has_duration_usec()) { |
| buffer->set_duration(base::Microseconds(buffer_message.duration_usec())); |
| } |
| VLOG(3) << "timestamp:" << buffer_message.timestamp_usec() |
| << " duration:" << buffer_message.duration_usec(); |
| |
| if (buffer_message.has_is_key_frame()) |
| buffer->set_is_key_frame(buffer_message.is_key_frame()); |
| |
| bool has_discard = false; |
| base::TimeDelta front_discard; |
| if (buffer_message.has_front_discard_usec()) { |
| has_discard = true; |
| front_discard = base::Microseconds(buffer_message.front_discard_usec()); |
| } |
| base::TimeDelta back_discard; |
| if (buffer_message.has_back_discard_usec()) { |
| has_discard = true; |
| back_discard = base::Microseconds(buffer_message.back_discard_usec()); |
| } |
| |
| if (has_discard) { |
| buffer->set_discard_padding( |
| DecoderBuffer::DiscardPadding(front_discard, back_discard)); |
| } |
| |
| if (buffer_message.has_side_data()) { |
| buffer->CopySideDataFrom( |
| reinterpret_cast<const uint8_t*>(buffer_message.side_data().data()), |
| buffer_message.side_data().size()); |
| } |
| |
| return buffer; |
| } |
| |
| void ConvertDecoderBufferToProto( |
| const DecoderBuffer& decoder_buffer, |
| openscreen::cast::DecoderBuffer* buffer_message) { |
| if (decoder_buffer.end_of_stream()) { |
| buffer_message->set_is_eos(true); |
| return; |
| } |
| |
| VLOG(3) << "timestamp:" << decoder_buffer.timestamp().InMicroseconds() |
| << " duration:" << decoder_buffer.duration().InMicroseconds(); |
| buffer_message->set_timestamp_usec( |
| decoder_buffer.timestamp().InMicroseconds()); |
| buffer_message->set_duration_usec(decoder_buffer.duration().InMicroseconds()); |
| buffer_message->set_is_key_frame(decoder_buffer.is_key_frame()); |
| |
| buffer_message->set_front_discard_usec( |
| decoder_buffer.discard_padding().first.InMicroseconds()); |
| buffer_message->set_back_discard_usec( |
| decoder_buffer.discard_padding().second.InMicroseconds()); |
| |
| if (decoder_buffer.side_data_size()) { |
| buffer_message->set_side_data(decoder_buffer.side_data(), |
| decoder_buffer.side_data_size()); |
| } |
| } |
| |
| } // namespace |
| |
| scoped_refptr<DecoderBuffer> ByteArrayToDecoderBuffer(const uint8_t* data, |
| uint32_t size) { |
| base::BigEndianReader reader(reinterpret_cast<const char*>(data), size); |
| uint8_t payload_version = 0; |
| uint16_t proto_size = 0; |
| openscreen::cast::DecoderBuffer segment; |
| uint32_t buffer_size = 0; |
| if (reader.ReadU8(&payload_version) && payload_version == 0 && |
| reader.ReadU16(&proto_size) && proto_size < reader.remaining() && |
| segment.ParseFromArray(reader.ptr(), proto_size) && |
| reader.Skip(proto_size) && reader.ReadU32(&buffer_size) && |
| buffer_size <= reader.remaining()) { |
| // Deserialize proto buffer. It passes the pre allocated DecoderBuffer into |
| // the function because the proto buffer may overwrite DecoderBuffer since |
| // it may be EOS buffer. |
| scoped_refptr<DecoderBuffer> decoder_buffer = ConvertProtoToDecoderBuffer( |
| segment, |
| DecoderBuffer::CopyFrom(reinterpret_cast<const uint8_t*>(reader.ptr()), |
| buffer_size)); |
| return decoder_buffer; |
| } |
| |
| return nullptr; |
| } |
| |
| std::vector<uint8_t> DecoderBufferToByteArray( |
| const DecoderBuffer& decoder_buffer) { |
| openscreen::cast::DecoderBuffer decoder_buffer_message; |
| ConvertDecoderBufferToProto(decoder_buffer, &decoder_buffer_message); |
| |
| size_t decoder_buffer_size = |
| decoder_buffer.end_of_stream() ? 0 : decoder_buffer.data_size(); |
| size_t size = kPayloadVersionFieldSize + kProtoBufferHeaderSize + |
| decoder_buffer_message.ByteSize() + kDataBufferHeaderSize + |
| decoder_buffer_size; |
| std::vector<uint8_t> buffer(size); |
| base::BigEndianWriter writer(reinterpret_cast<char*>(buffer.data()), |
| buffer.size()); |
| if (writer.WriteU8(0) && |
| writer.WriteU16( |
| static_cast<uint16_t>(decoder_buffer_message.GetCachedSize())) && |
| decoder_buffer_message.SerializeToArray( |
| writer.ptr(), decoder_buffer_message.GetCachedSize()) && |
| writer.Skip(decoder_buffer_message.GetCachedSize()) && |
| writer.WriteU32(decoder_buffer_size)) { |
| if (decoder_buffer_size) { |
| // DecoderBuffer frame data. |
| writer.WriteBytes(reinterpret_cast<const void*>(decoder_buffer.data()), |
| decoder_buffer.data_size()); |
| } |
| return buffer; |
| } |
| |
| NOTREACHED(); |
| // Reset buffer since serialization of the data failed. |
| buffer.clear(); |
| return buffer; |
| } |
| |
| void ConvertAudioDecoderConfigToProto( |
| const AudioDecoderConfig& audio_config, |
| openscreen::cast::AudioDecoderConfig* audio_message) { |
| DCHECK(audio_config.IsValidConfig()); |
| DCHECK(audio_message); |
| |
| audio_message->set_codec( |
| ToProtoAudioDecoderConfigCodec(audio_config.codec()).value()); |
| audio_message->set_sample_format( |
| ToProtoAudioDecoderConfigSampleFormat(audio_config.sample_format()) |
| .value()); |
| audio_message->set_channel_layout( |
| ToProtoAudioDecoderConfigChannelLayout(audio_config.channel_layout()) |
| .value()); |
| audio_message->set_samples_per_second(audio_config.samples_per_second()); |
| audio_message->set_seek_preroll_usec( |
| audio_config.seek_preroll().InMicroseconds()); |
| audio_message->set_codec_delay(audio_config.codec_delay()); |
| |
| if (!audio_config.extra_data().empty()) { |
| audio_message->set_extra_data(audio_config.extra_data().data(), |
| audio_config.extra_data().size()); |
| } |
| } |
| |
| bool ConvertProtoToAudioDecoderConfig( |
| const openscreen::cast::AudioDecoderConfig& audio_message, |
| AudioDecoderConfig* audio_config) { |
| DCHECK(audio_config); |
| audio_config->Initialize( |
| ToMediaAudioCodec(audio_message.codec()).value(), |
| ToMediaSampleFormat(audio_message.sample_format()).value(), |
| ToMediaChannelLayout(audio_message.channel_layout()).value(), |
| audio_message.samples_per_second(), |
| std::vector<uint8_t>(audio_message.extra_data().begin(), |
| audio_message.extra_data().end()), |
| EncryptionScheme::kUnencrypted, |
| base::Microseconds(audio_message.seek_preroll_usec()), |
| audio_message.codec_delay()); |
| return audio_config->IsValidConfig(); |
| } |
| |
| void ConvertVideoDecoderConfigToProto( |
| const VideoDecoderConfig& video_config, |
| openscreen::cast::VideoDecoderConfig* video_message) { |
| DCHECK(video_config.IsValidConfig()); |
| DCHECK(video_message); |
| |
| video_message->set_codec( |
| ToProtoVideoDecoderConfigCodec(video_config.codec()).value()); |
| video_message->set_profile( |
| ToProtoVideoDecoderConfigProfile(video_config.profile()).value()); |
| // TODO(dalecurtis): Remove |format| it's now unused. |
| video_message->set_format( |
| video_config.alpha_mode() == VideoDecoderConfig::AlphaMode::kHasAlpha |
| ? openscreen::cast::VideoDecoderConfig::PIXEL_FORMAT_I420A |
| : openscreen::cast::VideoDecoderConfig::PIXEL_FORMAT_I420); |
| |
| // TODO(hubbe): Update proto to use color_space_info() |
| if (video_config.color_space_info() == VideoColorSpace::JPEG()) { |
| video_message->set_color_space( |
| openscreen::cast::VideoDecoderConfig::COLOR_SPACE_JPEG); |
| } else if (video_config.color_space_info() == VideoColorSpace::REC709()) { |
| video_message->set_color_space( |
| openscreen::cast::VideoDecoderConfig::COLOR_SPACE_HD_REC709); |
| } else if (video_config.color_space_info() == VideoColorSpace::REC601()) { |
| video_message->set_color_space( |
| openscreen::cast::VideoDecoderConfig::COLOR_SPACE_SD_REC601); |
| } else { |
| video_message->set_color_space( |
| openscreen::cast::VideoDecoderConfig::COLOR_SPACE_SD_REC601); |
| } |
| |
| openscreen::cast::Size* coded_size_message = |
| video_message->mutable_coded_size(); |
| coded_size_message->set_width(video_config.coded_size().width()); |
| coded_size_message->set_height(video_config.coded_size().height()); |
| |
| openscreen::cast::Rect* visible_rect_message = |
| video_message->mutable_visible_rect(); |
| visible_rect_message->set_x(video_config.visible_rect().x()); |
| visible_rect_message->set_y(video_config.visible_rect().y()); |
| visible_rect_message->set_width(video_config.visible_rect().width()); |
| visible_rect_message->set_height(video_config.visible_rect().height()); |
| |
| openscreen::cast::Size* natural_size_message = |
| video_message->mutable_natural_size(); |
| natural_size_message->set_width(video_config.natural_size().width()); |
| natural_size_message->set_height(video_config.natural_size().height()); |
| |
| if (!video_config.extra_data().empty()) { |
| video_message->set_extra_data(video_config.extra_data().data(), |
| video_config.extra_data().size()); |
| } |
| } |
| |
| bool ConvertProtoToVideoDecoderConfig( |
| const openscreen::cast::VideoDecoderConfig& video_message, |
| VideoDecoderConfig* video_config) { |
| DCHECK(video_config); |
| |
| // TODO(hubbe): Update pb to use VideoColorSpace |
| VideoColorSpace color_space; |
| switch (video_message.color_space()) { |
| case openscreen::cast::VideoDecoderConfig::COLOR_SPACE_UNSPECIFIED: |
| break; |
| case openscreen::cast::VideoDecoderConfig::COLOR_SPACE_JPEG: |
| color_space = VideoColorSpace::JPEG(); |
| break; |
| case openscreen::cast::VideoDecoderConfig::COLOR_SPACE_HD_REC709: |
| color_space = VideoColorSpace::REC709(); |
| break; |
| case openscreen::cast::VideoDecoderConfig::COLOR_SPACE_SD_REC601: |
| color_space = VideoColorSpace::REC601(); |
| break; |
| } |
| video_config->Initialize( |
| ToMediaVideoCodec(video_message.codec()).value(), |
| ToMediaVideoCodecProfile(video_message.profile()).value(), |
| IsOpaque(ToMediaVideoPixelFormat(video_message.format()).value()) |
| ? VideoDecoderConfig::AlphaMode::kIsOpaque |
| : VideoDecoderConfig::AlphaMode::kHasAlpha, |
| color_space, kNoTransformation, |
| gfx::Size(video_message.coded_size().width(), |
| video_message.coded_size().height()), |
| gfx::Rect(video_message.visible_rect().x(), |
| video_message.visible_rect().y(), |
| video_message.visible_rect().width(), |
| video_message.visible_rect().height()), |
| gfx::Size(video_message.natural_size().width(), |
| video_message.natural_size().height()), |
| std::vector<uint8_t>(video_message.extra_data().begin(), |
| video_message.extra_data().end()), |
| EncryptionScheme::kUnencrypted); |
| return video_config->IsValidConfig(); |
| } |
| |
| void ConvertProtoToPipelineStatistics( |
| const openscreen::cast::PipelineStatistics& stats_message, |
| PipelineStatistics* stats) { |
| stats->audio_bytes_decoded = stats_message.audio_bytes_decoded(); |
| stats->video_bytes_decoded = stats_message.video_bytes_decoded(); |
| stats->video_frames_decoded = stats_message.video_frames_decoded(); |
| stats->video_frames_dropped = stats_message.video_frames_dropped(); |
| stats->audio_memory_usage = stats_message.audio_memory_usage(); |
| stats->video_memory_usage = stats_message.video_memory_usage(); |
| // HACK: Set the following to prevent "disable video when hidden" logic in |
| // media::blink::WebMediaPlayerImpl. |
| stats->video_keyframe_distance_average = base::TimeDelta::Max(); |
| |
| // This field is not used by the rpc field. |
| stats->video_frames_decoded_power_efficient = 0; |
| |
| // The following fields were added after the initial message definition. Check |
| // that sender provided the values. |
| if (stats_message.has_audio_decoder_info()) { |
| auto audio_info = stats_message.audio_decoder_info(); |
| stats->audio_pipeline_info.decoder_type = |
| static_cast<AudioDecoderType>(audio_info.decoder_type()); |
| stats->audio_pipeline_info.is_platform_decoder = |
| audio_info.is_platform_decoder(); |
| stats->audio_pipeline_info.has_decrypting_demuxer_stream = false; |
| stats->audio_pipeline_info.encryption_type = EncryptionType::kClear; |
| } |
| if (stats_message.has_video_decoder_info()) { |
| auto video_info = stats_message.video_decoder_info(); |
| stats->video_pipeline_info.decoder_type = |
| static_cast<VideoDecoderType>(video_info.decoder_type()); |
| stats->video_pipeline_info.is_platform_decoder = |
| video_info.is_platform_decoder(); |
| stats->video_pipeline_info.has_decrypting_demuxer_stream = false; |
| stats->video_pipeline_info.encryption_type = EncryptionType::kClear; |
| } |
| if (stats_message.has_video_frame_duration_average_usec()) { |
| stats->video_frame_duration_average = |
| base::Microseconds(stats_message.video_frame_duration_average_usec()); |
| } |
| } |
| |
| } // namespace remoting |
| } // namespace media |