2020-08-14 16:58:22 +00:00
|
|
|
/*
|
|
|
|
* Copyright (c) 2018 The WebRTC project authors. All Rights Reserved.
|
|
|
|
*
|
|
|
|
* Use of this source code is governed by a BSD-style license
|
|
|
|
* that can be found in the LICENSE file in the root of the source
|
|
|
|
* tree. An additional intellectual property rights grant can be found
|
|
|
|
* in the file PATENTS. All contributing project authors may
|
|
|
|
* be found in the AUTHORS file in the root of the source tree.
|
|
|
|
*/
|
|
|
|
|
|
|
|
#include "call/rtp_payload_params.h"
|
|
|
|
|
|
|
|
#include <stddef.h>
|
|
|
|
|
|
|
|
#include <algorithm>
|
|
|
|
|
|
|
|
#include "absl/container/inlined_vector.h"
|
|
|
|
#include "absl/strings/match.h"
|
|
|
|
#include "absl/types/variant.h"
|
|
|
|
#include "api/video/video_timing.h"
|
|
|
|
#include "modules/video_coding/codecs/h264/include/h264_globals.h"
|
|
|
|
#include "modules/video_coding/codecs/interface/common_constants.h"
|
|
|
|
#include "modules/video_coding/codecs/vp8/include/vp8_globals.h"
|
|
|
|
#include "modules/video_coding/codecs/vp9/include/vp9_globals.h"
|
|
|
|
#include "modules/video_coding/frame_dependencies_calculator.h"
|
|
|
|
#include "rtc_base/arraysize.h"
|
|
|
|
#include "rtc_base/checks.h"
|
|
|
|
#include "rtc_base/logging.h"
|
|
|
|
#include "rtc_base/random.h"
|
|
|
|
#include "rtc_base/time_utils.h"
|
|
|
|
|
|
|
|
namespace webrtc {
|
|
|
|
|
|
|
|
namespace {
|
|
|
|
void PopulateRtpWithCodecSpecifics(const CodecSpecificInfo& info,
|
|
|
|
absl::optional<int> spatial_index,
|
|
|
|
RTPVideoHeader* rtp) {
|
|
|
|
rtp->codec = info.codecType;
|
2020-12-23 07:48:30 +00:00
|
|
|
rtp->is_last_frame_in_picture = info.end_of_picture;
|
2020-08-14 16:58:22 +00:00
|
|
|
switch (info.codecType) {
|
|
|
|
case kVideoCodecVP8: {
|
|
|
|
auto& vp8_header = rtp->video_type_header.emplace<RTPVideoHeaderVP8>();
|
|
|
|
vp8_header.InitRTPVideoHeaderVP8();
|
|
|
|
vp8_header.nonReference = info.codecSpecific.VP8.nonReference;
|
|
|
|
vp8_header.temporalIdx = info.codecSpecific.VP8.temporalIdx;
|
|
|
|
vp8_header.layerSync = info.codecSpecific.VP8.layerSync;
|
|
|
|
vp8_header.keyIdx = info.codecSpecific.VP8.keyIdx;
|
|
|
|
rtp->simulcastIdx = spatial_index.value_or(0);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
case kVideoCodecVP9: {
|
|
|
|
auto& vp9_header = rtp->video_type_header.emplace<RTPVideoHeaderVP9>();
|
|
|
|
vp9_header.InitRTPVideoHeaderVP9();
|
|
|
|
vp9_header.inter_pic_predicted =
|
|
|
|
info.codecSpecific.VP9.inter_pic_predicted;
|
|
|
|
vp9_header.flexible_mode = info.codecSpecific.VP9.flexible_mode;
|
|
|
|
vp9_header.ss_data_available = info.codecSpecific.VP9.ss_data_available;
|
|
|
|
vp9_header.non_ref_for_inter_layer_pred =
|
|
|
|
info.codecSpecific.VP9.non_ref_for_inter_layer_pred;
|
|
|
|
vp9_header.temporal_idx = info.codecSpecific.VP9.temporal_idx;
|
|
|
|
vp9_header.temporal_up_switch = info.codecSpecific.VP9.temporal_up_switch;
|
|
|
|
vp9_header.inter_layer_predicted =
|
|
|
|
info.codecSpecific.VP9.inter_layer_predicted;
|
|
|
|
vp9_header.gof_idx = info.codecSpecific.VP9.gof_idx;
|
|
|
|
vp9_header.num_spatial_layers = info.codecSpecific.VP9.num_spatial_layers;
|
|
|
|
vp9_header.first_active_layer = info.codecSpecific.VP9.first_active_layer;
|
|
|
|
if (vp9_header.num_spatial_layers > 1) {
|
|
|
|
vp9_header.spatial_idx = spatial_index.value_or(kNoSpatialIdx);
|
|
|
|
} else {
|
|
|
|
vp9_header.spatial_idx = kNoSpatialIdx;
|
|
|
|
}
|
|
|
|
if (info.codecSpecific.VP9.ss_data_available) {
|
|
|
|
vp9_header.spatial_layer_resolution_present =
|
|
|
|
info.codecSpecific.VP9.spatial_layer_resolution_present;
|
|
|
|
if (info.codecSpecific.VP9.spatial_layer_resolution_present) {
|
|
|
|
for (size_t i = 0; i < info.codecSpecific.VP9.num_spatial_layers;
|
|
|
|
++i) {
|
|
|
|
vp9_header.width[i] = info.codecSpecific.VP9.width[i];
|
|
|
|
vp9_header.height[i] = info.codecSpecific.VP9.height[i];
|
|
|
|
}
|
|
|
|
}
|
|
|
|
vp9_header.gof.CopyGofInfoVP9(info.codecSpecific.VP9.gof);
|
|
|
|
}
|
|
|
|
|
|
|
|
vp9_header.num_ref_pics = info.codecSpecific.VP9.num_ref_pics;
|
|
|
|
for (int i = 0; i < info.codecSpecific.VP9.num_ref_pics; ++i) {
|
|
|
|
vp9_header.pid_diff[i] = info.codecSpecific.VP9.p_diff[i];
|
|
|
|
}
|
2020-12-23 07:48:30 +00:00
|
|
|
vp9_header.end_of_picture = info.end_of_picture;
|
2020-08-14 16:58:22 +00:00
|
|
|
return;
|
|
|
|
}
|
|
|
|
case kVideoCodecH264: {
|
|
|
|
auto& h264_header = rtp->video_type_header.emplace<RTPVideoHeaderH264>();
|
|
|
|
h264_header.packetization_mode =
|
|
|
|
info.codecSpecific.H264.packetization_mode;
|
|
|
|
rtp->simulcastIdx = spatial_index.value_or(0);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
#ifndef DISABLE_H265
|
|
|
|
case kVideoCodecH265: {
|
|
|
|
auto& h265_header = rtp->video_type_header.emplace<RTPVideoHeaderH265>();
|
|
|
|
h265_header.packetization_mode =
|
|
|
|
info.codecSpecific.H265.packetization_mode;
|
|
|
|
}
|
|
|
|
return;
|
|
|
|
#endif
|
|
|
|
case kVideoCodecMultiplex:
|
|
|
|
case kVideoCodecGeneric:
|
|
|
|
rtp->codec = kVideoCodecGeneric;
|
|
|
|
rtp->simulcastIdx = spatial_index.value_or(0);
|
|
|
|
return;
|
|
|
|
default:
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
void SetVideoTiming(const EncodedImage& image, VideoSendTiming* timing) {
|
|
|
|
if (image.timing_.flags == VideoSendTiming::TimingFrameFlags::kInvalid ||
|
|
|
|
image.timing_.flags == VideoSendTiming::TimingFrameFlags::kNotTriggered) {
|
|
|
|
timing->flags = VideoSendTiming::TimingFrameFlags::kInvalid;
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
timing->encode_start_delta_ms = VideoSendTiming::GetDeltaCappedMs(
|
|
|
|
image.capture_time_ms_, image.timing_.encode_start_ms);
|
|
|
|
timing->encode_finish_delta_ms = VideoSendTiming::GetDeltaCappedMs(
|
|
|
|
image.capture_time_ms_, image.timing_.encode_finish_ms);
|
|
|
|
timing->packetization_finish_delta_ms = 0;
|
|
|
|
timing->pacer_exit_delta_ms = 0;
|
|
|
|
timing->network_timestamp_delta_ms = 0;
|
|
|
|
timing->network2_timestamp_delta_ms = 0;
|
|
|
|
timing->flags = image.timing_.flags;
|
|
|
|
}
|
|
|
|
} // namespace
|
|
|
|
|
|
|
|
RtpPayloadParams::RtpPayloadParams(const uint32_t ssrc,
|
|
|
|
const RtpPayloadState* state,
|
|
|
|
const WebRtcKeyValueConfig& trials)
|
|
|
|
: ssrc_(ssrc),
|
|
|
|
generic_picture_id_experiment_(
|
|
|
|
absl::StartsWith(trials.Lookup("WebRTC-GenericPictureId"),
|
2021-06-25 00:43:10 +00:00
|
|
|
"Enabled")),
|
|
|
|
simulate_generic_vp9_(
|
|
|
|
absl::StartsWith(trials.Lookup("WebRTC-Vp9DependencyDescriptor"),
|
2020-08-14 16:58:22 +00:00
|
|
|
"Enabled")) {
|
|
|
|
for (auto& spatial_layer : last_shared_frame_id_)
|
|
|
|
spatial_layer.fill(-1);
|
|
|
|
|
|
|
|
buffer_id_to_frame_id_.fill(-1);
|
|
|
|
|
|
|
|
Random random(rtc::TimeMicros());
|
|
|
|
state_.picture_id =
|
|
|
|
state ? state->picture_id : (random.Rand<int16_t>() & 0x7FFF);
|
|
|
|
state_.tl0_pic_idx = state ? state->tl0_pic_idx : (random.Rand<uint8_t>());
|
|
|
|
}
|
|
|
|
|
|
|
|
RtpPayloadParams::RtpPayloadParams(const RtpPayloadParams& other) = default;
|
|
|
|
|
|
|
|
RtpPayloadParams::~RtpPayloadParams() {}
|
|
|
|
|
|
|
|
RTPVideoHeader RtpPayloadParams::GetRtpVideoHeader(
|
|
|
|
const EncodedImage& image,
|
|
|
|
const CodecSpecificInfo* codec_specific_info,
|
|
|
|
int64_t shared_frame_id) {
|
|
|
|
RTPVideoHeader rtp_video_header;
|
|
|
|
if (codec_specific_info) {
|
|
|
|
PopulateRtpWithCodecSpecifics(*codec_specific_info, image.SpatialIndex(),
|
|
|
|
&rtp_video_header);
|
|
|
|
}
|
2021-06-25 00:43:10 +00:00
|
|
|
rtp_video_header.frame_type = image._frameType;
|
2020-08-14 16:58:22 +00:00
|
|
|
rtp_video_header.rotation = image.rotation_;
|
|
|
|
rtp_video_header.content_type = image.content_type_;
|
|
|
|
rtp_video_header.playout_delay = image.playout_delay_;
|
|
|
|
rtp_video_header.width = image._encodedWidth;
|
|
|
|
rtp_video_header.height = image._encodedHeight;
|
|
|
|
rtp_video_header.color_space = image.ColorSpace()
|
|
|
|
? absl::make_optional(*image.ColorSpace())
|
|
|
|
: absl::nullopt;
|
2021-06-25 00:43:10 +00:00
|
|
|
rtp_video_header.video_frame_tracking_id = image.VideoFrameTrackingId();
|
2020-08-14 16:58:22 +00:00
|
|
|
SetVideoTiming(image, &rtp_video_header.video_timing);
|
|
|
|
|
|
|
|
const bool is_keyframe = image._frameType == VideoFrameType::kVideoFrameKey;
|
|
|
|
const bool first_frame_in_picture =
|
|
|
|
(codec_specific_info && codec_specific_info->codecType == kVideoCodecVP9)
|
|
|
|
? codec_specific_info->codecSpecific.VP9.first_frame_in_picture
|
|
|
|
: true;
|
|
|
|
|
|
|
|
SetCodecSpecific(&rtp_video_header, first_frame_in_picture);
|
|
|
|
|
|
|
|
SetGeneric(codec_specific_info, shared_frame_id, is_keyframe,
|
|
|
|
&rtp_video_header);
|
|
|
|
|
|
|
|
return rtp_video_header;
|
|
|
|
}
|
|
|
|
|
|
|
|
uint32_t RtpPayloadParams::ssrc() const {
|
|
|
|
return ssrc_;
|
|
|
|
}
|
|
|
|
|
|
|
|
RtpPayloadState RtpPayloadParams::state() const {
|
|
|
|
return state_;
|
|
|
|
}
|
|
|
|
|
|
|
|
void RtpPayloadParams::SetCodecSpecific(RTPVideoHeader* rtp_video_header,
|
|
|
|
bool first_frame_in_picture) {
|
|
|
|
// Always set picture id. Set tl0_pic_idx iff temporal index is set.
|
|
|
|
if (first_frame_in_picture) {
|
|
|
|
state_.picture_id = (static_cast<uint16_t>(state_.picture_id) + 1) & 0x7FFF;
|
|
|
|
}
|
|
|
|
if (rtp_video_header->codec == kVideoCodecVP8) {
|
|
|
|
auto& vp8_header =
|
|
|
|
absl::get<RTPVideoHeaderVP8>(rtp_video_header->video_type_header);
|
|
|
|
vp8_header.pictureId = state_.picture_id;
|
|
|
|
|
|
|
|
if (vp8_header.temporalIdx != kNoTemporalIdx) {
|
|
|
|
if (vp8_header.temporalIdx == 0) {
|
|
|
|
++state_.tl0_pic_idx;
|
|
|
|
}
|
|
|
|
vp8_header.tl0PicIdx = state_.tl0_pic_idx;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if (rtp_video_header->codec == kVideoCodecVP9) {
|
|
|
|
auto& vp9_header =
|
|
|
|
absl::get<RTPVideoHeaderVP9>(rtp_video_header->video_type_header);
|
|
|
|
vp9_header.picture_id = state_.picture_id;
|
|
|
|
|
|
|
|
// Note that in the case that we have no temporal layers but we do have
|
|
|
|
// spatial layers, packets will carry layering info with a temporal_idx of
|
|
|
|
// zero, and we then have to set and increment tl0_pic_idx.
|
|
|
|
if (vp9_header.temporal_idx != kNoTemporalIdx ||
|
|
|
|
vp9_header.spatial_idx != kNoSpatialIdx) {
|
|
|
|
if (first_frame_in_picture &&
|
|
|
|
(vp9_header.temporal_idx == 0 ||
|
|
|
|
vp9_header.temporal_idx == kNoTemporalIdx)) {
|
|
|
|
++state_.tl0_pic_idx;
|
|
|
|
}
|
|
|
|
vp9_header.tl0_pic_idx = state_.tl0_pic_idx;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if (generic_picture_id_experiment_ &&
|
|
|
|
rtp_video_header->codec == kVideoCodecGeneric) {
|
|
|
|
rtp_video_header->video_type_header.emplace<RTPVideoHeaderLegacyGeneric>()
|
|
|
|
.picture_id = state_.picture_id;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
RTPVideoHeader::GenericDescriptorInfo
|
|
|
|
RtpPayloadParams::GenericDescriptorFromFrameInfo(
|
|
|
|
const GenericFrameInfo& frame_info,
|
2021-06-25 00:43:10 +00:00
|
|
|
int64_t frame_id) {
|
2020-08-14 16:58:22 +00:00
|
|
|
RTPVideoHeader::GenericDescriptorInfo generic;
|
|
|
|
generic.frame_id = frame_id;
|
|
|
|
generic.dependencies = dependencies_calculator_.FromBuffersUsage(
|
2021-06-25 00:43:10 +00:00
|
|
|
frame_id, frame_info.encoder_buffers);
|
2020-08-14 16:58:22 +00:00
|
|
|
generic.chain_diffs =
|
|
|
|
chains_calculator_.From(frame_id, frame_info.part_of_chain);
|
|
|
|
generic.spatial_index = frame_info.spatial_id;
|
|
|
|
generic.temporal_index = frame_info.temporal_id;
|
|
|
|
generic.decode_target_indications = frame_info.decode_target_indications;
|
|
|
|
generic.active_decode_targets = frame_info.active_decode_targets;
|
|
|
|
return generic;
|
|
|
|
}
|
|
|
|
|
|
|
|
void RtpPayloadParams::SetGeneric(const CodecSpecificInfo* codec_specific_info,
|
|
|
|
int64_t frame_id,
|
|
|
|
bool is_keyframe,
|
|
|
|
RTPVideoHeader* rtp_video_header) {
|
|
|
|
if (codec_specific_info && codec_specific_info->generic_frame_info &&
|
|
|
|
!codec_specific_info->generic_frame_info->encoder_buffers.empty()) {
|
|
|
|
if (is_keyframe) {
|
|
|
|
// Key frame resets all chains it is in.
|
|
|
|
chains_calculator_.Reset(
|
|
|
|
codec_specific_info->generic_frame_info->part_of_chain);
|
|
|
|
}
|
2021-06-25 00:43:10 +00:00
|
|
|
rtp_video_header->generic = GenericDescriptorFromFrameInfo(
|
|
|
|
*codec_specific_info->generic_frame_info, frame_id);
|
2020-08-14 16:58:22 +00:00
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
switch (rtp_video_header->codec) {
|
|
|
|
case VideoCodecType::kVideoCodecGeneric:
|
|
|
|
GenericToGeneric(frame_id, is_keyframe, rtp_video_header);
|
|
|
|
return;
|
|
|
|
case VideoCodecType::kVideoCodecVP8:
|
|
|
|
if (codec_specific_info) {
|
|
|
|
Vp8ToGeneric(codec_specific_info->codecSpecific.VP8, frame_id,
|
|
|
|
is_keyframe, rtp_video_header);
|
|
|
|
}
|
|
|
|
return;
|
|
|
|
case VideoCodecType::kVideoCodecVP9:
|
2021-06-25 00:43:10 +00:00
|
|
|
if (simulate_generic_vp9_ && codec_specific_info != nullptr) {
|
|
|
|
Vp9ToGeneric(codec_specific_info->codecSpecific.VP9, frame_id,
|
|
|
|
*rtp_video_header);
|
|
|
|
}
|
|
|
|
return;
|
2020-08-14 16:58:22 +00:00
|
|
|
case VideoCodecType::kVideoCodecAV1:
|
2021-06-25 00:43:10 +00:00
|
|
|
// TODO(philipel): Implement AV1 to generic descriptor.
|
2020-08-14 16:58:22 +00:00
|
|
|
return;
|
|
|
|
case VideoCodecType::kVideoCodecH264:
|
|
|
|
if (codec_specific_info) {
|
|
|
|
H264ToGeneric(codec_specific_info->codecSpecific.H264, frame_id,
|
|
|
|
is_keyframe, rtp_video_header);
|
|
|
|
}
|
|
|
|
return;
|
|
|
|
#ifndef DISABLE_H265
|
|
|
|
case VideoCodecType::kVideoCodecH265:
|
|
|
|
#endif
|
|
|
|
case VideoCodecType::kVideoCodecMultiplex:
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
RTC_NOTREACHED() << "Unsupported codec.";
|
|
|
|
}
|
|
|
|
|
|
|
|
void RtpPayloadParams::GenericToGeneric(int64_t shared_frame_id,
|
|
|
|
bool is_keyframe,
|
|
|
|
RTPVideoHeader* rtp_video_header) {
|
|
|
|
RTPVideoHeader::GenericDescriptorInfo& generic =
|
|
|
|
rtp_video_header->generic.emplace();
|
|
|
|
|
|
|
|
generic.frame_id = shared_frame_id;
|
|
|
|
|
|
|
|
if (is_keyframe) {
|
|
|
|
last_shared_frame_id_[0].fill(-1);
|
|
|
|
} else {
|
|
|
|
int64_t frame_id = last_shared_frame_id_[0][0];
|
|
|
|
RTC_DCHECK_NE(frame_id, -1);
|
|
|
|
RTC_DCHECK_LT(frame_id, shared_frame_id);
|
|
|
|
generic.dependencies.push_back(frame_id);
|
|
|
|
}
|
|
|
|
|
|
|
|
last_shared_frame_id_[0][0] = shared_frame_id;
|
|
|
|
}
|
|
|
|
|
|
|
|
void RtpPayloadParams::H264ToGeneric(const CodecSpecificInfoH264& h264_info,
|
|
|
|
int64_t shared_frame_id,
|
|
|
|
bool is_keyframe,
|
|
|
|
RTPVideoHeader* rtp_video_header) {
|
|
|
|
const int temporal_index =
|
|
|
|
h264_info.temporal_idx != kNoTemporalIdx ? h264_info.temporal_idx : 0;
|
|
|
|
|
|
|
|
if (temporal_index >= RtpGenericFrameDescriptor::kMaxTemporalLayers) {
|
|
|
|
RTC_LOG(LS_WARNING) << "Temporal and/or spatial index is too high to be "
|
|
|
|
"used with generic frame descriptor.";
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
RTPVideoHeader::GenericDescriptorInfo& generic =
|
|
|
|
rtp_video_header->generic.emplace();
|
|
|
|
|
|
|
|
generic.frame_id = shared_frame_id;
|
|
|
|
generic.temporal_index = temporal_index;
|
|
|
|
|
|
|
|
if (is_keyframe) {
|
|
|
|
RTC_DCHECK_EQ(temporal_index, 0);
|
|
|
|
last_shared_frame_id_[/*spatial index*/ 0].fill(-1);
|
|
|
|
last_shared_frame_id_[/*spatial index*/ 0][temporal_index] =
|
|
|
|
shared_frame_id;
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (h264_info.base_layer_sync) {
|
|
|
|
int64_t tl0_frame_id = last_shared_frame_id_[/*spatial index*/ 0][0];
|
|
|
|
|
|
|
|
for (int i = 1; i < RtpGenericFrameDescriptor::kMaxTemporalLayers; ++i) {
|
|
|
|
if (last_shared_frame_id_[/*spatial index*/ 0][i] < tl0_frame_id) {
|
|
|
|
last_shared_frame_id_[/*spatial index*/ 0][i] = -1;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
RTC_DCHECK_GE(tl0_frame_id, 0);
|
|
|
|
RTC_DCHECK_LT(tl0_frame_id, shared_frame_id);
|
|
|
|
generic.dependencies.push_back(tl0_frame_id);
|
|
|
|
} else {
|
|
|
|
for (int i = 0; i <= temporal_index; ++i) {
|
|
|
|
int64_t frame_id = last_shared_frame_id_[/*spatial index*/ 0][i];
|
|
|
|
|
|
|
|
if (frame_id != -1) {
|
|
|
|
RTC_DCHECK_LT(frame_id, shared_frame_id);
|
|
|
|
generic.dependencies.push_back(frame_id);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
last_shared_frame_id_[/*spatial_index*/ 0][temporal_index] = shared_frame_id;
|
|
|
|
}
|
|
|
|
|
|
|
|
void RtpPayloadParams::Vp8ToGeneric(const CodecSpecificInfoVP8& vp8_info,
|
|
|
|
int64_t shared_frame_id,
|
|
|
|
bool is_keyframe,
|
|
|
|
RTPVideoHeader* rtp_video_header) {
|
|
|
|
const auto& vp8_header =
|
|
|
|
absl::get<RTPVideoHeaderVP8>(rtp_video_header->video_type_header);
|
|
|
|
const int spatial_index = 0;
|
|
|
|
const int temporal_index =
|
|
|
|
vp8_header.temporalIdx != kNoTemporalIdx ? vp8_header.temporalIdx : 0;
|
|
|
|
|
|
|
|
if (temporal_index >= RtpGenericFrameDescriptor::kMaxTemporalLayers ||
|
|
|
|
spatial_index >= RtpGenericFrameDescriptor::kMaxSpatialLayers) {
|
|
|
|
RTC_LOG(LS_WARNING) << "Temporal and/or spatial index is too high to be "
|
|
|
|
"used with generic frame descriptor.";
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
RTPVideoHeader::GenericDescriptorInfo& generic =
|
|
|
|
rtp_video_header->generic.emplace();
|
|
|
|
|
|
|
|
generic.frame_id = shared_frame_id;
|
|
|
|
generic.spatial_index = spatial_index;
|
|
|
|
generic.temporal_index = temporal_index;
|
|
|
|
|
|
|
|
if (vp8_info.useExplicitDependencies) {
|
|
|
|
SetDependenciesVp8New(vp8_info, shared_frame_id, is_keyframe,
|
|
|
|
vp8_header.layerSync, &generic);
|
|
|
|
} else {
|
|
|
|
SetDependenciesVp8Deprecated(vp8_info, shared_frame_id, is_keyframe,
|
|
|
|
spatial_index, temporal_index,
|
|
|
|
vp8_header.layerSync, &generic);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2021-06-25 00:43:10 +00:00
|
|
|
FrameDependencyStructure RtpPayloadParams::MinimalisticVp9Structure(
|
|
|
|
const CodecSpecificInfoVP9& vp9) {
|
|
|
|
const int num_spatial_layers = vp9.num_spatial_layers;
|
|
|
|
const int num_temporal_layers = kMaxTemporalStreams;
|
|
|
|
FrameDependencyStructure structure;
|
|
|
|
structure.num_decode_targets = num_spatial_layers * num_temporal_layers;
|
|
|
|
structure.num_chains = num_spatial_layers;
|
|
|
|
structure.templates.reserve(num_spatial_layers * num_temporal_layers);
|
|
|
|
for (int sid = 0; sid < num_spatial_layers; ++sid) {
|
|
|
|
for (int tid = 0; tid < num_temporal_layers; ++tid) {
|
|
|
|
FrameDependencyTemplate a_template;
|
|
|
|
a_template.spatial_id = sid;
|
|
|
|
a_template.temporal_id = tid;
|
|
|
|
for (int s = 0; s < num_spatial_layers; ++s) {
|
|
|
|
for (int t = 0; t < num_temporal_layers; ++t) {
|
|
|
|
// Prefer kSwitch for indication frame is part of the decode target
|
|
|
|
// because RtpPayloadParams::Vp9ToGeneric uses that indication more
|
|
|
|
// often that kRequired, increasing chance custom dti need not to
|
|
|
|
// use more bits in dependency descriptor on the wire.
|
|
|
|
a_template.decode_target_indications.push_back(
|
|
|
|
sid <= s && tid <= t ? DecodeTargetIndication::kSwitch
|
|
|
|
: DecodeTargetIndication::kNotPresent);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
a_template.frame_diffs.push_back(tid == 0 ? num_spatial_layers *
|
|
|
|
num_temporal_layers
|
|
|
|
: num_spatial_layers);
|
|
|
|
a_template.chain_diffs.assign(structure.num_chains, 1);
|
|
|
|
structure.templates.push_back(a_template);
|
|
|
|
|
|
|
|
structure.decode_target_protected_by_chain.push_back(sid);
|
|
|
|
}
|
|
|
|
if (vp9.ss_data_available && vp9.spatial_layer_resolution_present) {
|
|
|
|
structure.resolutions.emplace_back(vp9.width[sid], vp9.height[sid]);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return structure;
|
|
|
|
}
|
|
|
|
|
|
|
|
void RtpPayloadParams::Vp9ToGeneric(const CodecSpecificInfoVP9& vp9_info,
|
|
|
|
int64_t shared_frame_id,
|
|
|
|
RTPVideoHeader& rtp_video_header) {
|
|
|
|
const auto& vp9_header =
|
|
|
|
absl::get<RTPVideoHeaderVP9>(rtp_video_header.video_type_header);
|
|
|
|
const int num_spatial_layers = vp9_header.num_spatial_layers;
|
|
|
|
const int num_temporal_layers = kMaxTemporalStreams;
|
|
|
|
|
|
|
|
int spatial_index =
|
|
|
|
vp9_header.spatial_idx != kNoSpatialIdx ? vp9_header.spatial_idx : 0;
|
|
|
|
int temporal_index =
|
|
|
|
vp9_header.temporal_idx != kNoTemporalIdx ? vp9_header.temporal_idx : 0;
|
|
|
|
|
|
|
|
if (spatial_index >= num_spatial_layers ||
|
|
|
|
temporal_index >= num_temporal_layers ||
|
|
|
|
num_spatial_layers > RtpGenericFrameDescriptor::kMaxSpatialLayers) {
|
|
|
|
// Prefer to generate no generic layering than an inconsistent one.
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
RTPVideoHeader::GenericDescriptorInfo& result =
|
|
|
|
rtp_video_header.generic.emplace();
|
|
|
|
|
|
|
|
result.frame_id = shared_frame_id;
|
|
|
|
result.spatial_index = spatial_index;
|
|
|
|
result.temporal_index = temporal_index;
|
|
|
|
|
|
|
|
result.decode_target_indications.reserve(num_spatial_layers *
|
|
|
|
num_temporal_layers);
|
|
|
|
for (int sid = 0; sid < num_spatial_layers; ++sid) {
|
|
|
|
for (int tid = 0; tid < num_temporal_layers; ++tid) {
|
|
|
|
DecodeTargetIndication dti;
|
|
|
|
if (sid < spatial_index || tid < temporal_index) {
|
|
|
|
dti = DecodeTargetIndication::kNotPresent;
|
|
|
|
} else if (spatial_index != sid &&
|
|
|
|
vp9_header.non_ref_for_inter_layer_pred) {
|
|
|
|
dti = DecodeTargetIndication::kNotPresent;
|
|
|
|
} else if (sid == spatial_index && tid == temporal_index) {
|
|
|
|
// Assume that if frame is decodable, all of its own layer is decodable.
|
|
|
|
dti = DecodeTargetIndication::kSwitch;
|
|
|
|
} else if (sid == spatial_index && vp9_header.temporal_up_switch) {
|
|
|
|
dti = DecodeTargetIndication::kSwitch;
|
|
|
|
} else if (!vp9_header.inter_pic_predicted) {
|
|
|
|
// Key frame or spatial upswitch
|
|
|
|
dti = DecodeTargetIndication::kSwitch;
|
|
|
|
} else {
|
|
|
|
// Make no other assumptions. That should be safe, though suboptimal.
|
|
|
|
// To provide more accurate dti, encoder wrapper should fill in
|
|
|
|
// CodecSpecificInfo::generic_frame_info
|
|
|
|
dti = DecodeTargetIndication::kRequired;
|
|
|
|
}
|
|
|
|
result.decode_target_indications.push_back(dti);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Calculate frame dependencies.
|
|
|
|
static constexpr int kPictureDiffLimit = 128;
|
|
|
|
if (last_vp9_frame_id_.empty()) {
|
|
|
|
// Create the array only if it is ever used.
|
|
|
|
last_vp9_frame_id_.resize(kPictureDiffLimit);
|
|
|
|
}
|
|
|
|
if (vp9_header.inter_layer_predicted && spatial_index > 0) {
|
|
|
|
result.dependencies.push_back(
|
|
|
|
last_vp9_frame_id_[vp9_header.picture_id % kPictureDiffLimit]
|
|
|
|
[spatial_index - 1]);
|
|
|
|
}
|
|
|
|
if (vp9_header.inter_pic_predicted) {
|
|
|
|
for (size_t i = 0; i < vp9_header.num_ref_pics; ++i) {
|
|
|
|
// picture_id is 15 bit number that wraps around. Though undeflow may
|
|
|
|
// produce picture that exceeds 2^15, it is ok because in this
|
|
|
|
// code block only last 7 bits of the picture_id are used.
|
|
|
|
uint16_t depend_on = vp9_header.picture_id - vp9_header.pid_diff[i];
|
|
|
|
result.dependencies.push_back(
|
|
|
|
last_vp9_frame_id_[depend_on % kPictureDiffLimit][spatial_index]);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
last_vp9_frame_id_[vp9_header.picture_id % kPictureDiffLimit][spatial_index] =
|
|
|
|
shared_frame_id;
|
|
|
|
|
|
|
|
// Calculate chains, asuming chain includes all frames with temporal_id = 0
|
|
|
|
if (!vp9_header.inter_pic_predicted && !vp9_header.inter_layer_predicted) {
|
|
|
|
// Assume frames without dependencies also reset chains.
|
|
|
|
for (int sid = spatial_index; sid < num_spatial_layers; ++sid) {
|
|
|
|
chain_last_frame_id_[sid] = -1;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
result.chain_diffs.resize(num_spatial_layers);
|
|
|
|
for (int sid = 0; sid < num_spatial_layers; ++sid) {
|
|
|
|
if (chain_last_frame_id_[sid] == -1) {
|
|
|
|
result.chain_diffs[sid] = 0;
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
result.chain_diffs[sid] = shared_frame_id - chain_last_frame_id_[sid];
|
|
|
|
}
|
|
|
|
|
|
|
|
if (temporal_index == 0) {
|
|
|
|
chain_last_frame_id_[spatial_index] = shared_frame_id;
|
|
|
|
if (!vp9_header.non_ref_for_inter_layer_pred) {
|
|
|
|
for (int sid = spatial_index + 1; sid < num_spatial_layers; ++sid) {
|
|
|
|
chain_last_frame_id_[sid] = shared_frame_id;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2020-08-14 16:58:22 +00:00
|
|
|
void RtpPayloadParams::SetDependenciesVp8Deprecated(
|
|
|
|
const CodecSpecificInfoVP8& vp8_info,
|
|
|
|
int64_t shared_frame_id,
|
|
|
|
bool is_keyframe,
|
|
|
|
int spatial_index,
|
|
|
|
int temporal_index,
|
|
|
|
bool layer_sync,
|
|
|
|
RTPVideoHeader::GenericDescriptorInfo* generic) {
|
|
|
|
RTC_DCHECK(!vp8_info.useExplicitDependencies);
|
|
|
|
RTC_DCHECK(!new_version_used_.has_value() || !new_version_used_.value());
|
|
|
|
new_version_used_ = false;
|
|
|
|
|
|
|
|
if (is_keyframe) {
|
|
|
|
RTC_DCHECK_EQ(temporal_index, 0);
|
|
|
|
last_shared_frame_id_[spatial_index].fill(-1);
|
|
|
|
last_shared_frame_id_[spatial_index][temporal_index] = shared_frame_id;
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (layer_sync) {
|
|
|
|
int64_t tl0_frame_id = last_shared_frame_id_[spatial_index][0];
|
|
|
|
|
|
|
|
for (int i = 1; i < RtpGenericFrameDescriptor::kMaxTemporalLayers; ++i) {
|
|
|
|
if (last_shared_frame_id_[spatial_index][i] < tl0_frame_id) {
|
|
|
|
last_shared_frame_id_[spatial_index][i] = -1;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
RTC_DCHECK_GE(tl0_frame_id, 0);
|
|
|
|
RTC_DCHECK_LT(tl0_frame_id, shared_frame_id);
|
|
|
|
generic->dependencies.push_back(tl0_frame_id);
|
|
|
|
} else {
|
|
|
|
for (int i = 0; i <= temporal_index; ++i) {
|
|
|
|
int64_t frame_id = last_shared_frame_id_[spatial_index][i];
|
|
|
|
|
|
|
|
if (frame_id != -1) {
|
|
|
|
RTC_DCHECK_LT(frame_id, shared_frame_id);
|
|
|
|
generic->dependencies.push_back(frame_id);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
last_shared_frame_id_[spatial_index][temporal_index] = shared_frame_id;
|
|
|
|
}
|
|
|
|
|
|
|
|
void RtpPayloadParams::SetDependenciesVp8New(
|
|
|
|
const CodecSpecificInfoVP8& vp8_info,
|
|
|
|
int64_t shared_frame_id,
|
|
|
|
bool is_keyframe,
|
|
|
|
bool layer_sync,
|
|
|
|
RTPVideoHeader::GenericDescriptorInfo* generic) {
|
|
|
|
RTC_DCHECK(vp8_info.useExplicitDependencies);
|
|
|
|
RTC_DCHECK(!new_version_used_.has_value() || new_version_used_.value());
|
|
|
|
new_version_used_ = true;
|
|
|
|
|
|
|
|
if (is_keyframe) {
|
|
|
|
RTC_DCHECK_EQ(vp8_info.referencedBuffersCount, 0u);
|
|
|
|
buffer_id_to_frame_id_.fill(shared_frame_id);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
constexpr size_t kBuffersCountVp8 = CodecSpecificInfoVP8::kBuffersCount;
|
|
|
|
|
|
|
|
RTC_DCHECK_GT(vp8_info.referencedBuffersCount, 0u);
|
|
|
|
RTC_DCHECK_LE(vp8_info.referencedBuffersCount,
|
|
|
|
arraysize(vp8_info.referencedBuffers));
|
|
|
|
|
|
|
|
for (size_t i = 0; i < vp8_info.referencedBuffersCount; ++i) {
|
|
|
|
const size_t referenced_buffer = vp8_info.referencedBuffers[i];
|
|
|
|
RTC_DCHECK_LT(referenced_buffer, kBuffersCountVp8);
|
|
|
|
RTC_DCHECK_LT(referenced_buffer, buffer_id_to_frame_id_.size());
|
|
|
|
|
|
|
|
const int64_t dependency_frame_id =
|
|
|
|
buffer_id_to_frame_id_[referenced_buffer];
|
|
|
|
RTC_DCHECK_GE(dependency_frame_id, 0);
|
|
|
|
RTC_DCHECK_LT(dependency_frame_id, shared_frame_id);
|
|
|
|
|
|
|
|
const bool is_new_dependency =
|
|
|
|
std::find(generic->dependencies.begin(), generic->dependencies.end(),
|
|
|
|
dependency_frame_id) == generic->dependencies.end();
|
|
|
|
if (is_new_dependency) {
|
|
|
|
generic->dependencies.push_back(dependency_frame_id);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
RTC_DCHECK_LE(vp8_info.updatedBuffersCount, kBuffersCountVp8);
|
|
|
|
for (size_t i = 0; i < vp8_info.updatedBuffersCount; ++i) {
|
|
|
|
const size_t updated_id = vp8_info.updatedBuffers[i];
|
|
|
|
buffer_id_to_frame_id_[updated_id] = shared_frame_id;
|
|
|
|
}
|
|
|
|
|
|
|
|
RTC_DCHECK_LE(buffer_id_to_frame_id_.size(), kBuffersCountVp8);
|
|
|
|
}
|
|
|
|
|
|
|
|
} // namespace webrtc
|