mirror of
https://github.com/mollyim/webrtc.git
synced 2025-05-12 21:30:45 +01:00

Implemented behind `WebRTC-Video-SimulcastIndependentFrameIds`. Bug: b/329063481 Change-Id: I683e567bb5b449f998be57ec3a11bb3b95e3ace4 Reviewed-on: https://webrtc-review.googlesource.com/c/src/+/343382 Commit-Queue: Philip Eliasson <philipel@webrtc.org> Reviewed-by: Danil Chapovalov <danilchap@webrtc.org> Cr-Commit-Position: refs/heads/main@{#41927}
1330 lines
54 KiB
C++
1330 lines
54 KiB
C++
/*
|
|
* Copyright (c) 2015 The WebRTC project authors. All Rights Reserved.
|
|
*
|
|
* Use of this source code is governed by a BSD-style license
|
|
* that can be found in the LICENSE file in the root of the source
|
|
* tree. An additional intellectual property rights grant can be found
|
|
* in the file PATENTS. All contributing project authors may
|
|
* be found in the AUTHORS file in the root of the source tree.
|
|
*/
|
|
|
|
#include "call/rtp_video_sender.h"
|
|
|
|
#include <atomic>
|
|
#include <memory>
|
|
#include <string>
|
|
#include <utility>
|
|
|
|
#include "absl/functional/any_invocable.h"
|
|
#include "api/environment/environment.h"
|
|
#include "api/environment/environment_factory.h"
|
|
#include "api/test/mock_frame_transformer.h"
|
|
#include "call/rtp_transport_controller_send.h"
|
|
#include "modules/rtp_rtcp/include/rtp_rtcp_defines.h"
|
|
#include "modules/rtp_rtcp/source/byte_io.h"
|
|
#include "modules/rtp_rtcp/source/rtcp_packet/nack.h"
|
|
#include "modules/rtp_rtcp/source/rtp_dependency_descriptor_extension.h"
|
|
#include "modules/rtp_rtcp/source/rtp_packet.h"
|
|
#include "modules/video_coding/fec_controller_default.h"
|
|
#include "modules/video_coding/include/video_codec_interface.h"
|
|
#include "rtc_base/rate_limiter.h"
|
|
#include "test/explicit_key_value_config.h"
|
|
#include "test/gmock.h"
|
|
#include "test/gtest.h"
|
|
#include "test/mock_transport.h"
|
|
#include "test/scenario/scenario.h"
|
|
#include "test/scoped_key_value_config.h"
|
|
#include "test/time_controller/simulated_time_controller.h"
|
|
#include "video/send_statistics_proxy.h"
|
|
|
|
namespace webrtc {
|
|
namespace {
|
|
|
|
using ::testing::_;
|
|
using ::testing::NiceMock;
|
|
using ::testing::SaveArg;
|
|
using ::testing::SizeIs;
|
|
|
|
const int8_t kPayloadType = 96;
|
|
const uint32_t kSsrc1 = 12345;
|
|
const uint32_t kSsrc2 = 23456;
|
|
const uint32_t kRtxSsrc1 = 34567;
|
|
const uint32_t kRtxSsrc2 = 45678;
|
|
const int16_t kInitialPictureId1 = 222;
|
|
const int16_t kInitialPictureId2 = 44;
|
|
const int16_t kInitialTl0PicIdx1 = 99;
|
|
const int16_t kInitialTl0PicIdx2 = 199;
|
|
const int64_t kRetransmitWindowSizeMs = 500;
|
|
const int kTransportsSequenceExtensionId = 7;
|
|
const int kDependencyDescriptorExtensionId = 8;
|
|
|
|
class MockRtcpIntraFrameObserver : public RtcpIntraFrameObserver {
|
|
public:
|
|
MOCK_METHOD(void, OnReceivedIntraFrameRequest, (uint32_t), (override));
|
|
};
|
|
|
|
RtpSenderObservers CreateObservers(
|
|
RtcpIntraFrameObserver* intra_frame_callback,
|
|
ReportBlockDataObserver* report_block_data_observer,
|
|
StreamDataCountersCallback* rtp_stats,
|
|
BitrateStatisticsObserver* bitrate_observer,
|
|
FrameCountObserver* frame_count_observer,
|
|
RtcpPacketTypeCounterObserver* rtcp_type_observer) {
|
|
RtpSenderObservers observers;
|
|
observers.rtcp_rtt_stats = nullptr;
|
|
observers.intra_frame_callback = intra_frame_callback;
|
|
observers.rtcp_loss_notification_observer = nullptr;
|
|
observers.report_block_data_observer = report_block_data_observer;
|
|
observers.rtp_stats = rtp_stats;
|
|
observers.bitrate_observer = bitrate_observer;
|
|
observers.frame_count_observer = frame_count_observer;
|
|
observers.rtcp_type_observer = rtcp_type_observer;
|
|
observers.send_packet_observer = nullptr;
|
|
return observers;
|
|
}
|
|
|
|
BitrateConstraints GetBitrateConfig() {
|
|
BitrateConstraints bitrate_config;
|
|
bitrate_config.min_bitrate_bps = 30000;
|
|
bitrate_config.start_bitrate_bps = 300000;
|
|
bitrate_config.max_bitrate_bps = 3000000;
|
|
return bitrate_config;
|
|
}
|
|
|
|
VideoSendStream::Config CreateVideoSendStreamConfig(
|
|
Transport* transport,
|
|
const std::vector<uint32_t>& ssrcs,
|
|
const std::vector<uint32_t>& rtx_ssrcs,
|
|
int payload_type) {
|
|
VideoSendStream::Config config(transport);
|
|
config.rtp.ssrcs = ssrcs;
|
|
config.rtp.rtx.ssrcs = rtx_ssrcs;
|
|
config.rtp.payload_type = payload_type;
|
|
config.rtp.rtx.payload_type = payload_type + 1;
|
|
config.rtp.nack.rtp_history_ms = 1000;
|
|
config.rtp.extensions.emplace_back(RtpExtension::kTransportSequenceNumberUri,
|
|
kTransportsSequenceExtensionId);
|
|
config.rtp.extensions.emplace_back(RtpDependencyDescriptorExtension::Uri(),
|
|
kDependencyDescriptorExtensionId);
|
|
config.rtp.extmap_allow_mixed = true;
|
|
return config;
|
|
}
|
|
|
|
class RtpVideoSenderTestFixture {
|
|
public:
|
|
RtpVideoSenderTestFixture(
|
|
const std::vector<uint32_t>& ssrcs,
|
|
const std::vector<uint32_t>& rtx_ssrcs,
|
|
int payload_type,
|
|
const std::map<uint32_t, RtpPayloadState>& suspended_payload_states,
|
|
FrameCountObserver* frame_count_observer,
|
|
rtc::scoped_refptr<FrameTransformerInterface> frame_transformer,
|
|
const FieldTrialsView* field_trials = nullptr)
|
|
: time_controller_(Timestamp::Millis(1000000)),
|
|
env_(CreateEnvironment(&field_trials_,
|
|
field_trials,
|
|
time_controller_.GetClock(),
|
|
time_controller_.CreateTaskQueueFactory())),
|
|
config_(CreateVideoSendStreamConfig(&transport_,
|
|
ssrcs,
|
|
rtx_ssrcs,
|
|
payload_type)),
|
|
bitrate_config_(GetBitrateConfig()),
|
|
transport_controller_(
|
|
RtpTransportConfig{.env = env_, .bitrate_config = bitrate_config_}),
|
|
stats_proxy_(time_controller_.GetClock(),
|
|
config_,
|
|
VideoEncoderConfig::ContentType::kRealtimeVideo,
|
|
env_.field_trials()),
|
|
retransmission_rate_limiter_(time_controller_.GetClock(),
|
|
kRetransmitWindowSizeMs) {
|
|
transport_controller_.EnsureStarted();
|
|
std::map<uint32_t, RtpState> suspended_ssrcs;
|
|
router_ = std::make_unique<RtpVideoSender>(
|
|
time_controller_.GetClock(), suspended_ssrcs, suspended_payload_states,
|
|
config_.rtp, config_.rtcp_report_interval_ms, &transport_,
|
|
CreateObservers(&encoder_feedback_, &stats_proxy_, &stats_proxy_,
|
|
&stats_proxy_, frame_count_observer, &stats_proxy_),
|
|
&transport_controller_, &env_.event_log(),
|
|
&retransmission_rate_limiter_,
|
|
std::make_unique<FecControllerDefault>(env_), nullptr, CryptoOptions{},
|
|
frame_transformer, env_.field_trials(),
|
|
time_controller_.GetTaskQueueFactory());
|
|
}
|
|
|
|
RtpVideoSenderTestFixture(
|
|
const std::vector<uint32_t>& ssrcs,
|
|
const std::vector<uint32_t>& rtx_ssrcs,
|
|
int payload_type,
|
|
const std::map<uint32_t, RtpPayloadState>& suspended_payload_states,
|
|
FrameCountObserver* frame_count_observer,
|
|
const FieldTrialsView* field_trials = nullptr)
|
|
: RtpVideoSenderTestFixture(ssrcs,
|
|
rtx_ssrcs,
|
|
payload_type,
|
|
suspended_payload_states,
|
|
frame_count_observer,
|
|
/*frame_transformer=*/nullptr,
|
|
field_trials) {}
|
|
|
|
RtpVideoSenderTestFixture(
|
|
const std::vector<uint32_t>& ssrcs,
|
|
const std::vector<uint32_t>& rtx_ssrcs,
|
|
int payload_type,
|
|
const std::map<uint32_t, RtpPayloadState>& suspended_payload_states,
|
|
const FieldTrialsView* field_trials = nullptr)
|
|
: RtpVideoSenderTestFixture(ssrcs,
|
|
rtx_ssrcs,
|
|
payload_type,
|
|
suspended_payload_states,
|
|
/*frame_count_observer=*/nullptr,
|
|
/*frame_transformer=*/nullptr,
|
|
field_trials) {}
|
|
|
|
~RtpVideoSenderTestFixture() { SetSending(false); }
|
|
|
|
RtpVideoSender* router() { return router_.get(); }
|
|
MockTransport& transport() { return transport_; }
|
|
void AdvanceTime(TimeDelta delta) { time_controller_.AdvanceTime(delta); }
|
|
|
|
void SetSending(bool sending) { router_->SetSending(sending); }
|
|
|
|
private:
|
|
test::ScopedKeyValueConfig field_trials_;
|
|
NiceMock<MockTransport> transport_;
|
|
NiceMock<MockRtcpIntraFrameObserver> encoder_feedback_;
|
|
GlobalSimulatedTimeController time_controller_;
|
|
Environment env_;
|
|
VideoSendStream::Config config_;
|
|
BitrateConstraints bitrate_config_;
|
|
RtpTransportControllerSend transport_controller_;
|
|
SendStatisticsProxy stats_proxy_;
|
|
RateLimiter retransmission_rate_limiter_;
|
|
std::unique_ptr<RtpVideoSender> router_;
|
|
};
|
|
|
|
BitrateAllocationUpdate CreateBitrateAllocationUpdate(int target_bitrate_bps) {
|
|
BitrateAllocationUpdate update;
|
|
update.target_bitrate = DataRate::BitsPerSec(target_bitrate_bps);
|
|
update.round_trip_time = TimeDelta::Zero();
|
|
return update;
|
|
}
|
|
|
|
} // namespace
|
|
|
|
TEST(RtpVideoSenderTest, SendOnOneModule) {
|
|
constexpr uint8_t kPayload = 'a';
|
|
EncodedImage encoded_image;
|
|
encoded_image.SetRtpTimestamp(1);
|
|
encoded_image.capture_time_ms_ = 2;
|
|
encoded_image._frameType = VideoFrameType::kVideoFrameKey;
|
|
encoded_image.SetEncodedData(EncodedImageBuffer::Create(&kPayload, 1));
|
|
|
|
RtpVideoSenderTestFixture test({kSsrc1}, {kRtxSsrc1}, kPayloadType, {});
|
|
EXPECT_NE(EncodedImageCallback::Result::OK,
|
|
test.router()->OnEncodedImage(encoded_image, nullptr).error);
|
|
|
|
test.SetSending(true);
|
|
EXPECT_EQ(EncodedImageCallback::Result::OK,
|
|
test.router()->OnEncodedImage(encoded_image, nullptr).error);
|
|
|
|
test.SetSending(false);
|
|
EXPECT_NE(EncodedImageCallback::Result::OK,
|
|
test.router()->OnEncodedImage(encoded_image, nullptr).error);
|
|
|
|
test.SetSending(true);
|
|
EXPECT_EQ(EncodedImageCallback::Result::OK,
|
|
test.router()->OnEncodedImage(encoded_image, nullptr).error);
|
|
}
|
|
|
|
TEST(RtpVideoSenderTest, OnEncodedImageReturnOkWhenSendingTrue) {
|
|
constexpr uint8_t kPayload = 'a';
|
|
EncodedImage encoded_image_1;
|
|
encoded_image_1.SetRtpTimestamp(1);
|
|
encoded_image_1.capture_time_ms_ = 2;
|
|
encoded_image_1._frameType = VideoFrameType::kVideoFrameKey;
|
|
encoded_image_1.SetEncodedData(EncodedImageBuffer::Create(&kPayload, 1));
|
|
|
|
RtpVideoSenderTestFixture test({kSsrc1, kSsrc2}, {kRtxSsrc1, kRtxSsrc2},
|
|
kPayloadType, {});
|
|
|
|
CodecSpecificInfo codec_info;
|
|
codec_info.codecType = kVideoCodecVP8;
|
|
|
|
test.SetSending(true);
|
|
EXPECT_EQ(EncodedImageCallback::Result::OK,
|
|
test.router()->OnEncodedImage(encoded_image_1, &codec_info).error);
|
|
|
|
EncodedImage encoded_image_2(encoded_image_1);
|
|
encoded_image_2.SetSimulcastIndex(1);
|
|
EXPECT_EQ(EncodedImageCallback::Result::OK,
|
|
test.router()->OnEncodedImage(encoded_image_2, &codec_info).error);
|
|
}
|
|
|
|
TEST(RtpVideoSenderTest, OnEncodedImageReturnErrorCodeWhenSendingFalse) {
|
|
constexpr uint8_t kPayload = 'a';
|
|
EncodedImage encoded_image_1;
|
|
encoded_image_1.SetRtpTimestamp(1);
|
|
encoded_image_1.capture_time_ms_ = 2;
|
|
encoded_image_1._frameType = VideoFrameType::kVideoFrameKey;
|
|
encoded_image_1.SetEncodedData(EncodedImageBuffer::Create(&kPayload, 1));
|
|
|
|
EncodedImage encoded_image_2(encoded_image_1);
|
|
encoded_image_2.SetSimulcastIndex(1);
|
|
|
|
RtpVideoSenderTestFixture test({kSsrc1, kSsrc2}, {kRtxSsrc1, kRtxSsrc2},
|
|
kPayloadType, {});
|
|
CodecSpecificInfo codec_info;
|
|
codec_info.codecType = kVideoCodecVP8;
|
|
|
|
// Setting rtp streams to inactive will turn the payload router to
|
|
// inactive.
|
|
test.SetSending(false);
|
|
// An incoming encoded image will not ask the module to send outgoing data
|
|
// because the payload router is inactive.
|
|
EXPECT_NE(EncodedImageCallback::Result::OK,
|
|
test.router()->OnEncodedImage(encoded_image_1, &codec_info).error);
|
|
EXPECT_NE(EncodedImageCallback::Result::OK,
|
|
test.router()->OnEncodedImage(encoded_image_2, &codec_info).error);
|
|
}
|
|
|
|
TEST(RtpVideoSenderTest,
|
|
DiscardsHigherSimulcastFramesAfterLayerDisabledInVideoLayersAllocation) {
|
|
constexpr uint8_t kPayload = 'a';
|
|
EncodedImage encoded_image_1;
|
|
encoded_image_1.SetRtpTimestamp(1);
|
|
encoded_image_1.capture_time_ms_ = 2;
|
|
encoded_image_1._frameType = VideoFrameType::kVideoFrameKey;
|
|
encoded_image_1.SetEncodedData(EncodedImageBuffer::Create(&kPayload, 1));
|
|
EncodedImage encoded_image_2(encoded_image_1);
|
|
encoded_image_2.SetSimulcastIndex(1);
|
|
CodecSpecificInfo codec_info;
|
|
codec_info.codecType = kVideoCodecVP8;
|
|
RtpVideoSenderTestFixture test({kSsrc1, kSsrc2}, {kRtxSsrc1, kRtxSsrc2},
|
|
kPayloadType, {});
|
|
test.SetSending(true);
|
|
// A layer is sent on both rtp streams.
|
|
test.router()->OnVideoLayersAllocationUpdated(
|
|
{.active_spatial_layers = {{.rtp_stream_index = 0},
|
|
{.rtp_stream_index = 1}}});
|
|
|
|
EXPECT_EQ(EncodedImageCallback::Result::OK,
|
|
test.router()->OnEncodedImage(encoded_image_1, &codec_info).error);
|
|
EXPECT_EQ(EncodedImageCallback::Result::OK,
|
|
test.router()->OnEncodedImage(encoded_image_2, &codec_info).error);
|
|
|
|
// Only rtp stream index 0 is configured to send a stream.
|
|
test.router()->OnVideoLayersAllocationUpdated(
|
|
{.active_spatial_layers = {{.rtp_stream_index = 0}}});
|
|
EXPECT_EQ(EncodedImageCallback::Result::OK,
|
|
test.router()->OnEncodedImage(encoded_image_1, &codec_info).error);
|
|
EXPECT_NE(EncodedImageCallback::Result::OK,
|
|
test.router()->OnEncodedImage(encoded_image_2, &codec_info).error);
|
|
}
|
|
|
|
TEST(RtpVideoSenderTest, CreateWithNoPreviousStates) {
|
|
RtpVideoSenderTestFixture test({kSsrc1, kSsrc2}, {kRtxSsrc1, kRtxSsrc2},
|
|
kPayloadType, {});
|
|
test.SetSending(true);
|
|
|
|
std::map<uint32_t, RtpPayloadState> initial_states =
|
|
test.router()->GetRtpPayloadStates();
|
|
EXPECT_EQ(2u, initial_states.size());
|
|
EXPECT_NE(initial_states.find(kSsrc1), initial_states.end());
|
|
EXPECT_NE(initial_states.find(kSsrc2), initial_states.end());
|
|
}
|
|
|
|
TEST(RtpVideoSenderTest, CreateWithPreviousStates) {
|
|
const int64_t kState1SharedFrameId = 123;
|
|
const int64_t kState2SharedFrameId = 234;
|
|
RtpPayloadState state1;
|
|
state1.picture_id = kInitialPictureId1;
|
|
state1.tl0_pic_idx = kInitialTl0PicIdx1;
|
|
state1.shared_frame_id = kState1SharedFrameId;
|
|
RtpPayloadState state2;
|
|
state2.picture_id = kInitialPictureId2;
|
|
state2.tl0_pic_idx = kInitialTl0PicIdx2;
|
|
state2.shared_frame_id = kState2SharedFrameId;
|
|
std::map<uint32_t, RtpPayloadState> states = {{kSsrc1, state1},
|
|
{kSsrc2, state2}};
|
|
|
|
RtpVideoSenderTestFixture test({kSsrc1, kSsrc2}, {kRtxSsrc1, kRtxSsrc2},
|
|
kPayloadType, states);
|
|
test.SetSending(true);
|
|
|
|
std::map<uint32_t, RtpPayloadState> initial_states =
|
|
test.router()->GetRtpPayloadStates();
|
|
EXPECT_EQ(2u, initial_states.size());
|
|
EXPECT_EQ(kInitialPictureId1, initial_states[kSsrc1].picture_id);
|
|
EXPECT_EQ(kInitialTl0PicIdx1, initial_states[kSsrc1].tl0_pic_idx);
|
|
EXPECT_EQ(kInitialPictureId2, initial_states[kSsrc2].picture_id);
|
|
EXPECT_EQ(kInitialTl0PicIdx2, initial_states[kSsrc2].tl0_pic_idx);
|
|
EXPECT_EQ(kState2SharedFrameId, initial_states[kSsrc1].shared_frame_id);
|
|
EXPECT_EQ(kState2SharedFrameId, initial_states[kSsrc2].shared_frame_id);
|
|
}
|
|
|
|
TEST(RtpVideoSenderTest, FrameCountCallbacks) {
|
|
class MockFrameCountObserver : public FrameCountObserver {
|
|
public:
|
|
MOCK_METHOD(void,
|
|
FrameCountUpdated,
|
|
(const FrameCounts& frame_counts, uint32_t ssrc),
|
|
(override));
|
|
} callback;
|
|
|
|
RtpVideoSenderTestFixture test({kSsrc1}, {kRtxSsrc1}, kPayloadType, {},
|
|
&callback);
|
|
|
|
constexpr uint8_t kPayload = 'a';
|
|
EncodedImage encoded_image;
|
|
encoded_image.SetRtpTimestamp(1);
|
|
encoded_image.capture_time_ms_ = 2;
|
|
encoded_image._frameType = VideoFrameType::kVideoFrameKey;
|
|
encoded_image.SetEncodedData(EncodedImageBuffer::Create(&kPayload, 1));
|
|
|
|
encoded_image._frameType = VideoFrameType::kVideoFrameKey;
|
|
|
|
// No callbacks when not active.
|
|
EXPECT_CALL(callback, FrameCountUpdated).Times(0);
|
|
EXPECT_NE(EncodedImageCallback::Result::OK,
|
|
test.router()->OnEncodedImage(encoded_image, nullptr).error);
|
|
::testing::Mock::VerifyAndClearExpectations(&callback);
|
|
|
|
test.SetSending(true);
|
|
|
|
FrameCounts frame_counts;
|
|
EXPECT_CALL(callback, FrameCountUpdated(_, kSsrc1))
|
|
.WillOnce(SaveArg<0>(&frame_counts));
|
|
EXPECT_EQ(EncodedImageCallback::Result::OK,
|
|
test.router()->OnEncodedImage(encoded_image, nullptr).error);
|
|
|
|
EXPECT_EQ(1, frame_counts.key_frames);
|
|
EXPECT_EQ(0, frame_counts.delta_frames);
|
|
|
|
::testing::Mock::VerifyAndClearExpectations(&callback);
|
|
|
|
encoded_image._frameType = VideoFrameType::kVideoFrameDelta;
|
|
EXPECT_CALL(callback, FrameCountUpdated(_, kSsrc1))
|
|
.WillOnce(SaveArg<0>(&frame_counts));
|
|
EXPECT_EQ(EncodedImageCallback::Result::OK,
|
|
test.router()->OnEncodedImage(encoded_image, nullptr).error);
|
|
|
|
EXPECT_EQ(1, frame_counts.key_frames);
|
|
EXPECT_EQ(1, frame_counts.delta_frames);
|
|
}
|
|
|
|
// Integration test verifying that ack of packet via TransportFeedback means
|
|
// that the packet is removed from RtpPacketHistory and won't be retransmitted
|
|
// again.
|
|
TEST(RtpVideoSenderTest, DoesNotRetrasmitAckedPackets) {
|
|
RtpVideoSenderTestFixture test({kSsrc1, kSsrc2}, {kRtxSsrc1, kRtxSsrc2},
|
|
kPayloadType, {});
|
|
test.SetSending(true);
|
|
|
|
constexpr uint8_t kPayload = 'a';
|
|
EncodedImage encoded_image;
|
|
encoded_image.SetRtpTimestamp(1);
|
|
encoded_image.capture_time_ms_ = 2;
|
|
encoded_image._frameType = VideoFrameType::kVideoFrameKey;
|
|
encoded_image.SetEncodedData(EncodedImageBuffer::Create(&kPayload, 1));
|
|
|
|
// Send two tiny images, mapping to two RTP packets. Capture sequence numbers.
|
|
std::vector<uint16_t> rtp_sequence_numbers;
|
|
std::vector<uint16_t> transport_sequence_numbers;
|
|
EXPECT_CALL(test.transport(), SendRtp)
|
|
.Times(2)
|
|
.WillRepeatedly([&rtp_sequence_numbers, &transport_sequence_numbers](
|
|
rtc::ArrayView<const uint8_t> packet,
|
|
const PacketOptions& options) {
|
|
RtpPacket rtp_packet;
|
|
EXPECT_TRUE(rtp_packet.Parse(packet));
|
|
rtp_sequence_numbers.push_back(rtp_packet.SequenceNumber());
|
|
transport_sequence_numbers.push_back(options.packet_id);
|
|
return true;
|
|
});
|
|
EXPECT_EQ(EncodedImageCallback::Result::OK,
|
|
test.router()->OnEncodedImage(encoded_image, nullptr).error);
|
|
encoded_image.SetRtpTimestamp(2);
|
|
encoded_image.capture_time_ms_ = 3;
|
|
EXPECT_EQ(EncodedImageCallback::Result::OK,
|
|
test.router()->OnEncodedImage(encoded_image, nullptr).error);
|
|
|
|
test.AdvanceTime(TimeDelta::Millis(33));
|
|
|
|
// Construct a NACK message for requesting retransmission of both packet.
|
|
rtcp::Nack nack;
|
|
nack.SetMediaSsrc(kSsrc1);
|
|
nack.SetPacketIds(rtp_sequence_numbers);
|
|
rtc::Buffer nack_buffer = nack.Build();
|
|
|
|
std::vector<uint16_t> retransmitted_rtp_sequence_numbers;
|
|
EXPECT_CALL(test.transport(), SendRtp)
|
|
.Times(2)
|
|
.WillRepeatedly([&retransmitted_rtp_sequence_numbers](
|
|
rtc::ArrayView<const uint8_t> packet,
|
|
const PacketOptions& options) {
|
|
RtpPacket rtp_packet;
|
|
EXPECT_TRUE(rtp_packet.Parse(packet));
|
|
EXPECT_EQ(rtp_packet.Ssrc(), kRtxSsrc1);
|
|
// Capture the retransmitted sequence number from the RTX header.
|
|
rtc::ArrayView<const uint8_t> payload = rtp_packet.payload();
|
|
retransmitted_rtp_sequence_numbers.push_back(
|
|
ByteReader<uint16_t>::ReadBigEndian(payload.data()));
|
|
return true;
|
|
});
|
|
test.router()->DeliverRtcp(nack_buffer.data(), nack_buffer.size());
|
|
test.AdvanceTime(TimeDelta::Millis(33));
|
|
|
|
// Verify that both packets were retransmitted.
|
|
EXPECT_EQ(retransmitted_rtp_sequence_numbers, rtp_sequence_numbers);
|
|
|
|
// Simulate transport feedback indicating fist packet received, next packet
|
|
// lost (not other way around as that would trigger early retransmit).
|
|
StreamFeedbackObserver::StreamPacketInfo lost_packet_feedback;
|
|
lost_packet_feedback.rtp_sequence_number = rtp_sequence_numbers[0];
|
|
lost_packet_feedback.ssrc = kSsrc1;
|
|
lost_packet_feedback.received = false;
|
|
lost_packet_feedback.is_retransmission = false;
|
|
|
|
StreamFeedbackObserver::StreamPacketInfo received_packet_feedback;
|
|
received_packet_feedback.rtp_sequence_number = rtp_sequence_numbers[1];
|
|
received_packet_feedback.ssrc = kSsrc1;
|
|
received_packet_feedback.received = true;
|
|
lost_packet_feedback.is_retransmission = false;
|
|
|
|
test.router()->OnPacketFeedbackVector(
|
|
{lost_packet_feedback, received_packet_feedback});
|
|
|
|
// Advance time to make sure retransmission would be allowed and try again.
|
|
// This time the retransmission should not happen for the first packet since
|
|
// the history has been notified of the ack and removed the packet. The
|
|
// second packet, included in the feedback but not marked as received, should
|
|
// still be retransmitted.
|
|
test.AdvanceTime(TimeDelta::Millis(33));
|
|
EXPECT_CALL(test.transport(), SendRtp)
|
|
.WillOnce([&lost_packet_feedback](rtc::ArrayView<const uint8_t> packet,
|
|
const PacketOptions& options) {
|
|
RtpPacket rtp_packet;
|
|
EXPECT_TRUE(rtp_packet.Parse(packet));
|
|
EXPECT_EQ(rtp_packet.Ssrc(), kRtxSsrc1);
|
|
// Capture the retransmitted sequence number from the RTX header.
|
|
rtc::ArrayView<const uint8_t> payload = rtp_packet.payload();
|
|
EXPECT_EQ(lost_packet_feedback.rtp_sequence_number,
|
|
ByteReader<uint16_t>::ReadBigEndian(payload.data()));
|
|
return true;
|
|
});
|
|
test.router()->DeliverRtcp(nack_buffer.data(), nack_buffer.size());
|
|
test.AdvanceTime(TimeDelta::Millis(33));
|
|
}
|
|
|
|
// This tests that we utilize transport wide feedback to retransmit lost
|
|
// packets. This is tested by dropping all ordinary packets from a "lossy"
|
|
// stream sent along with a secondary untouched stream. The transport wide
|
|
// feedback packets from the secondary stream allows the sending side to
|
|
// detect and retreansmit the lost packets from the lossy stream.
|
|
TEST(RtpVideoSenderTest, RetransmitsOnTransportWideLossInfo) {
|
|
int rtx_packets;
|
|
test::Scenario s(test_info_);
|
|
test::CallClientConfig call_conf;
|
|
// Keeping the bitrate fixed to avoid RTX due to probing.
|
|
call_conf.transport.rates.max_rate = DataRate::KilobitsPerSec(300);
|
|
call_conf.transport.rates.start_rate = DataRate::KilobitsPerSec(300);
|
|
test::NetworkSimulationConfig net_conf;
|
|
net_conf.bandwidth = DataRate::KilobitsPerSec(300);
|
|
auto send_node = s.CreateSimulationNode(net_conf);
|
|
auto* callee = s.CreateClient("return", call_conf);
|
|
auto* route = s.CreateRoutes(s.CreateClient("send", call_conf), {send_node},
|
|
callee, {s.CreateSimulationNode(net_conf)});
|
|
|
|
test::VideoStreamConfig lossy_config;
|
|
lossy_config.source.framerate = 5;
|
|
auto* lossy = s.CreateVideoStream(route->forward(), lossy_config);
|
|
// The secondary stream acts a driver for transport feedback messages,
|
|
// ensuring that lost packets on the lossy stream are retransmitted.
|
|
s.CreateVideoStream(route->forward(), test::VideoStreamConfig());
|
|
|
|
send_node->router()->SetFilter([&](const EmulatedIpPacket& packet) {
|
|
RtpPacket rtp;
|
|
if (rtp.Parse(packet.data)) {
|
|
// Drops all regular packets for the lossy stream and counts all RTX
|
|
// packets. Since no packets are let trough, NACKs can't be triggered
|
|
// by the receiving side.
|
|
if (lossy->send()->UsingSsrc(rtp.Ssrc())) {
|
|
return false;
|
|
} else if (lossy->send()->UsingRtxSsrc(rtp.Ssrc())) {
|
|
++rtx_packets;
|
|
}
|
|
}
|
|
return true;
|
|
});
|
|
|
|
// Run for a short duration and reset counters to avoid counting RTX packets
|
|
// from initial probing.
|
|
s.RunFor(TimeDelta::Seconds(1));
|
|
rtx_packets = 0;
|
|
int decoded_baseline = 0;
|
|
callee->SendTask([&decoded_baseline, &lossy]() {
|
|
decoded_baseline = lossy->receive()->GetStats().frames_decoded;
|
|
});
|
|
s.RunFor(TimeDelta::Seconds(1));
|
|
// We expect both that RTX packets were sent and that an appropriate number of
|
|
// frames were received. This is somewhat redundant but reduces the risk of
|
|
// false positives in future regressions (e.g. RTX is send due to probing).
|
|
EXPECT_GE(rtx_packets, 1);
|
|
int frames_decoded = 0;
|
|
callee->SendTask([&decoded_baseline, &frames_decoded, &lossy]() {
|
|
frames_decoded =
|
|
lossy->receive()->GetStats().frames_decoded - decoded_baseline;
|
|
});
|
|
EXPECT_EQ(frames_decoded, 5);
|
|
}
|
|
|
|
// Integration test verifying that retransmissions are sent for packets which
|
|
// can be detected as lost early, using transport wide feedback.
|
|
TEST(RtpVideoSenderTest, EarlyRetransmits) {
|
|
RtpVideoSenderTestFixture test({kSsrc1, kSsrc2}, {kRtxSsrc1, kRtxSsrc2},
|
|
kPayloadType, {});
|
|
test.SetSending(true);
|
|
|
|
const uint8_t kPayload[1] = {'a'};
|
|
EncodedImage encoded_image;
|
|
encoded_image.SetRtpTimestamp(1);
|
|
encoded_image.capture_time_ms_ = 2;
|
|
encoded_image._frameType = VideoFrameType::kVideoFrameKey;
|
|
encoded_image.SetEncodedData(
|
|
EncodedImageBuffer::Create(kPayload, sizeof(kPayload)));
|
|
encoded_image.SetSimulcastIndex(0);
|
|
|
|
CodecSpecificInfo codec_specific;
|
|
codec_specific.codecType = VideoCodecType::kVideoCodecGeneric;
|
|
|
|
// Send two tiny images, mapping to single RTP packets. Capture sequence
|
|
// numbers.
|
|
uint16_t frame1_rtp_sequence_number = 0;
|
|
uint16_t frame1_transport_sequence_number = 0;
|
|
EXPECT_CALL(test.transport(), SendRtp)
|
|
.WillOnce(
|
|
[&frame1_rtp_sequence_number, &frame1_transport_sequence_number](
|
|
rtc::ArrayView<const uint8_t> packet,
|
|
const PacketOptions& options) {
|
|
RtpPacket rtp_packet;
|
|
EXPECT_TRUE(rtp_packet.Parse(packet));
|
|
frame1_rtp_sequence_number = rtp_packet.SequenceNumber();
|
|
frame1_transport_sequence_number = options.packet_id;
|
|
EXPECT_EQ(rtp_packet.Ssrc(), kSsrc1);
|
|
return true;
|
|
});
|
|
EXPECT_EQ(test.router()->OnEncodedImage(encoded_image, &codec_specific).error,
|
|
EncodedImageCallback::Result::OK);
|
|
|
|
test.AdvanceTime(TimeDelta::Millis(33));
|
|
|
|
uint16_t frame2_rtp_sequence_number = 0;
|
|
uint16_t frame2_transport_sequence_number = 0;
|
|
encoded_image.SetSimulcastIndex(1);
|
|
EXPECT_CALL(test.transport(), SendRtp)
|
|
.WillOnce(
|
|
[&frame2_rtp_sequence_number, &frame2_transport_sequence_number](
|
|
rtc::ArrayView<const uint8_t> packet,
|
|
const PacketOptions& options) {
|
|
RtpPacket rtp_packet;
|
|
EXPECT_TRUE(rtp_packet.Parse(packet));
|
|
frame2_rtp_sequence_number = rtp_packet.SequenceNumber();
|
|
frame2_transport_sequence_number = options.packet_id;
|
|
EXPECT_EQ(rtp_packet.Ssrc(), kSsrc2);
|
|
return true;
|
|
});
|
|
EXPECT_EQ(test.router()->OnEncodedImage(encoded_image, &codec_specific).error,
|
|
EncodedImageCallback::Result::OK);
|
|
test.AdvanceTime(TimeDelta::Millis(33));
|
|
|
|
EXPECT_NE(frame1_transport_sequence_number, frame2_transport_sequence_number);
|
|
|
|
// Inject a transport feedback where the packet for the first frame is lost,
|
|
// expect a retransmission for it.
|
|
EXPECT_CALL(test.transport(), SendRtp)
|
|
.WillOnce([&frame1_rtp_sequence_number](
|
|
rtc::ArrayView<const uint8_t> packet,
|
|
const PacketOptions& options) {
|
|
RtpPacket rtp_packet;
|
|
EXPECT_TRUE(rtp_packet.Parse(packet));
|
|
EXPECT_EQ(rtp_packet.Ssrc(), kRtxSsrc1);
|
|
|
|
// Retransmitted sequence number from the RTX header should match
|
|
// the lost packet.
|
|
rtc::ArrayView<const uint8_t> payload = rtp_packet.payload();
|
|
EXPECT_EQ(ByteReader<uint16_t>::ReadBigEndian(payload.data()),
|
|
frame1_rtp_sequence_number);
|
|
return true;
|
|
});
|
|
|
|
StreamFeedbackObserver::StreamPacketInfo first_packet_feedback;
|
|
first_packet_feedback.rtp_sequence_number = frame1_rtp_sequence_number;
|
|
first_packet_feedback.ssrc = kSsrc1;
|
|
first_packet_feedback.received = false;
|
|
first_packet_feedback.is_retransmission = false;
|
|
|
|
StreamFeedbackObserver::StreamPacketInfo second_packet_feedback;
|
|
second_packet_feedback.rtp_sequence_number = frame2_rtp_sequence_number;
|
|
second_packet_feedback.ssrc = kSsrc2;
|
|
second_packet_feedback.received = true;
|
|
first_packet_feedback.is_retransmission = false;
|
|
|
|
test.router()->OnPacketFeedbackVector(
|
|
{first_packet_feedback, second_packet_feedback});
|
|
|
|
// Wait for pacer to run and send the RTX packet.
|
|
test.AdvanceTime(TimeDelta::Millis(33));
|
|
}
|
|
|
|
TEST(RtpVideoSenderTest, SupportsDependencyDescriptor) {
|
|
RtpVideoSenderTestFixture test({kSsrc1}, {}, kPayloadType, {});
|
|
test.SetSending(true);
|
|
|
|
RtpHeaderExtensionMap extensions;
|
|
extensions.Register<RtpDependencyDescriptorExtension>(
|
|
kDependencyDescriptorExtensionId);
|
|
std::vector<RtpPacket> sent_packets;
|
|
ON_CALL(test.transport(), SendRtp)
|
|
.WillByDefault([&](rtc::ArrayView<const uint8_t> packet,
|
|
const PacketOptions& options) {
|
|
sent_packets.emplace_back(&extensions);
|
|
EXPECT_TRUE(sent_packets.back().Parse(packet));
|
|
return true;
|
|
});
|
|
|
|
const uint8_t kPayload[1] = {'a'};
|
|
EncodedImage encoded_image;
|
|
encoded_image.SetRtpTimestamp(1);
|
|
encoded_image.capture_time_ms_ = 2;
|
|
encoded_image.SetEncodedData(
|
|
EncodedImageBuffer::Create(kPayload, sizeof(kPayload)));
|
|
|
|
CodecSpecificInfo codec_specific;
|
|
codec_specific.codecType = VideoCodecType::kVideoCodecGeneric;
|
|
codec_specific.template_structure.emplace();
|
|
codec_specific.template_structure->num_decode_targets = 1;
|
|
codec_specific.template_structure->templates = {
|
|
FrameDependencyTemplate().T(0).Dtis("S"),
|
|
FrameDependencyTemplate().T(0).Dtis("S").FrameDiffs({2}),
|
|
FrameDependencyTemplate().T(1).Dtis("D").FrameDiffs({1}),
|
|
};
|
|
|
|
// Send two tiny images, mapping to single RTP packets.
|
|
// Send in key frame.
|
|
encoded_image._frameType = VideoFrameType::kVideoFrameKey;
|
|
codec_specific.generic_frame_info =
|
|
GenericFrameInfo::Builder().T(0).Dtis("S").Build();
|
|
codec_specific.generic_frame_info->encoder_buffers = {{0, false, true}};
|
|
EXPECT_EQ(test.router()->OnEncodedImage(encoded_image, &codec_specific).error,
|
|
EncodedImageCallback::Result::OK);
|
|
test.AdvanceTime(TimeDelta::Millis(33));
|
|
ASSERT_THAT(sent_packets, SizeIs(1));
|
|
EXPECT_TRUE(
|
|
sent_packets.back().HasExtension<RtpDependencyDescriptorExtension>());
|
|
|
|
// Send in delta frame.
|
|
encoded_image._frameType = VideoFrameType::kVideoFrameDelta;
|
|
codec_specific.template_structure = absl::nullopt;
|
|
codec_specific.generic_frame_info =
|
|
GenericFrameInfo::Builder().T(1).Dtis("D").Build();
|
|
codec_specific.generic_frame_info->encoder_buffers = {{0, true, false}};
|
|
EXPECT_EQ(test.router()->OnEncodedImage(encoded_image, &codec_specific).error,
|
|
EncodedImageCallback::Result::OK);
|
|
test.AdvanceTime(TimeDelta::Millis(33));
|
|
ASSERT_THAT(sent_packets, SizeIs(2));
|
|
EXPECT_TRUE(
|
|
sent_packets.back().HasExtension<RtpDependencyDescriptorExtension>());
|
|
}
|
|
|
|
TEST(RtpVideoSenderTest, SimulcastIndependentFrameIds) {
|
|
test::ExplicitKeyValueConfig field_trials(
|
|
"WebRTC-Video-SimulcastIndependentFrameIds/Enabled/"
|
|
"WebRTC-GenericDescriptorAuth/Disabled/");
|
|
const std::map<uint32_t, RtpPayloadState> kPayloadStates = {
|
|
{kSsrc1, {.frame_id = 100}}, {kSsrc2, {.frame_id = 200}}};
|
|
RtpVideoSenderTestFixture test({kSsrc1, kSsrc2}, {}, kPayloadType,
|
|
kPayloadStates, &field_trials);
|
|
test.SetSending(true);
|
|
|
|
RtpHeaderExtensionMap extensions;
|
|
extensions.Register<RtpDependencyDescriptorExtension>(
|
|
kDependencyDescriptorExtensionId);
|
|
std::vector<RtpPacket> sent_packets;
|
|
ON_CALL(test.transport(), SendRtp)
|
|
.WillByDefault([&](rtc::ArrayView<const uint8_t> packet,
|
|
const PacketOptions& options) {
|
|
sent_packets.emplace_back(&extensions);
|
|
EXPECT_TRUE(sent_packets.back().Parse(packet));
|
|
return true;
|
|
});
|
|
|
|
const uint8_t kPayload[1] = {'a'};
|
|
EncodedImage encoded_image;
|
|
encoded_image.SetEncodedData(
|
|
EncodedImageBuffer::Create(kPayload, sizeof(kPayload)));
|
|
|
|
CodecSpecificInfo codec_specific;
|
|
codec_specific.codecType = VideoCodecType::kVideoCodecGeneric;
|
|
codec_specific.template_structure.emplace();
|
|
codec_specific.template_structure->num_decode_targets = 1;
|
|
codec_specific.template_structure->templates = {
|
|
FrameDependencyTemplate().T(0).Dtis("S"),
|
|
FrameDependencyTemplate().T(0).Dtis("S").FrameDiffs({1}),
|
|
};
|
|
codec_specific.generic_frame_info =
|
|
GenericFrameInfo::Builder().T(0).Dtis("S").Build();
|
|
encoded_image._frameType = VideoFrameType::kVideoFrameKey;
|
|
codec_specific.generic_frame_info->encoder_buffers = {{0, false, true}};
|
|
|
|
encoded_image.SetSimulcastIndex(0);
|
|
EXPECT_EQ(test.router()->OnEncodedImage(encoded_image, &codec_specific).error,
|
|
EncodedImageCallback::Result::OK);
|
|
encoded_image.SetSimulcastIndex(1);
|
|
EXPECT_EQ(test.router()->OnEncodedImage(encoded_image, &codec_specific).error,
|
|
EncodedImageCallback::Result::OK);
|
|
|
|
test.AdvanceTime(TimeDelta::Millis(33));
|
|
ASSERT_THAT(sent_packets, SizeIs(2));
|
|
DependencyDescriptorMandatory dd_s0;
|
|
DependencyDescriptorMandatory dd_s1;
|
|
ASSERT_TRUE(
|
|
sent_packets[0].GetExtension<RtpDependencyDescriptorExtension>(&dd_s0));
|
|
ASSERT_TRUE(
|
|
sent_packets[1].GetExtension<RtpDependencyDescriptorExtension>(&dd_s1));
|
|
EXPECT_EQ(dd_s0.frame_number(), 100);
|
|
EXPECT_EQ(dd_s1.frame_number(), 200);
|
|
}
|
|
|
|
TEST(RtpVideoSenderTest,
|
|
SimulcastNoIndependentFrameIdsIfGenericDescriptorAuthIsEnabled) {
|
|
test::ExplicitKeyValueConfig field_trials(
|
|
"WebRTC-Video-SimulcastIndependentFrameIds/Enabled/"
|
|
"WebRTC-GenericDescriptorAuth/Enabled/");
|
|
const std::map<uint32_t, RtpPayloadState> kPayloadStates = {
|
|
{kSsrc1, {.shared_frame_id = 1000, .frame_id = 100}},
|
|
{kSsrc2, {.shared_frame_id = 1000, .frame_id = 200}}};
|
|
RtpVideoSenderTestFixture test({kSsrc1, kSsrc2}, {}, kPayloadType,
|
|
kPayloadStates, &field_trials);
|
|
test.SetSending(true);
|
|
|
|
RtpHeaderExtensionMap extensions;
|
|
extensions.Register<RtpDependencyDescriptorExtension>(
|
|
kDependencyDescriptorExtensionId);
|
|
std::vector<RtpPacket> sent_packets;
|
|
ON_CALL(test.transport(), SendRtp)
|
|
.WillByDefault([&](rtc::ArrayView<const uint8_t> packet,
|
|
const PacketOptions& options) {
|
|
sent_packets.emplace_back(&extensions);
|
|
EXPECT_TRUE(sent_packets.back().Parse(packet));
|
|
return true;
|
|
});
|
|
|
|
const uint8_t kPayload[1] = {'a'};
|
|
EncodedImage encoded_image;
|
|
encoded_image.SetEncodedData(
|
|
EncodedImageBuffer::Create(kPayload, sizeof(kPayload)));
|
|
|
|
CodecSpecificInfo codec_specific;
|
|
codec_specific.codecType = VideoCodecType::kVideoCodecGeneric;
|
|
codec_specific.template_structure.emplace();
|
|
codec_specific.template_structure->num_decode_targets = 1;
|
|
codec_specific.template_structure->templates = {
|
|
FrameDependencyTemplate().T(0).Dtis("S"),
|
|
FrameDependencyTemplate().T(0).Dtis("S").FrameDiffs({1}),
|
|
};
|
|
codec_specific.generic_frame_info =
|
|
GenericFrameInfo::Builder().T(0).Dtis("S").Build();
|
|
encoded_image._frameType = VideoFrameType::kVideoFrameKey;
|
|
codec_specific.generic_frame_info->encoder_buffers = {{0, false, true}};
|
|
|
|
encoded_image.SetSimulcastIndex(0);
|
|
EXPECT_EQ(test.router()->OnEncodedImage(encoded_image, &codec_specific).error,
|
|
EncodedImageCallback::Result::OK);
|
|
encoded_image.SetSimulcastIndex(1);
|
|
EXPECT_EQ(test.router()->OnEncodedImage(encoded_image, &codec_specific).error,
|
|
EncodedImageCallback::Result::OK);
|
|
|
|
test.AdvanceTime(TimeDelta::Millis(33));
|
|
ASSERT_THAT(sent_packets, SizeIs(2));
|
|
DependencyDescriptorMandatory dd_s0;
|
|
DependencyDescriptorMandatory dd_s1;
|
|
ASSERT_TRUE(
|
|
sent_packets[0].GetExtension<RtpDependencyDescriptorExtension>(&dd_s0));
|
|
ASSERT_TRUE(
|
|
sent_packets[1].GetExtension<RtpDependencyDescriptorExtension>(&dd_s1));
|
|
EXPECT_EQ(dd_s0.frame_number(), 1001);
|
|
EXPECT_EQ(dd_s1.frame_number(), 1002);
|
|
}
|
|
|
|
TEST(RtpVideoSenderTest,
|
|
SupportsDependencyDescriptorForVp8NotProvidedByEncoder) {
|
|
constexpr uint8_t kPayload[1] = {'a'};
|
|
RtpVideoSenderTestFixture test({kSsrc1}, {}, kPayloadType, {});
|
|
RtpHeaderExtensionMap extensions;
|
|
extensions.Register<RtpDependencyDescriptorExtension>(
|
|
kDependencyDescriptorExtensionId);
|
|
std::vector<RtpPacket> sent_packets;
|
|
ON_CALL(test.transport(), SendRtp)
|
|
.WillByDefault(
|
|
[&](rtc::ArrayView<const uint8_t> packet, const PacketOptions&) {
|
|
EXPECT_TRUE(sent_packets.emplace_back(&extensions).Parse(packet));
|
|
return true;
|
|
});
|
|
test.SetSending(true);
|
|
|
|
EncodedImage key_frame_image;
|
|
key_frame_image._frameType = VideoFrameType::kVideoFrameKey;
|
|
key_frame_image.SetEncodedData(
|
|
EncodedImageBuffer::Create(kPayload, sizeof(kPayload)));
|
|
CodecSpecificInfo key_frame_info;
|
|
key_frame_info.codecType = VideoCodecType::kVideoCodecVP8;
|
|
ASSERT_EQ(
|
|
test.router()->OnEncodedImage(key_frame_image, &key_frame_info).error,
|
|
EncodedImageCallback::Result::OK);
|
|
|
|
EncodedImage delta_image;
|
|
delta_image._frameType = VideoFrameType::kVideoFrameDelta;
|
|
delta_image.SetEncodedData(
|
|
EncodedImageBuffer::Create(kPayload, sizeof(kPayload)));
|
|
CodecSpecificInfo delta_info;
|
|
delta_info.codecType = VideoCodecType::kVideoCodecVP8;
|
|
ASSERT_EQ(test.router()->OnEncodedImage(delta_image, &delta_info).error,
|
|
EncodedImageCallback::Result::OK);
|
|
|
|
test.AdvanceTime(TimeDelta::Millis(123));
|
|
|
|
DependencyDescriptor key_frame_dd;
|
|
DependencyDescriptor delta_dd;
|
|
ASSERT_THAT(sent_packets, SizeIs(2));
|
|
EXPECT_TRUE(sent_packets[0].GetExtension<RtpDependencyDescriptorExtension>(
|
|
/*structure=*/nullptr, &key_frame_dd));
|
|
EXPECT_TRUE(sent_packets[1].GetExtension<RtpDependencyDescriptorExtension>(
|
|
key_frame_dd.attached_structure.get(), &delta_dd));
|
|
}
|
|
|
|
TEST(RtpVideoSenderTest, SupportsDependencyDescriptorForVp9) {
|
|
RtpVideoSenderTestFixture test({kSsrc1}, {}, kPayloadType, {});
|
|
test.SetSending(true);
|
|
|
|
RtpHeaderExtensionMap extensions;
|
|
extensions.Register<RtpDependencyDescriptorExtension>(
|
|
kDependencyDescriptorExtensionId);
|
|
std::vector<RtpPacket> sent_packets;
|
|
ON_CALL(test.transport(), SendRtp)
|
|
.WillByDefault([&](rtc::ArrayView<const uint8_t> packet,
|
|
const PacketOptions& options) {
|
|
sent_packets.emplace_back(&extensions);
|
|
EXPECT_TRUE(sent_packets.back().Parse(packet));
|
|
return true;
|
|
});
|
|
|
|
const uint8_t kPayload[1] = {'a'};
|
|
EncodedImage encoded_image;
|
|
encoded_image.SetRtpTimestamp(1);
|
|
encoded_image.capture_time_ms_ = 2;
|
|
encoded_image._frameType = VideoFrameType::kVideoFrameKey;
|
|
encoded_image.SetEncodedData(
|
|
EncodedImageBuffer::Create(kPayload, sizeof(kPayload)));
|
|
|
|
CodecSpecificInfo codec_specific;
|
|
codec_specific.codecType = VideoCodecType::kVideoCodecVP9;
|
|
codec_specific.template_structure.emplace();
|
|
codec_specific.template_structure->num_decode_targets = 2;
|
|
codec_specific.template_structure->templates = {
|
|
FrameDependencyTemplate().S(0).Dtis("SS"),
|
|
FrameDependencyTemplate().S(1).Dtis("-S").FrameDiffs({1}),
|
|
};
|
|
|
|
// Send two tiny images, each mapping to single RTP packet.
|
|
// Send in key frame for the base spatial layer.
|
|
codec_specific.generic_frame_info =
|
|
GenericFrameInfo::Builder().S(0).Dtis("SS").Build();
|
|
codec_specific.generic_frame_info->encoder_buffers = {{0, false, true}};
|
|
EXPECT_EQ(test.router()->OnEncodedImage(encoded_image, &codec_specific).error,
|
|
EncodedImageCallback::Result::OK);
|
|
// Send in 2nd spatial layer.
|
|
codec_specific.template_structure = absl::nullopt;
|
|
codec_specific.generic_frame_info =
|
|
GenericFrameInfo::Builder().S(1).Dtis("-S").Build();
|
|
codec_specific.generic_frame_info->encoder_buffers = {{0, true, false},
|
|
{1, false, true}};
|
|
EXPECT_EQ(test.router()->OnEncodedImage(encoded_image, &codec_specific).error,
|
|
EncodedImageCallback::Result::OK);
|
|
|
|
test.AdvanceTime(TimeDelta::Millis(33));
|
|
ASSERT_THAT(sent_packets, SizeIs(2));
|
|
EXPECT_TRUE(sent_packets[0].HasExtension<RtpDependencyDescriptorExtension>());
|
|
EXPECT_TRUE(sent_packets[1].HasExtension<RtpDependencyDescriptorExtension>());
|
|
}
|
|
|
|
TEST(RtpVideoSenderTest,
|
|
SupportsDependencyDescriptorForVp9NotProvidedByEncoder) {
|
|
RtpVideoSenderTestFixture test({kSsrc1}, {}, kPayloadType, {});
|
|
test.SetSending(true);
|
|
|
|
RtpHeaderExtensionMap extensions;
|
|
extensions.Register<RtpDependencyDescriptorExtension>(
|
|
kDependencyDescriptorExtensionId);
|
|
std::vector<RtpPacket> sent_packets;
|
|
ON_CALL(test.transport(), SendRtp)
|
|
.WillByDefault([&](rtc::ArrayView<const uint8_t> packet,
|
|
const PacketOptions& options) {
|
|
sent_packets.emplace_back(&extensions);
|
|
EXPECT_TRUE(sent_packets.back().Parse(packet));
|
|
return true;
|
|
});
|
|
|
|
const uint8_t kPayload[1] = {'a'};
|
|
EncodedImage encoded_image;
|
|
encoded_image.SetRtpTimestamp(1);
|
|
encoded_image.capture_time_ms_ = 2;
|
|
encoded_image._frameType = VideoFrameType::kVideoFrameKey;
|
|
encoded_image._encodedWidth = 320;
|
|
encoded_image._encodedHeight = 180;
|
|
encoded_image.SetEncodedData(
|
|
EncodedImageBuffer::Create(kPayload, sizeof(kPayload)));
|
|
|
|
CodecSpecificInfo codec_specific;
|
|
codec_specific.codecType = VideoCodecType::kVideoCodecVP9;
|
|
codec_specific.codecSpecific.VP9.num_spatial_layers = 1;
|
|
codec_specific.codecSpecific.VP9.temporal_idx = kNoTemporalIdx;
|
|
codec_specific.codecSpecific.VP9.first_frame_in_picture = true;
|
|
codec_specific.end_of_picture = true;
|
|
codec_specific.codecSpecific.VP9.inter_pic_predicted = false;
|
|
|
|
// Send two tiny images, each mapping to single RTP packet.
|
|
EXPECT_EQ(test.router()->OnEncodedImage(encoded_image, &codec_specific).error,
|
|
EncodedImageCallback::Result::OK);
|
|
|
|
// Send in 2nd picture.
|
|
encoded_image._frameType = VideoFrameType::kVideoFrameDelta;
|
|
encoded_image.SetRtpTimestamp(3000);
|
|
codec_specific.codecSpecific.VP9.inter_pic_predicted = true;
|
|
codec_specific.codecSpecific.VP9.num_ref_pics = 1;
|
|
codec_specific.codecSpecific.VP9.p_diff[0] = 1;
|
|
EXPECT_EQ(test.router()->OnEncodedImage(encoded_image, &codec_specific).error,
|
|
EncodedImageCallback::Result::OK);
|
|
|
|
test.AdvanceTime(TimeDelta::Millis(33));
|
|
ASSERT_THAT(sent_packets, SizeIs(2));
|
|
EXPECT_TRUE(sent_packets[0].HasExtension<RtpDependencyDescriptorExtension>());
|
|
EXPECT_TRUE(sent_packets[1].HasExtension<RtpDependencyDescriptorExtension>());
|
|
}
|
|
|
|
TEST(RtpVideoSenderTest, GenerateDependecyDescriptorForGenericCodecs) {
|
|
test::ScopedKeyValueConfig field_trials(
|
|
"WebRTC-GenericCodecDependencyDescriptor/Enabled/");
|
|
RtpVideoSenderTestFixture test({kSsrc1}, {}, kPayloadType, {}, &field_trials);
|
|
test.SetSending(true);
|
|
|
|
RtpHeaderExtensionMap extensions;
|
|
extensions.Register<RtpDependencyDescriptorExtension>(
|
|
kDependencyDescriptorExtensionId);
|
|
std::vector<RtpPacket> sent_packets;
|
|
ON_CALL(test.transport(), SendRtp)
|
|
.WillByDefault([&](rtc::ArrayView<const uint8_t> packet,
|
|
const PacketOptions& options) {
|
|
sent_packets.emplace_back(&extensions);
|
|
EXPECT_TRUE(sent_packets.back().Parse(packet));
|
|
return true;
|
|
});
|
|
|
|
const uint8_t kPayload[1] = {'a'};
|
|
EncodedImage encoded_image;
|
|
encoded_image.SetRtpTimestamp(1);
|
|
encoded_image.capture_time_ms_ = 2;
|
|
encoded_image._frameType = VideoFrameType::kVideoFrameKey;
|
|
encoded_image._encodedWidth = 320;
|
|
encoded_image._encodedHeight = 180;
|
|
encoded_image.SetEncodedData(
|
|
EncodedImageBuffer::Create(kPayload, sizeof(kPayload)));
|
|
|
|
CodecSpecificInfo codec_specific;
|
|
codec_specific.codecType = VideoCodecType::kVideoCodecGeneric;
|
|
codec_specific.end_of_picture = true;
|
|
|
|
// Send two tiny images, each mapping to single RTP packet.
|
|
EXPECT_EQ(test.router()->OnEncodedImage(encoded_image, &codec_specific).error,
|
|
EncodedImageCallback::Result::OK);
|
|
|
|
// Send in 2nd picture.
|
|
encoded_image._frameType = VideoFrameType::kVideoFrameDelta;
|
|
encoded_image.SetRtpTimestamp(3000);
|
|
EXPECT_EQ(test.router()->OnEncodedImage(encoded_image, &codec_specific).error,
|
|
EncodedImageCallback::Result::OK);
|
|
|
|
test.AdvanceTime(TimeDelta::Millis(33));
|
|
ASSERT_THAT(sent_packets, SizeIs(2));
|
|
EXPECT_TRUE(sent_packets[0].HasExtension<RtpDependencyDescriptorExtension>());
|
|
EXPECT_TRUE(sent_packets[1].HasExtension<RtpDependencyDescriptorExtension>());
|
|
}
|
|
|
|
TEST(RtpVideoSenderTest, SupportsStoppingUsingDependencyDescriptor) {
|
|
RtpVideoSenderTestFixture test({kSsrc1}, {}, kPayloadType, {});
|
|
test.SetSending(true);
|
|
|
|
RtpHeaderExtensionMap extensions;
|
|
extensions.Register<RtpDependencyDescriptorExtension>(
|
|
kDependencyDescriptorExtensionId);
|
|
std::vector<RtpPacket> sent_packets;
|
|
ON_CALL(test.transport(), SendRtp)
|
|
.WillByDefault([&](rtc::ArrayView<const uint8_t> packet,
|
|
const PacketOptions& options) {
|
|
sent_packets.emplace_back(&extensions);
|
|
EXPECT_TRUE(sent_packets.back().Parse(packet));
|
|
return true;
|
|
});
|
|
|
|
const uint8_t kPayload[1] = {'a'};
|
|
EncodedImage encoded_image;
|
|
encoded_image.SetRtpTimestamp(1);
|
|
encoded_image.capture_time_ms_ = 2;
|
|
encoded_image.SetEncodedData(
|
|
EncodedImageBuffer::Create(kPayload, sizeof(kPayload)));
|
|
|
|
CodecSpecificInfo codec_specific;
|
|
codec_specific.codecType = VideoCodecType::kVideoCodecGeneric;
|
|
codec_specific.template_structure.emplace();
|
|
codec_specific.template_structure->num_decode_targets = 1;
|
|
codec_specific.template_structure->templates = {
|
|
FrameDependencyTemplate().T(0).Dtis("S"),
|
|
FrameDependencyTemplate().T(0).Dtis("S").FrameDiffs({2}),
|
|
FrameDependencyTemplate().T(1).Dtis("D").FrameDiffs({1}),
|
|
};
|
|
|
|
// Send two tiny images, mapping to single RTP packets.
|
|
// Send in a key frame.
|
|
encoded_image._frameType = VideoFrameType::kVideoFrameKey;
|
|
codec_specific.generic_frame_info =
|
|
GenericFrameInfo::Builder().T(0).Dtis("S").Build();
|
|
codec_specific.generic_frame_info->encoder_buffers = {{0, false, true}};
|
|
EXPECT_EQ(test.router()->OnEncodedImage(encoded_image, &codec_specific).error,
|
|
EncodedImageCallback::Result::OK);
|
|
test.AdvanceTime(TimeDelta::Millis(33));
|
|
ASSERT_THAT(sent_packets, SizeIs(1));
|
|
EXPECT_TRUE(
|
|
sent_packets.back().HasExtension<RtpDependencyDescriptorExtension>());
|
|
|
|
// Send in a new key frame without the support for the dependency descriptor.
|
|
encoded_image._frameType = VideoFrameType::kVideoFrameKey;
|
|
codec_specific.template_structure = absl::nullopt;
|
|
EXPECT_EQ(test.router()->OnEncodedImage(encoded_image, &codec_specific).error,
|
|
EncodedImageCallback::Result::OK);
|
|
test.AdvanceTime(TimeDelta::Millis(33));
|
|
ASSERT_THAT(sent_packets, SizeIs(2));
|
|
EXPECT_FALSE(
|
|
sent_packets.back().HasExtension<RtpDependencyDescriptorExtension>());
|
|
}
|
|
|
|
TEST(RtpVideoSenderTest, CanSetZeroBitrate) {
|
|
RtpVideoSenderTestFixture test({kSsrc1}, {kRtxSsrc1}, kPayloadType, {});
|
|
test.router()->OnBitrateUpdated(CreateBitrateAllocationUpdate(0),
|
|
/*framerate*/ 0);
|
|
}
|
|
|
|
TEST(RtpVideoSenderTest, SimulcastSenderRegistersFrameTransformers) {
|
|
rtc::scoped_refptr<MockFrameTransformer> transformer =
|
|
rtc::make_ref_counted<MockFrameTransformer>();
|
|
|
|
EXPECT_CALL(*transformer, RegisterTransformedFrameSinkCallback(_, kSsrc1));
|
|
EXPECT_CALL(*transformer, RegisterTransformedFrameSinkCallback(_, kSsrc2));
|
|
RtpVideoSenderTestFixture test({kSsrc1, kSsrc2}, {kRtxSsrc1, kRtxSsrc2},
|
|
kPayloadType, {}, nullptr, transformer);
|
|
|
|
EXPECT_CALL(*transformer, UnregisterTransformedFrameSinkCallback(kSsrc1));
|
|
EXPECT_CALL(*transformer, UnregisterTransformedFrameSinkCallback(kSsrc2));
|
|
}
|
|
|
|
TEST(RtpVideoSenderTest, OverheadIsSubtractedFromTargetBitrate) {
|
|
test::ScopedKeyValueConfig field_trials(
|
|
"WebRTC-Video-UseFrameRateForOverhead/Enabled/");
|
|
|
|
// TODO(jakobi): RTP header size should not be hard coded.
|
|
constexpr uint32_t kRtpHeaderSizeBytes = 20;
|
|
constexpr uint32_t kTransportPacketOverheadBytes = 40;
|
|
constexpr uint32_t kOverheadPerPacketBytes =
|
|
kRtpHeaderSizeBytes + kTransportPacketOverheadBytes;
|
|
RtpVideoSenderTestFixture test({kSsrc1}, {}, kPayloadType, {}, &field_trials);
|
|
test.router()->OnTransportOverheadChanged(kTransportPacketOverheadBytes);
|
|
test.SetSending(true);
|
|
|
|
{
|
|
test.router()->OnBitrateUpdated(CreateBitrateAllocationUpdate(300000),
|
|
/*framerate*/ 15);
|
|
// 1 packet per frame.
|
|
EXPECT_EQ(test.router()->GetPayloadBitrateBps(),
|
|
300000 - kOverheadPerPacketBytes * 8 * 30);
|
|
}
|
|
{
|
|
test.router()->OnBitrateUpdated(CreateBitrateAllocationUpdate(150000),
|
|
/*framerate*/ 15);
|
|
// 1 packet per frame.
|
|
EXPECT_EQ(test.router()->GetPayloadBitrateBps(),
|
|
150000 - kOverheadPerPacketBytes * 8 * 15);
|
|
}
|
|
{
|
|
test.router()->OnBitrateUpdated(CreateBitrateAllocationUpdate(1000000),
|
|
/*framerate*/ 30);
|
|
// 3 packets per frame.
|
|
EXPECT_EQ(test.router()->GetPayloadBitrateBps(),
|
|
1000000 - kOverheadPerPacketBytes * 8 * 30 * 3);
|
|
}
|
|
}
|
|
|
|
TEST(RtpVideoSenderTest, ClearsPendingPacketsOnInactivation) {
|
|
RtpVideoSenderTestFixture test({kSsrc1}, {kRtxSsrc1}, kPayloadType, {});
|
|
test.SetSending(true);
|
|
|
|
RtpHeaderExtensionMap extensions;
|
|
extensions.Register<RtpDependencyDescriptorExtension>(
|
|
kDependencyDescriptorExtensionId);
|
|
std::vector<RtpPacket> sent_packets;
|
|
ON_CALL(test.transport(), SendRtp)
|
|
.WillByDefault([&](rtc::ArrayView<const uint8_t> packet,
|
|
const PacketOptions& options) {
|
|
sent_packets.emplace_back(&extensions);
|
|
EXPECT_TRUE(sent_packets.back().Parse(packet));
|
|
return true;
|
|
});
|
|
|
|
// Set a very low bitrate.
|
|
test.router()->OnBitrateUpdated(
|
|
CreateBitrateAllocationUpdate(/*rate_bps=*/10'000),
|
|
/*framerate=*/30);
|
|
|
|
// Create and send a large keyframe.
|
|
const size_t kImageSizeBytes = 10000;
|
|
constexpr uint8_t kPayload[kImageSizeBytes] = {'a'};
|
|
EncodedImage encoded_image;
|
|
encoded_image.SetRtpTimestamp(1);
|
|
encoded_image.capture_time_ms_ = 2;
|
|
encoded_image._frameType = VideoFrameType::kVideoFrameKey;
|
|
encoded_image.SetEncodedData(
|
|
EncodedImageBuffer::Create(kPayload, sizeof(kPayload)));
|
|
EXPECT_EQ(test.router()
|
|
->OnEncodedImage(encoded_image, /*codec_specific=*/nullptr)
|
|
.error,
|
|
EncodedImageCallback::Result::OK);
|
|
|
|
// Advance time a small amount, check that sent data is only part of the
|
|
// image.
|
|
test.AdvanceTime(TimeDelta::Millis(5));
|
|
DataSize transmittedPayload = DataSize::Zero();
|
|
for (const RtpPacket& packet : sent_packets) {
|
|
transmittedPayload += DataSize::Bytes(packet.payload_size());
|
|
// Make sure we don't see the end of the frame.
|
|
EXPECT_FALSE(packet.Marker());
|
|
}
|
|
EXPECT_GT(transmittedPayload, DataSize::Zero());
|
|
EXPECT_LT(transmittedPayload, DataSize::Bytes(kImageSizeBytes / 3));
|
|
|
|
// Record the RTP timestamp of the first frame.
|
|
const uint32_t first_frame_timestamp = sent_packets[0].Timestamp();
|
|
sent_packets.clear();
|
|
|
|
// Disable the sending module and advance time slightly. No packets should be
|
|
// sent.
|
|
test.SetSending(false);
|
|
test.AdvanceTime(TimeDelta::Millis(20));
|
|
EXPECT_TRUE(sent_packets.empty());
|
|
|
|
// Reactive the send module - any packets should have been removed, so nothing
|
|
// should be transmitted.
|
|
test.SetSending(true);
|
|
test.AdvanceTime(TimeDelta::Millis(33));
|
|
EXPECT_TRUE(sent_packets.empty());
|
|
|
|
// Send a new frame.
|
|
encoded_image.SetRtpTimestamp(3);
|
|
encoded_image.capture_time_ms_ = 4;
|
|
EXPECT_EQ(test.router()
|
|
->OnEncodedImage(encoded_image, /*codec_specific=*/nullptr)
|
|
.error,
|
|
EncodedImageCallback::Result::OK);
|
|
test.AdvanceTime(TimeDelta::Millis(33));
|
|
|
|
// Advance time, check we get new packets - but only for the second frame.
|
|
EXPECT_FALSE(sent_packets.empty());
|
|
EXPECT_NE(sent_packets[0].Timestamp(), first_frame_timestamp);
|
|
}
|
|
|
|
// Integration test verifying that when retransmission mode is set to
|
|
// kRetransmitBaseLayer,only base layer is retransmitted.
|
|
TEST(RtpVideoSenderTest, RetransmitsBaseLayerOnly) {
|
|
RtpVideoSenderTestFixture test({kSsrc1, kSsrc2}, {kRtxSsrc1, kRtxSsrc2},
|
|
kPayloadType, {});
|
|
test.SetSending(true);
|
|
|
|
test.router()->SetRetransmissionMode(kRetransmitBaseLayer);
|
|
constexpr uint8_t kPayload = 'a';
|
|
EncodedImage encoded_image;
|
|
encoded_image.SetRtpTimestamp(1);
|
|
encoded_image.capture_time_ms_ = 2;
|
|
encoded_image._frameType = VideoFrameType::kVideoFrameKey;
|
|
encoded_image.SetEncodedData(EncodedImageBuffer::Create(&kPayload, 1));
|
|
|
|
// Send two tiny images, mapping to two RTP packets. Capture sequence numbers.
|
|
std::vector<uint16_t> rtp_sequence_numbers;
|
|
std::vector<uint16_t> transport_sequence_numbers;
|
|
std::vector<uint16_t> base_sequence_numbers;
|
|
EXPECT_CALL(test.transport(), SendRtp)
|
|
.Times(2)
|
|
.WillRepeatedly([&rtp_sequence_numbers, &transport_sequence_numbers](
|
|
rtc::ArrayView<const uint8_t> packet,
|
|
const PacketOptions& options) {
|
|
RtpPacket rtp_packet;
|
|
EXPECT_TRUE(rtp_packet.Parse(packet));
|
|
rtp_sequence_numbers.push_back(rtp_packet.SequenceNumber());
|
|
transport_sequence_numbers.push_back(options.packet_id);
|
|
return true;
|
|
});
|
|
CodecSpecificInfo key_codec_info;
|
|
key_codec_info.codecType = kVideoCodecVP8;
|
|
key_codec_info.codecSpecific.VP8.temporalIdx = 0;
|
|
EXPECT_EQ(EncodedImageCallback::Result::OK,
|
|
test.router()->OnEncodedImage(
|
|
encoded_image, &key_codec_info).error);
|
|
encoded_image.SetRtpTimestamp(2);
|
|
encoded_image.capture_time_ms_ = 3;
|
|
encoded_image._frameType = VideoFrameType::kVideoFrameDelta;
|
|
CodecSpecificInfo delta_codec_info;
|
|
delta_codec_info.codecType = kVideoCodecVP8;
|
|
delta_codec_info.codecSpecific.VP8.temporalIdx = 1;
|
|
EXPECT_EQ(EncodedImageCallback::Result::OK,
|
|
test.router()->OnEncodedImage(
|
|
encoded_image, &delta_codec_info).error);
|
|
|
|
test.AdvanceTime(TimeDelta::Millis(33));
|
|
|
|
// Construct a NACK message for requesting retransmission of both packet.
|
|
rtcp::Nack nack;
|
|
nack.SetMediaSsrc(kSsrc1);
|
|
nack.SetPacketIds(rtp_sequence_numbers);
|
|
rtc::Buffer nack_buffer = nack.Build();
|
|
|
|
std::vector<uint16_t> retransmitted_rtp_sequence_numbers;
|
|
EXPECT_CALL(test.transport(), SendRtp)
|
|
.Times(1)
|
|
.WillRepeatedly([&retransmitted_rtp_sequence_numbers](
|
|
rtc::ArrayView<const uint8_t> packet,
|
|
const PacketOptions& options) {
|
|
RtpPacket rtp_packet;
|
|
EXPECT_TRUE(rtp_packet.Parse(packet));
|
|
EXPECT_EQ(rtp_packet.Ssrc(), kRtxSsrc1);
|
|
// Capture the retransmitted sequence number from the RTX header.
|
|
rtc::ArrayView<const uint8_t> payload = rtp_packet.payload();
|
|
retransmitted_rtp_sequence_numbers.push_back(
|
|
ByteReader<uint16_t>::ReadBigEndian(payload.data()));
|
|
return true;
|
|
});
|
|
test.router()->DeliverRtcp(nack_buffer.data(), nack_buffer.size());
|
|
test.AdvanceTime(TimeDelta::Millis(33));
|
|
|
|
// Verify that only base layer packet was retransmitted.
|
|
std::vector<uint16_t> base_rtp_sequence_numbers(rtp_sequence_numbers.begin(),
|
|
rtp_sequence_numbers.begin() + 1);
|
|
EXPECT_EQ(retransmitted_rtp_sequence_numbers, base_rtp_sequence_numbers);
|
|
}
|
|
|
|
} // namespace webrtc
|