mirror of
https://github.com/mollyim/webrtc.git
synced 2025-05-13 13:50:40 +01:00

Because a single Resource only has a single ResourceListener, injected Resources only gets wired up to the stream's ResourceAdaptationProcessor that was last to call SetResourceListener. This could potentially lead to the relevant stream not adapting based on the injected resource because it got wired up to the wrong stream's processor. This CL fixes this issue by introducing BroadcastResourceListener. By listening to 1 resource (the injected one), it can spawn N "adapter" resources that mirror's the injected resource's usage signal, allowing all ResourceAdaptationProcessor's to react to the signal. This is wired up in Call, and tests are updated to verify the signal gets through. Bug: chromium:1101263, webrtc:11720 Change-Id: I8a37284cb9a68f08ca1bdb1ee050b7144c451297 Reviewed-on: https://webrtc-review.googlesource.com/c/src/+/178386 Reviewed-by: Stefan Holmer <stefan@webrtc.org> Reviewed-by: Evan Shrubsole <eshr@google.com> Commit-Queue: Henrik Boström <hbos@webrtc.org> Cr-Commit-Position: refs/heads/master@{#31612}
1474 lines
56 KiB
C++
1474 lines
56 KiB
C++
/*
|
|
* Copyright (c) 2013 The WebRTC project authors. All Rights Reserved.
|
|
*
|
|
* Use of this source code is governed by a BSD-style license
|
|
* that can be found in the LICENSE file in the root of the source
|
|
* tree. An additional intellectual property rights grant can be found
|
|
* in the file PATENTS. All contributing project authors may
|
|
* be found in the AUTHORS file in the root of the source tree.
|
|
*/
|
|
|
|
#include "call/call.h"
|
|
|
|
#include <string.h>
|
|
|
|
#include <algorithm>
|
|
#include <map>
|
|
#include <memory>
|
|
#include <set>
|
|
#include <utility>
|
|
#include <vector>
|
|
|
|
#include "absl/types/optional.h"
|
|
#include "api/rtc_event_log/rtc_event_log.h"
|
|
#include "api/transport/network_control.h"
|
|
#include "audio/audio_receive_stream.h"
|
|
#include "audio/audio_send_stream.h"
|
|
#include "audio/audio_state.h"
|
|
#include "call/adaptation/broadcast_resource_listener.h"
|
|
#include "call/bitrate_allocator.h"
|
|
#include "call/flexfec_receive_stream_impl.h"
|
|
#include "call/receive_time_calculator.h"
|
|
#include "call/rtp_stream_receiver_controller.h"
|
|
#include "call/rtp_transport_controller_send.h"
|
|
#include "logging/rtc_event_log/events/rtc_event_audio_receive_stream_config.h"
|
|
#include "logging/rtc_event_log/events/rtc_event_rtcp_packet_incoming.h"
|
|
#include "logging/rtc_event_log/events/rtc_event_rtp_packet_incoming.h"
|
|
#include "logging/rtc_event_log/events/rtc_event_video_receive_stream_config.h"
|
|
#include "logging/rtc_event_log/events/rtc_event_video_send_stream_config.h"
|
|
#include "logging/rtc_event_log/rtc_stream_config.h"
|
|
#include "modules/congestion_controller/include/receive_side_congestion_controller.h"
|
|
#include "modules/rtp_rtcp/include/flexfec_receiver.h"
|
|
#include "modules/rtp_rtcp/include/rtp_header_extension_map.h"
|
|
#include "modules/rtp_rtcp/source/byte_io.h"
|
|
#include "modules/rtp_rtcp/source/rtp_packet_received.h"
|
|
#include "modules/rtp_rtcp/source/rtp_utility.h"
|
|
#include "modules/utility/include/process_thread.h"
|
|
#include "modules/video_coding/fec_controller_default.h"
|
|
#include "rtc_base/checks.h"
|
|
#include "rtc_base/constructor_magic.h"
|
|
#include "rtc_base/location.h"
|
|
#include "rtc_base/logging.h"
|
|
#include "rtc_base/strings/string_builder.h"
|
|
#include "rtc_base/synchronization/sequence_checker.h"
|
|
#include "rtc_base/task_utils/pending_task_safety_flag.h"
|
|
#include "rtc_base/thread_annotations.h"
|
|
#include "rtc_base/time_utils.h"
|
|
#include "rtc_base/trace_event.h"
|
|
#include "system_wrappers/include/clock.h"
|
|
#include "system_wrappers/include/cpu_info.h"
|
|
#include "system_wrappers/include/field_trial.h"
|
|
#include "system_wrappers/include/metrics.h"
|
|
#include "video/call_stats2.h"
|
|
#include "video/send_delay_stats.h"
|
|
#include "video/stats_counter.h"
|
|
#include "video/video_receive_stream2.h"
|
|
#include "video/video_send_stream.h"
|
|
|
|
namespace webrtc {
|
|
|
|
namespace {
|
|
bool SendPeriodicFeedback(const std::vector<RtpExtension>& extensions) {
|
|
for (const auto& extension : extensions) {
|
|
if (extension.uri == RtpExtension::kTransportSequenceNumberV2Uri)
|
|
return false;
|
|
}
|
|
return true;
|
|
}
|
|
|
|
// TODO(nisse): This really begs for a shared context struct.
|
|
bool UseSendSideBwe(const std::vector<RtpExtension>& extensions,
|
|
bool transport_cc) {
|
|
if (!transport_cc)
|
|
return false;
|
|
for (const auto& extension : extensions) {
|
|
if (extension.uri == RtpExtension::kTransportSequenceNumberUri ||
|
|
extension.uri == RtpExtension::kTransportSequenceNumberV2Uri)
|
|
return true;
|
|
}
|
|
return false;
|
|
}
|
|
|
|
bool UseSendSideBwe(const VideoReceiveStream::Config& config) {
|
|
return UseSendSideBwe(config.rtp.extensions, config.rtp.transport_cc);
|
|
}
|
|
|
|
bool UseSendSideBwe(const AudioReceiveStream::Config& config) {
|
|
return UseSendSideBwe(config.rtp.extensions, config.rtp.transport_cc);
|
|
}
|
|
|
|
bool UseSendSideBwe(const FlexfecReceiveStream::Config& config) {
|
|
return UseSendSideBwe(config.rtp_header_extensions, config.transport_cc);
|
|
}
|
|
|
|
const int* FindKeyByValue(const std::map<int, int>& m, int v) {
|
|
for (const auto& kv : m) {
|
|
if (kv.second == v)
|
|
return &kv.first;
|
|
}
|
|
return nullptr;
|
|
}
|
|
|
|
std::unique_ptr<rtclog::StreamConfig> CreateRtcLogStreamConfig(
|
|
const VideoReceiveStream::Config& config) {
|
|
auto rtclog_config = std::make_unique<rtclog::StreamConfig>();
|
|
rtclog_config->remote_ssrc = config.rtp.remote_ssrc;
|
|
rtclog_config->local_ssrc = config.rtp.local_ssrc;
|
|
rtclog_config->rtx_ssrc = config.rtp.rtx_ssrc;
|
|
rtclog_config->rtcp_mode = config.rtp.rtcp_mode;
|
|
rtclog_config->rtp_extensions = config.rtp.extensions;
|
|
|
|
for (const auto& d : config.decoders) {
|
|
const int* search =
|
|
FindKeyByValue(config.rtp.rtx_associated_payload_types, d.payload_type);
|
|
rtclog_config->codecs.emplace_back(d.video_format.name, d.payload_type,
|
|
search ? *search : 0);
|
|
}
|
|
return rtclog_config;
|
|
}
|
|
|
|
std::unique_ptr<rtclog::StreamConfig> CreateRtcLogStreamConfig(
|
|
const VideoSendStream::Config& config,
|
|
size_t ssrc_index) {
|
|
auto rtclog_config = std::make_unique<rtclog::StreamConfig>();
|
|
rtclog_config->local_ssrc = config.rtp.ssrcs[ssrc_index];
|
|
if (ssrc_index < config.rtp.rtx.ssrcs.size()) {
|
|
rtclog_config->rtx_ssrc = config.rtp.rtx.ssrcs[ssrc_index];
|
|
}
|
|
rtclog_config->rtcp_mode = config.rtp.rtcp_mode;
|
|
rtclog_config->rtp_extensions = config.rtp.extensions;
|
|
|
|
rtclog_config->codecs.emplace_back(config.rtp.payload_name,
|
|
config.rtp.payload_type,
|
|
config.rtp.rtx.payload_type);
|
|
return rtclog_config;
|
|
}
|
|
|
|
std::unique_ptr<rtclog::StreamConfig> CreateRtcLogStreamConfig(
|
|
const AudioReceiveStream::Config& config) {
|
|
auto rtclog_config = std::make_unique<rtclog::StreamConfig>();
|
|
rtclog_config->remote_ssrc = config.rtp.remote_ssrc;
|
|
rtclog_config->local_ssrc = config.rtp.local_ssrc;
|
|
rtclog_config->rtp_extensions = config.rtp.extensions;
|
|
return rtclog_config;
|
|
}
|
|
|
|
bool IsRtcp(const uint8_t* packet, size_t length) {
|
|
RtpUtility::RtpHeaderParser rtp_parser(packet, length);
|
|
return rtp_parser.RTCP();
|
|
}
|
|
|
|
TaskQueueBase* GetCurrentTaskQueueOrThread() {
|
|
TaskQueueBase* current = TaskQueueBase::Current();
|
|
if (!current)
|
|
current = rtc::ThreadManager::Instance()->CurrentThread();
|
|
return current;
|
|
}
|
|
|
|
} // namespace
|
|
|
|
namespace internal {
|
|
|
|
// Wraps an injected resource in a BroadcastResourceListener and handles adding
|
|
// and removing adapter resources to individual VideoSendStreams.
|
|
class ResourceVideoSendStreamForwarder {
|
|
public:
|
|
ResourceVideoSendStreamForwarder(
|
|
rtc::scoped_refptr<webrtc::Resource> resource)
|
|
: broadcast_resource_listener_(resource) {
|
|
broadcast_resource_listener_.StartListening();
|
|
}
|
|
~ResourceVideoSendStreamForwarder() {
|
|
RTC_DCHECK(adapter_resources_.empty());
|
|
broadcast_resource_listener_.StopListening();
|
|
}
|
|
|
|
rtc::scoped_refptr<webrtc::Resource> Resource() const {
|
|
return broadcast_resource_listener_.SourceResource();
|
|
}
|
|
|
|
void OnCreateVideoSendStream(VideoSendStream* video_send_stream) {
|
|
RTC_DCHECK(adapter_resources_.find(video_send_stream) ==
|
|
adapter_resources_.end());
|
|
auto adapter_resource =
|
|
broadcast_resource_listener_.CreateAdapterResource();
|
|
video_send_stream->AddAdaptationResource(adapter_resource);
|
|
adapter_resources_.insert(
|
|
std::make_pair(video_send_stream, adapter_resource));
|
|
}
|
|
|
|
void OnDestroyVideoSendStream(VideoSendStream* video_send_stream) {
|
|
auto it = adapter_resources_.find(video_send_stream);
|
|
RTC_DCHECK(it != adapter_resources_.end());
|
|
broadcast_resource_listener_.RemoveAdapterResource(it->second);
|
|
adapter_resources_.erase(it);
|
|
}
|
|
|
|
private:
|
|
BroadcastResourceListener broadcast_resource_listener_;
|
|
std::map<VideoSendStream*, rtc::scoped_refptr<webrtc::Resource>>
|
|
adapter_resources_;
|
|
};
|
|
|
|
class Call final : public webrtc::Call,
|
|
public PacketReceiver,
|
|
public RecoveredPacketReceiver,
|
|
public TargetTransferRateObserver,
|
|
public BitrateAllocator::LimitObserver {
|
|
public:
|
|
Call(Clock* clock,
|
|
const Call::Config& config,
|
|
std::unique_ptr<RtpTransportControllerSendInterface> transport_send,
|
|
rtc::scoped_refptr<SharedModuleThread> module_process_thread,
|
|
TaskQueueFactory* task_queue_factory);
|
|
~Call() override;
|
|
|
|
// Implements webrtc::Call.
|
|
PacketReceiver* Receiver() override;
|
|
|
|
webrtc::AudioSendStream* CreateAudioSendStream(
|
|
const webrtc::AudioSendStream::Config& config) override;
|
|
void DestroyAudioSendStream(webrtc::AudioSendStream* send_stream) override;
|
|
|
|
webrtc::AudioReceiveStream* CreateAudioReceiveStream(
|
|
const webrtc::AudioReceiveStream::Config& config) override;
|
|
void DestroyAudioReceiveStream(
|
|
webrtc::AudioReceiveStream* receive_stream) override;
|
|
|
|
webrtc::VideoSendStream* CreateVideoSendStream(
|
|
webrtc::VideoSendStream::Config config,
|
|
VideoEncoderConfig encoder_config) override;
|
|
webrtc::VideoSendStream* CreateVideoSendStream(
|
|
webrtc::VideoSendStream::Config config,
|
|
VideoEncoderConfig encoder_config,
|
|
std::unique_ptr<FecController> fec_controller) override;
|
|
void DestroyVideoSendStream(webrtc::VideoSendStream* send_stream) override;
|
|
|
|
webrtc::VideoReceiveStream* CreateVideoReceiveStream(
|
|
webrtc::VideoReceiveStream::Config configuration) override;
|
|
void DestroyVideoReceiveStream(
|
|
webrtc::VideoReceiveStream* receive_stream) override;
|
|
|
|
FlexfecReceiveStream* CreateFlexfecReceiveStream(
|
|
const FlexfecReceiveStream::Config& config) override;
|
|
void DestroyFlexfecReceiveStream(
|
|
FlexfecReceiveStream* receive_stream) override;
|
|
|
|
void AddAdaptationResource(rtc::scoped_refptr<Resource> resource) override;
|
|
|
|
RtpTransportControllerSendInterface* GetTransportControllerSend() override;
|
|
|
|
Stats GetStats() const override;
|
|
|
|
// Implements PacketReceiver.
|
|
DeliveryStatus DeliverPacket(MediaType media_type,
|
|
rtc::CopyOnWriteBuffer packet,
|
|
int64_t packet_time_us) override;
|
|
|
|
// Implements RecoveredPacketReceiver.
|
|
void OnRecoveredPacket(const uint8_t* packet, size_t length) override;
|
|
|
|
void SignalChannelNetworkState(MediaType media, NetworkState state) override;
|
|
|
|
void OnAudioTransportOverheadChanged(
|
|
int transport_overhead_per_packet) override;
|
|
|
|
void OnSentPacket(const rtc::SentPacket& sent_packet) override;
|
|
|
|
// Implements TargetTransferRateObserver,
|
|
void OnTargetTransferRate(TargetTransferRate msg) override;
|
|
void OnStartRateUpdate(DataRate start_rate) override;
|
|
|
|
// Implements BitrateAllocator::LimitObserver.
|
|
void OnAllocationLimitsChanged(BitrateAllocationLimits limits) override;
|
|
|
|
void SetClientBitratePreferences(const BitrateSettings& preferences) override;
|
|
|
|
private:
|
|
DeliveryStatus DeliverRtcp(MediaType media_type,
|
|
const uint8_t* packet,
|
|
size_t length)
|
|
RTC_EXCLUSIVE_LOCKS_REQUIRED(worker_thread_);
|
|
DeliveryStatus DeliverRtp(MediaType media_type,
|
|
rtc::CopyOnWriteBuffer packet,
|
|
int64_t packet_time_us)
|
|
RTC_EXCLUSIVE_LOCKS_REQUIRED(worker_thread_);
|
|
void ConfigureSync(const std::string& sync_group)
|
|
RTC_EXCLUSIVE_LOCKS_REQUIRED(worker_thread_);
|
|
|
|
void NotifyBweOfReceivedPacket(const RtpPacketReceived& packet,
|
|
MediaType media_type)
|
|
RTC_SHARED_LOCKS_REQUIRED(worker_thread_);
|
|
|
|
void UpdateSendHistograms(Timestamp first_sent_packet)
|
|
RTC_EXCLUSIVE_LOCKS_REQUIRED(worker_thread_);
|
|
void UpdateReceiveHistograms();
|
|
void UpdateHistograms();
|
|
void UpdateAggregateNetworkState();
|
|
|
|
void RegisterRateObserver();
|
|
|
|
rtc::TaskQueue* send_transport_queue() const {
|
|
return transport_send_ptr_->GetWorkerQueue();
|
|
}
|
|
|
|
Clock* const clock_;
|
|
TaskQueueFactory* const task_queue_factory_;
|
|
TaskQueueBase* const worker_thread_;
|
|
|
|
const int num_cpu_cores_;
|
|
const rtc::scoped_refptr<SharedModuleThread> module_process_thread_;
|
|
const std::unique_ptr<CallStats> call_stats_;
|
|
const std::unique_ptr<BitrateAllocator> bitrate_allocator_;
|
|
Call::Config config_;
|
|
|
|
NetworkState audio_network_state_;
|
|
NetworkState video_network_state_;
|
|
bool aggregate_network_up_ RTC_GUARDED_BY(worker_thread_);
|
|
|
|
// Audio, Video, and FlexFEC receive streams are owned by the client that
|
|
// creates them.
|
|
std::set<AudioReceiveStream*> audio_receive_streams_
|
|
RTC_GUARDED_BY(worker_thread_);
|
|
std::set<VideoReceiveStream2*> video_receive_streams_
|
|
RTC_GUARDED_BY(worker_thread_);
|
|
|
|
std::map<std::string, AudioReceiveStream*> sync_stream_mapping_
|
|
RTC_GUARDED_BY(worker_thread_);
|
|
|
|
// TODO(nisse): Should eventually be injected at creation,
|
|
// with a single object in the bundled case.
|
|
RtpStreamReceiverController audio_receiver_controller_;
|
|
RtpStreamReceiverController video_receiver_controller_;
|
|
|
|
// This extra map is used for receive processing which is
|
|
// independent of media type.
|
|
|
|
// TODO(nisse): In the RTP transport refactoring, we should have a
|
|
// single mapping from ssrc to a more abstract receive stream, with
|
|
// accessor methods for all configuration we need at this level.
|
|
struct ReceiveRtpConfig {
|
|
explicit ReceiveRtpConfig(const webrtc::AudioReceiveStream::Config& config)
|
|
: extensions(config.rtp.extensions),
|
|
use_send_side_bwe(UseSendSideBwe(config)) {}
|
|
explicit ReceiveRtpConfig(const webrtc::VideoReceiveStream::Config& config)
|
|
: extensions(config.rtp.extensions),
|
|
use_send_side_bwe(UseSendSideBwe(config)) {}
|
|
explicit ReceiveRtpConfig(const FlexfecReceiveStream::Config& config)
|
|
: extensions(config.rtp_header_extensions),
|
|
use_send_side_bwe(UseSendSideBwe(config)) {}
|
|
|
|
// Registered RTP header extensions for each stream. Note that RTP header
|
|
// extensions are negotiated per track ("m= line") in the SDP, but we have
|
|
// no notion of tracks at the Call level. We therefore store the RTP header
|
|
// extensions per SSRC instead, which leads to some storage overhead.
|
|
const RtpHeaderExtensionMap extensions;
|
|
// Set if both RTP extension the RTCP feedback message needed for
|
|
// send side BWE are negotiated.
|
|
const bool use_send_side_bwe;
|
|
};
|
|
std::map<uint32_t, ReceiveRtpConfig> receive_rtp_config_
|
|
RTC_GUARDED_BY(worker_thread_);
|
|
|
|
// Audio and Video send streams are owned by the client that creates them.
|
|
std::map<uint32_t, AudioSendStream*> audio_send_ssrcs_
|
|
RTC_GUARDED_BY(worker_thread_);
|
|
std::map<uint32_t, VideoSendStream*> video_send_ssrcs_
|
|
RTC_GUARDED_BY(worker_thread_);
|
|
std::set<VideoSendStream*> video_send_streams_ RTC_GUARDED_BY(worker_thread_);
|
|
|
|
// Each forwarder wraps an adaptation resource that was added to the call.
|
|
std::vector<std::unique_ptr<ResourceVideoSendStreamForwarder>>
|
|
adaptation_resource_forwarders_ RTC_GUARDED_BY(worker_thread_);
|
|
|
|
using RtpStateMap = std::map<uint32_t, RtpState>;
|
|
RtpStateMap suspended_audio_send_ssrcs_ RTC_GUARDED_BY(worker_thread_);
|
|
RtpStateMap suspended_video_send_ssrcs_ RTC_GUARDED_BY(worker_thread_);
|
|
|
|
using RtpPayloadStateMap = std::map<uint32_t, RtpPayloadState>;
|
|
RtpPayloadStateMap suspended_video_payload_states_
|
|
RTC_GUARDED_BY(worker_thread_);
|
|
|
|
webrtc::RtcEventLog* event_log_;
|
|
|
|
// The following members are only accessed (exclusively) from one thread and
|
|
// from the destructor, and therefore doesn't need any explicit
|
|
// synchronization.
|
|
RateCounter received_bytes_per_second_counter_;
|
|
RateCounter received_audio_bytes_per_second_counter_;
|
|
RateCounter received_video_bytes_per_second_counter_;
|
|
RateCounter received_rtcp_bytes_per_second_counter_;
|
|
absl::optional<int64_t> first_received_rtp_audio_ms_;
|
|
absl::optional<int64_t> last_received_rtp_audio_ms_;
|
|
absl::optional<int64_t> first_received_rtp_video_ms_;
|
|
absl::optional<int64_t> last_received_rtp_video_ms_;
|
|
|
|
uint32_t last_bandwidth_bps_ RTC_GUARDED_BY(worker_thread_);
|
|
// TODO(holmer): Remove this lock once BitrateController no longer calls
|
|
// OnNetworkChanged from multiple threads.
|
|
uint32_t min_allocated_send_bitrate_bps_ RTC_GUARDED_BY(worker_thread_);
|
|
uint32_t configured_max_padding_bitrate_bps_ RTC_GUARDED_BY(worker_thread_);
|
|
AvgCounter estimated_send_bitrate_kbps_counter_
|
|
RTC_GUARDED_BY(worker_thread_);
|
|
AvgCounter pacer_bitrate_kbps_counter_ RTC_GUARDED_BY(worker_thread_);
|
|
|
|
ReceiveSideCongestionController receive_side_cc_;
|
|
|
|
const std::unique_ptr<ReceiveTimeCalculator> receive_time_calculator_;
|
|
|
|
const std::unique_ptr<SendDelayStats> video_send_delay_stats_;
|
|
const int64_t start_ms_;
|
|
|
|
// Note that |task_safety_| needs to be at a greater scope than the task queue
|
|
// owned by |transport_send_| since calls might arrive on the network thread
|
|
// while Call is being deleted and the task queue is being torn down.
|
|
ScopedTaskSafety task_safety_;
|
|
|
|
// Caches transport_send_.get(), to avoid racing with destructor.
|
|
// Note that this is declared before transport_send_ to ensure that it is not
|
|
// invalidated until no more tasks can be running on the transport_send_ task
|
|
// queue.
|
|
RtpTransportControllerSendInterface* const transport_send_ptr_;
|
|
// Declared last since it will issue callbacks from a task queue. Declaring it
|
|
// last ensures that it is destroyed first and any running tasks are finished.
|
|
std::unique_ptr<RtpTransportControllerSendInterface> transport_send_;
|
|
|
|
bool is_target_rate_observer_registered_ RTC_GUARDED_BY(worker_thread_) =
|
|
false;
|
|
|
|
RTC_DISALLOW_COPY_AND_ASSIGN(Call);
|
|
};
|
|
} // namespace internal
|
|
|
|
std::string Call::Stats::ToString(int64_t time_ms) const {
|
|
char buf[1024];
|
|
rtc::SimpleStringBuilder ss(buf);
|
|
ss << "Call stats: " << time_ms << ", {";
|
|
ss << "send_bw_bps: " << send_bandwidth_bps << ", ";
|
|
ss << "recv_bw_bps: " << recv_bandwidth_bps << ", ";
|
|
ss << "max_pad_bps: " << max_padding_bitrate_bps << ", ";
|
|
ss << "pacer_delay_ms: " << pacer_delay_ms << ", ";
|
|
ss << "rtt_ms: " << rtt_ms;
|
|
ss << '}';
|
|
return ss.str();
|
|
}
|
|
|
|
Call* Call::Create(const Call::Config& config) {
|
|
rtc::scoped_refptr<SharedModuleThread> call_thread =
|
|
SharedModuleThread::Create(ProcessThread::Create("ModuleProcessThread"),
|
|
nullptr);
|
|
return Create(config, std::move(call_thread));
|
|
}
|
|
|
|
Call* Call::Create(const Call::Config& config,
|
|
rtc::scoped_refptr<SharedModuleThread> call_thread) {
|
|
return Create(config, Clock::GetRealTimeClock(), std::move(call_thread),
|
|
ProcessThread::Create("PacerThread"));
|
|
}
|
|
|
|
Call* Call::Create(const Call::Config& config,
|
|
Clock* clock,
|
|
rtc::scoped_refptr<SharedModuleThread> call_thread,
|
|
std::unique_ptr<ProcessThread> pacer_thread) {
|
|
RTC_DCHECK(config.task_queue_factory);
|
|
return new internal::Call(
|
|
clock, config,
|
|
std::make_unique<RtpTransportControllerSend>(
|
|
clock, config.event_log, config.network_state_predictor_factory,
|
|
config.network_controller_factory, config.bitrate_config,
|
|
std::move(pacer_thread), config.task_queue_factory, config.trials),
|
|
std::move(call_thread), config.task_queue_factory);
|
|
}
|
|
|
|
class SharedModuleThread::Impl {
|
|
public:
|
|
Impl(std::unique_ptr<ProcessThread> process_thread,
|
|
std::function<void()> on_one_ref_remaining)
|
|
: module_thread_(std::move(process_thread)),
|
|
on_one_ref_remaining_(std::move(on_one_ref_remaining)) {}
|
|
|
|
void EnsureStarted() {
|
|
RTC_DCHECK_RUN_ON(&sequence_checker_);
|
|
if (started_)
|
|
return;
|
|
started_ = true;
|
|
module_thread_->Start();
|
|
}
|
|
|
|
ProcessThread* process_thread() {
|
|
RTC_DCHECK_RUN_ON(&sequence_checker_);
|
|
return module_thread_.get();
|
|
}
|
|
|
|
void AddRef() const {
|
|
RTC_DCHECK_RUN_ON(&sequence_checker_);
|
|
++ref_count_;
|
|
}
|
|
|
|
rtc::RefCountReleaseStatus Release() const {
|
|
RTC_DCHECK_RUN_ON(&sequence_checker_);
|
|
--ref_count_;
|
|
|
|
if (ref_count_ == 0) {
|
|
module_thread_->Stop();
|
|
return rtc::RefCountReleaseStatus::kDroppedLastRef;
|
|
}
|
|
|
|
if (ref_count_ == 1 && on_one_ref_remaining_) {
|
|
auto moved_fn = std::move(on_one_ref_remaining_);
|
|
// NOTE: after this function returns, chances are that |this| has been
|
|
// deleted - do not touch any member variables.
|
|
// If the owner of the last reference implements a lambda that releases
|
|
// that last reference inside of the callback (which is legal according
|
|
// to this implementation), we will recursively enter Release() above,
|
|
// call Stop() and release the last reference.
|
|
moved_fn();
|
|
}
|
|
|
|
return rtc::RefCountReleaseStatus::kOtherRefsRemained;
|
|
}
|
|
|
|
private:
|
|
SequenceChecker sequence_checker_;
|
|
mutable int ref_count_ RTC_GUARDED_BY(sequence_checker_) = 0;
|
|
std::unique_ptr<ProcessThread> const module_thread_;
|
|
std::function<void()> const on_one_ref_remaining_;
|
|
bool started_ = false;
|
|
};
|
|
|
|
SharedModuleThread::SharedModuleThread(
|
|
std::unique_ptr<ProcessThread> process_thread,
|
|
std::function<void()> on_one_ref_remaining)
|
|
: impl_(std::make_unique<Impl>(std::move(process_thread),
|
|
std::move(on_one_ref_remaining))) {}
|
|
|
|
SharedModuleThread::~SharedModuleThread() = default;
|
|
|
|
// static
|
|
|
|
rtc::scoped_refptr<SharedModuleThread> SharedModuleThread::Create(
|
|
std::unique_ptr<ProcessThread> process_thread,
|
|
std::function<void()> on_one_ref_remaining) {
|
|
return new SharedModuleThread(std::move(process_thread),
|
|
std::move(on_one_ref_remaining));
|
|
}
|
|
|
|
void SharedModuleThread::EnsureStarted() {
|
|
impl_->EnsureStarted();
|
|
}
|
|
|
|
ProcessThread* SharedModuleThread::process_thread() {
|
|
return impl_->process_thread();
|
|
}
|
|
|
|
void SharedModuleThread::AddRef() const {
|
|
impl_->AddRef();
|
|
}
|
|
|
|
rtc::RefCountReleaseStatus SharedModuleThread::Release() const {
|
|
auto ret = impl_->Release();
|
|
if (ret == rtc::RefCountReleaseStatus::kDroppedLastRef)
|
|
delete this;
|
|
return ret;
|
|
}
|
|
|
|
// This method here to avoid subclasses has to implement this method.
|
|
// Call perf test will use Internal::Call::CreateVideoSendStream() to inject
|
|
// FecController.
|
|
VideoSendStream* Call::CreateVideoSendStream(
|
|
VideoSendStream::Config config,
|
|
VideoEncoderConfig encoder_config,
|
|
std::unique_ptr<FecController> fec_controller) {
|
|
return nullptr;
|
|
}
|
|
|
|
namespace internal {
|
|
|
|
Call::Call(Clock* clock,
|
|
const Call::Config& config,
|
|
std::unique_ptr<RtpTransportControllerSendInterface> transport_send,
|
|
rtc::scoped_refptr<SharedModuleThread> module_process_thread,
|
|
TaskQueueFactory* task_queue_factory)
|
|
: clock_(clock),
|
|
task_queue_factory_(task_queue_factory),
|
|
worker_thread_(GetCurrentTaskQueueOrThread()),
|
|
num_cpu_cores_(CpuInfo::DetectNumberOfCores()),
|
|
module_process_thread_(std::move(module_process_thread)),
|
|
call_stats_(new CallStats(clock_, worker_thread_)),
|
|
bitrate_allocator_(new BitrateAllocator(this)),
|
|
config_(config),
|
|
audio_network_state_(kNetworkDown),
|
|
video_network_state_(kNetworkDown),
|
|
aggregate_network_up_(false),
|
|
event_log_(config.event_log),
|
|
received_bytes_per_second_counter_(clock_, nullptr, true),
|
|
received_audio_bytes_per_second_counter_(clock_, nullptr, true),
|
|
received_video_bytes_per_second_counter_(clock_, nullptr, true),
|
|
received_rtcp_bytes_per_second_counter_(clock_, nullptr, true),
|
|
last_bandwidth_bps_(0),
|
|
min_allocated_send_bitrate_bps_(0),
|
|
configured_max_padding_bitrate_bps_(0),
|
|
estimated_send_bitrate_kbps_counter_(clock_, nullptr, true),
|
|
pacer_bitrate_kbps_counter_(clock_, nullptr, true),
|
|
receive_side_cc_(clock_, transport_send->packet_router()),
|
|
receive_time_calculator_(ReceiveTimeCalculator::CreateFromFieldTrial()),
|
|
video_send_delay_stats_(new SendDelayStats(clock_)),
|
|
start_ms_(clock_->TimeInMilliseconds()),
|
|
transport_send_ptr_(transport_send.get()),
|
|
transport_send_(std::move(transport_send)) {
|
|
RTC_DCHECK(config.event_log != nullptr);
|
|
RTC_DCHECK(config.trials != nullptr);
|
|
RTC_DCHECK(worker_thread_->IsCurrent());
|
|
|
|
call_stats_->RegisterStatsObserver(&receive_side_cc_);
|
|
|
|
module_process_thread_->process_thread()->RegisterModule(
|
|
receive_side_cc_.GetRemoteBitrateEstimator(true), RTC_FROM_HERE);
|
|
module_process_thread_->process_thread()->RegisterModule(&receive_side_cc_,
|
|
RTC_FROM_HERE);
|
|
}
|
|
|
|
Call::~Call() {
|
|
RTC_DCHECK_RUN_ON(worker_thread_);
|
|
|
|
RTC_CHECK(audio_send_ssrcs_.empty());
|
|
RTC_CHECK(video_send_ssrcs_.empty());
|
|
RTC_CHECK(video_send_streams_.empty());
|
|
RTC_CHECK(audio_receive_streams_.empty());
|
|
RTC_CHECK(video_receive_streams_.empty());
|
|
|
|
module_process_thread_->process_thread()->DeRegisterModule(
|
|
receive_side_cc_.GetRemoteBitrateEstimator(true));
|
|
module_process_thread_->process_thread()->DeRegisterModule(&receive_side_cc_);
|
|
call_stats_->DeregisterStatsObserver(&receive_side_cc_);
|
|
|
|
absl::optional<Timestamp> first_sent_packet_ms =
|
|
transport_send_->GetFirstPacketTime();
|
|
|
|
// Only update histograms after process threads have been shut down, so that
|
|
// they won't try to concurrently update stats.
|
|
if (first_sent_packet_ms) {
|
|
UpdateSendHistograms(*first_sent_packet_ms);
|
|
}
|
|
|
|
UpdateReceiveHistograms();
|
|
UpdateHistograms();
|
|
}
|
|
|
|
void Call::RegisterRateObserver() {
|
|
RTC_DCHECK_RUN_ON(worker_thread_);
|
|
|
|
if (is_target_rate_observer_registered_)
|
|
return;
|
|
|
|
is_target_rate_observer_registered_ = true;
|
|
|
|
// This call seems to kick off a number of things, so probably better left
|
|
// off being kicked off on request rather than in the ctor.
|
|
transport_send_ptr_->RegisterTargetTransferRateObserver(this);
|
|
|
|
module_process_thread_->EnsureStarted();
|
|
}
|
|
|
|
void Call::SetClientBitratePreferences(const BitrateSettings& preferences) {
|
|
RTC_DCHECK_RUN_ON(worker_thread_);
|
|
GetTransportControllerSend()->SetClientBitratePreferences(preferences);
|
|
}
|
|
|
|
void Call::UpdateHistograms() {
|
|
RTC_HISTOGRAM_COUNTS_100000(
|
|
"WebRTC.Call.LifetimeInSeconds",
|
|
(clock_->TimeInMilliseconds() - start_ms_) / 1000);
|
|
}
|
|
|
|
// Called from the dtor.
|
|
void Call::UpdateSendHistograms(Timestamp first_sent_packet) {
|
|
int64_t elapsed_sec =
|
|
(clock_->TimeInMilliseconds() - first_sent_packet.ms()) / 1000;
|
|
if (elapsed_sec < metrics::kMinRunTimeInSeconds)
|
|
return;
|
|
const int kMinRequiredPeriodicSamples = 5;
|
|
AggregatedStats send_bitrate_stats =
|
|
estimated_send_bitrate_kbps_counter_.ProcessAndGetStats();
|
|
if (send_bitrate_stats.num_samples > kMinRequiredPeriodicSamples) {
|
|
RTC_HISTOGRAM_COUNTS_100000("WebRTC.Call.EstimatedSendBitrateInKbps",
|
|
send_bitrate_stats.average);
|
|
RTC_LOG(LS_INFO) << "WebRTC.Call.EstimatedSendBitrateInKbps, "
|
|
<< send_bitrate_stats.ToString();
|
|
}
|
|
AggregatedStats pacer_bitrate_stats =
|
|
pacer_bitrate_kbps_counter_.ProcessAndGetStats();
|
|
if (pacer_bitrate_stats.num_samples > kMinRequiredPeriodicSamples) {
|
|
RTC_HISTOGRAM_COUNTS_100000("WebRTC.Call.PacerBitrateInKbps",
|
|
pacer_bitrate_stats.average);
|
|
RTC_LOG(LS_INFO) << "WebRTC.Call.PacerBitrateInKbps, "
|
|
<< pacer_bitrate_stats.ToString();
|
|
}
|
|
}
|
|
|
|
void Call::UpdateReceiveHistograms() {
|
|
if (first_received_rtp_audio_ms_) {
|
|
RTC_HISTOGRAM_COUNTS_100000(
|
|
"WebRTC.Call.TimeReceivingAudioRtpPacketsInSeconds",
|
|
(*last_received_rtp_audio_ms_ - *first_received_rtp_audio_ms_) / 1000);
|
|
}
|
|
if (first_received_rtp_video_ms_) {
|
|
RTC_HISTOGRAM_COUNTS_100000(
|
|
"WebRTC.Call.TimeReceivingVideoRtpPacketsInSeconds",
|
|
(*last_received_rtp_video_ms_ - *first_received_rtp_video_ms_) / 1000);
|
|
}
|
|
const int kMinRequiredPeriodicSamples = 5;
|
|
AggregatedStats video_bytes_per_sec =
|
|
received_video_bytes_per_second_counter_.GetStats();
|
|
if (video_bytes_per_sec.num_samples > kMinRequiredPeriodicSamples) {
|
|
RTC_HISTOGRAM_COUNTS_100000("WebRTC.Call.VideoBitrateReceivedInKbps",
|
|
video_bytes_per_sec.average * 8 / 1000);
|
|
RTC_LOG(LS_INFO) << "WebRTC.Call.VideoBitrateReceivedInBps, "
|
|
<< video_bytes_per_sec.ToStringWithMultiplier(8);
|
|
}
|
|
AggregatedStats audio_bytes_per_sec =
|
|
received_audio_bytes_per_second_counter_.GetStats();
|
|
if (audio_bytes_per_sec.num_samples > kMinRequiredPeriodicSamples) {
|
|
RTC_HISTOGRAM_COUNTS_100000("WebRTC.Call.AudioBitrateReceivedInKbps",
|
|
audio_bytes_per_sec.average * 8 / 1000);
|
|
RTC_LOG(LS_INFO) << "WebRTC.Call.AudioBitrateReceivedInBps, "
|
|
<< audio_bytes_per_sec.ToStringWithMultiplier(8);
|
|
}
|
|
AggregatedStats rtcp_bytes_per_sec =
|
|
received_rtcp_bytes_per_second_counter_.GetStats();
|
|
if (rtcp_bytes_per_sec.num_samples > kMinRequiredPeriodicSamples) {
|
|
RTC_HISTOGRAM_COUNTS_100000("WebRTC.Call.RtcpBitrateReceivedInBps",
|
|
rtcp_bytes_per_sec.average * 8);
|
|
RTC_LOG(LS_INFO) << "WebRTC.Call.RtcpBitrateReceivedInBps, "
|
|
<< rtcp_bytes_per_sec.ToStringWithMultiplier(8);
|
|
}
|
|
AggregatedStats recv_bytes_per_sec =
|
|
received_bytes_per_second_counter_.GetStats();
|
|
if (recv_bytes_per_sec.num_samples > kMinRequiredPeriodicSamples) {
|
|
RTC_HISTOGRAM_COUNTS_100000("WebRTC.Call.BitrateReceivedInKbps",
|
|
recv_bytes_per_sec.average * 8 / 1000);
|
|
RTC_LOG(LS_INFO) << "WebRTC.Call.BitrateReceivedInBps, "
|
|
<< recv_bytes_per_sec.ToStringWithMultiplier(8);
|
|
}
|
|
}
|
|
|
|
PacketReceiver* Call::Receiver() {
|
|
RTC_DCHECK_RUN_ON(worker_thread_);
|
|
return this;
|
|
}
|
|
|
|
webrtc::AudioSendStream* Call::CreateAudioSendStream(
|
|
const webrtc::AudioSendStream::Config& config) {
|
|
TRACE_EVENT0("webrtc", "Call::CreateAudioSendStream");
|
|
RTC_DCHECK_RUN_ON(worker_thread_);
|
|
|
|
RegisterRateObserver();
|
|
|
|
// Stream config is logged in AudioSendStream::ConfigureStream, as it may
|
|
// change during the stream's lifetime.
|
|
absl::optional<RtpState> suspended_rtp_state;
|
|
{
|
|
const auto& iter = suspended_audio_send_ssrcs_.find(config.rtp.ssrc);
|
|
if (iter != suspended_audio_send_ssrcs_.end()) {
|
|
suspended_rtp_state.emplace(iter->second);
|
|
}
|
|
}
|
|
|
|
AudioSendStream* send_stream = new AudioSendStream(
|
|
clock_, config, config_.audio_state, task_queue_factory_,
|
|
module_process_thread_->process_thread(), transport_send_ptr_,
|
|
bitrate_allocator_.get(), event_log_, call_stats_->AsRtcpRttStats(),
|
|
suspended_rtp_state);
|
|
RTC_DCHECK(audio_send_ssrcs_.find(config.rtp.ssrc) ==
|
|
audio_send_ssrcs_.end());
|
|
audio_send_ssrcs_[config.rtp.ssrc] = send_stream;
|
|
|
|
for (AudioReceiveStream* stream : audio_receive_streams_) {
|
|
if (stream->config().rtp.local_ssrc == config.rtp.ssrc) {
|
|
stream->AssociateSendStream(send_stream);
|
|
}
|
|
}
|
|
|
|
UpdateAggregateNetworkState();
|
|
return send_stream;
|
|
}
|
|
|
|
void Call::DestroyAudioSendStream(webrtc::AudioSendStream* send_stream) {
|
|
TRACE_EVENT0("webrtc", "Call::DestroyAudioSendStream");
|
|
RTC_DCHECK_RUN_ON(worker_thread_);
|
|
RTC_DCHECK(send_stream != nullptr);
|
|
|
|
send_stream->Stop();
|
|
|
|
const uint32_t ssrc = send_stream->GetConfig().rtp.ssrc;
|
|
webrtc::internal::AudioSendStream* audio_send_stream =
|
|
static_cast<webrtc::internal::AudioSendStream*>(send_stream);
|
|
suspended_audio_send_ssrcs_[ssrc] = audio_send_stream->GetRtpState();
|
|
|
|
size_t num_deleted = audio_send_ssrcs_.erase(ssrc);
|
|
RTC_DCHECK_EQ(1, num_deleted);
|
|
|
|
for (AudioReceiveStream* stream : audio_receive_streams_) {
|
|
if (stream->config().rtp.local_ssrc == ssrc) {
|
|
stream->AssociateSendStream(nullptr);
|
|
}
|
|
}
|
|
|
|
UpdateAggregateNetworkState();
|
|
delete send_stream;
|
|
}
|
|
|
|
webrtc::AudioReceiveStream* Call::CreateAudioReceiveStream(
|
|
const webrtc::AudioReceiveStream::Config& config) {
|
|
TRACE_EVENT0("webrtc", "Call::CreateAudioReceiveStream");
|
|
RTC_DCHECK_RUN_ON(worker_thread_);
|
|
RegisterRateObserver();
|
|
event_log_->Log(std::make_unique<RtcEventAudioReceiveStreamConfig>(
|
|
CreateRtcLogStreamConfig(config)));
|
|
AudioReceiveStream* receive_stream = new AudioReceiveStream(
|
|
clock_, &audio_receiver_controller_, transport_send_ptr_->packet_router(),
|
|
module_process_thread_->process_thread(), config_.neteq_factory, config,
|
|
config_.audio_state, event_log_);
|
|
|
|
receive_rtp_config_.emplace(config.rtp.remote_ssrc, ReceiveRtpConfig(config));
|
|
audio_receive_streams_.insert(receive_stream);
|
|
|
|
ConfigureSync(config.sync_group);
|
|
|
|
auto it = audio_send_ssrcs_.find(config.rtp.local_ssrc);
|
|
if (it != audio_send_ssrcs_.end()) {
|
|
receive_stream->AssociateSendStream(it->second);
|
|
}
|
|
|
|
UpdateAggregateNetworkState();
|
|
return receive_stream;
|
|
}
|
|
|
|
void Call::DestroyAudioReceiveStream(
|
|
webrtc::AudioReceiveStream* receive_stream) {
|
|
TRACE_EVENT0("webrtc", "Call::DestroyAudioReceiveStream");
|
|
RTC_DCHECK_RUN_ON(worker_thread_);
|
|
RTC_DCHECK(receive_stream != nullptr);
|
|
webrtc::internal::AudioReceiveStream* audio_receive_stream =
|
|
static_cast<webrtc::internal::AudioReceiveStream*>(receive_stream);
|
|
|
|
const AudioReceiveStream::Config& config = audio_receive_stream->config();
|
|
uint32_t ssrc = config.rtp.remote_ssrc;
|
|
receive_side_cc_.GetRemoteBitrateEstimator(UseSendSideBwe(config))
|
|
->RemoveStream(ssrc);
|
|
audio_receive_streams_.erase(audio_receive_stream);
|
|
const std::string& sync_group = audio_receive_stream->config().sync_group;
|
|
const auto it = sync_stream_mapping_.find(sync_group);
|
|
if (it != sync_stream_mapping_.end() && it->second == audio_receive_stream) {
|
|
sync_stream_mapping_.erase(it);
|
|
ConfigureSync(sync_group);
|
|
}
|
|
receive_rtp_config_.erase(ssrc);
|
|
|
|
UpdateAggregateNetworkState();
|
|
delete audio_receive_stream;
|
|
}
|
|
|
|
// This method can be used for Call tests with external fec controller factory.
|
|
webrtc::VideoSendStream* Call::CreateVideoSendStream(
|
|
webrtc::VideoSendStream::Config config,
|
|
VideoEncoderConfig encoder_config,
|
|
std::unique_ptr<FecController> fec_controller) {
|
|
TRACE_EVENT0("webrtc", "Call::CreateVideoSendStream");
|
|
RTC_DCHECK_RUN_ON(worker_thread_);
|
|
|
|
RegisterRateObserver();
|
|
|
|
video_send_delay_stats_->AddSsrcs(config);
|
|
for (size_t ssrc_index = 0; ssrc_index < config.rtp.ssrcs.size();
|
|
++ssrc_index) {
|
|
event_log_->Log(std::make_unique<RtcEventVideoSendStreamConfig>(
|
|
CreateRtcLogStreamConfig(config, ssrc_index)));
|
|
}
|
|
|
|
// TODO(mflodman): Base the start bitrate on a current bandwidth estimate, if
|
|
// the call has already started.
|
|
// Copy ssrcs from |config| since |config| is moved.
|
|
std::vector<uint32_t> ssrcs = config.rtp.ssrcs;
|
|
|
|
VideoSendStream* send_stream = new VideoSendStream(
|
|
clock_, num_cpu_cores_, module_process_thread_->process_thread(),
|
|
task_queue_factory_, call_stats_->AsRtcpRttStats(), transport_send_ptr_,
|
|
bitrate_allocator_.get(), video_send_delay_stats_.get(), event_log_,
|
|
std::move(config), std::move(encoder_config), suspended_video_send_ssrcs_,
|
|
suspended_video_payload_states_, std::move(fec_controller));
|
|
|
|
for (uint32_t ssrc : ssrcs) {
|
|
RTC_DCHECK(video_send_ssrcs_.find(ssrc) == video_send_ssrcs_.end());
|
|
video_send_ssrcs_[ssrc] = send_stream;
|
|
}
|
|
video_send_streams_.insert(send_stream);
|
|
// Forward resources that were previously added to the call to the new stream.
|
|
for (const auto& resource_forwarder : adaptation_resource_forwarders_) {
|
|
resource_forwarder->OnCreateVideoSendStream(send_stream);
|
|
}
|
|
|
|
UpdateAggregateNetworkState();
|
|
|
|
return send_stream;
|
|
}
|
|
|
|
webrtc::VideoSendStream* Call::CreateVideoSendStream(
|
|
webrtc::VideoSendStream::Config config,
|
|
VideoEncoderConfig encoder_config) {
|
|
if (config_.fec_controller_factory) {
|
|
RTC_LOG(LS_INFO) << "External FEC Controller will be used.";
|
|
}
|
|
std::unique_ptr<FecController> fec_controller =
|
|
config_.fec_controller_factory
|
|
? config_.fec_controller_factory->CreateFecController()
|
|
: std::make_unique<FecControllerDefault>(clock_);
|
|
return CreateVideoSendStream(std::move(config), std::move(encoder_config),
|
|
std::move(fec_controller));
|
|
}
|
|
|
|
void Call::DestroyVideoSendStream(webrtc::VideoSendStream* send_stream) {
|
|
TRACE_EVENT0("webrtc", "Call::DestroyVideoSendStream");
|
|
RTC_DCHECK(send_stream != nullptr);
|
|
RTC_DCHECK_RUN_ON(worker_thread_);
|
|
|
|
send_stream->Stop();
|
|
|
|
VideoSendStream* send_stream_impl = nullptr;
|
|
|
|
auto it = video_send_ssrcs_.begin();
|
|
while (it != video_send_ssrcs_.end()) {
|
|
if (it->second == static_cast<VideoSendStream*>(send_stream)) {
|
|
send_stream_impl = it->second;
|
|
video_send_ssrcs_.erase(it++);
|
|
} else {
|
|
++it;
|
|
}
|
|
}
|
|
// Stop forwarding resources to the stream being destroyed.
|
|
for (const auto& resource_forwarder : adaptation_resource_forwarders_) {
|
|
resource_forwarder->OnDestroyVideoSendStream(send_stream_impl);
|
|
}
|
|
video_send_streams_.erase(send_stream_impl);
|
|
|
|
RTC_CHECK(send_stream_impl != nullptr);
|
|
|
|
VideoSendStream::RtpStateMap rtp_states;
|
|
VideoSendStream::RtpPayloadStateMap rtp_payload_states;
|
|
send_stream_impl->StopPermanentlyAndGetRtpStates(&rtp_states,
|
|
&rtp_payload_states);
|
|
for (const auto& kv : rtp_states) {
|
|
suspended_video_send_ssrcs_[kv.first] = kv.second;
|
|
}
|
|
for (const auto& kv : rtp_payload_states) {
|
|
suspended_video_payload_states_[kv.first] = kv.second;
|
|
}
|
|
|
|
UpdateAggregateNetworkState();
|
|
delete send_stream_impl;
|
|
}
|
|
|
|
webrtc::VideoReceiveStream* Call::CreateVideoReceiveStream(
|
|
webrtc::VideoReceiveStream::Config configuration) {
|
|
TRACE_EVENT0("webrtc", "Call::CreateVideoReceiveStream");
|
|
RTC_DCHECK_RUN_ON(worker_thread_);
|
|
|
|
receive_side_cc_.SetSendPeriodicFeedback(
|
|
SendPeriodicFeedback(configuration.rtp.extensions));
|
|
|
|
RegisterRateObserver();
|
|
|
|
TaskQueueBase* current = GetCurrentTaskQueueOrThread();
|
|
RTC_CHECK(current);
|
|
VideoReceiveStream2* receive_stream = new VideoReceiveStream2(
|
|
task_queue_factory_, current, &video_receiver_controller_, num_cpu_cores_,
|
|
transport_send_ptr_->packet_router(), std::move(configuration),
|
|
module_process_thread_->process_thread(), call_stats_.get(), clock_,
|
|
new VCMTiming(clock_));
|
|
|
|
const webrtc::VideoReceiveStream::Config& config = receive_stream->config();
|
|
if (config.rtp.rtx_ssrc) {
|
|
// We record identical config for the rtx stream as for the main
|
|
// stream. Since the transport_send_cc negotiation is per payload
|
|
// type, we may get an incorrect value for the rtx stream, but
|
|
// that is unlikely to matter in practice.
|
|
receive_rtp_config_.emplace(config.rtp.rtx_ssrc, ReceiveRtpConfig(config));
|
|
}
|
|
receive_rtp_config_.emplace(config.rtp.remote_ssrc, ReceiveRtpConfig(config));
|
|
video_receive_streams_.insert(receive_stream);
|
|
ConfigureSync(config.sync_group);
|
|
|
|
receive_stream->SignalNetworkState(video_network_state_);
|
|
UpdateAggregateNetworkState();
|
|
event_log_->Log(std::make_unique<RtcEventVideoReceiveStreamConfig>(
|
|
CreateRtcLogStreamConfig(config)));
|
|
return receive_stream;
|
|
}
|
|
|
|
void Call::DestroyVideoReceiveStream(
|
|
webrtc::VideoReceiveStream* receive_stream) {
|
|
TRACE_EVENT0("webrtc", "Call::DestroyVideoReceiveStream");
|
|
RTC_DCHECK_RUN_ON(worker_thread_);
|
|
RTC_DCHECK(receive_stream != nullptr);
|
|
VideoReceiveStream2* receive_stream_impl =
|
|
static_cast<VideoReceiveStream2*>(receive_stream);
|
|
const VideoReceiveStream::Config& config = receive_stream_impl->config();
|
|
|
|
// Remove all ssrcs pointing to a receive stream. As RTX retransmits on a
|
|
// separate SSRC there can be either one or two.
|
|
receive_rtp_config_.erase(config.rtp.remote_ssrc);
|
|
if (config.rtp.rtx_ssrc) {
|
|
receive_rtp_config_.erase(config.rtp.rtx_ssrc);
|
|
}
|
|
video_receive_streams_.erase(receive_stream_impl);
|
|
ConfigureSync(config.sync_group);
|
|
|
|
receive_side_cc_.GetRemoteBitrateEstimator(UseSendSideBwe(config))
|
|
->RemoveStream(config.rtp.remote_ssrc);
|
|
|
|
UpdateAggregateNetworkState();
|
|
delete receive_stream_impl;
|
|
}
|
|
|
|
FlexfecReceiveStream* Call::CreateFlexfecReceiveStream(
|
|
const FlexfecReceiveStream::Config& config) {
|
|
TRACE_EVENT0("webrtc", "Call::CreateFlexfecReceiveStream");
|
|
RTC_DCHECK_RUN_ON(worker_thread_);
|
|
|
|
RecoveredPacketReceiver* recovered_packet_receiver = this;
|
|
|
|
FlexfecReceiveStreamImpl* receive_stream;
|
|
|
|
// Unlike the video and audio receive streams, FlexfecReceiveStream implements
|
|
// RtpPacketSinkInterface itself, and hence its constructor passes its |this|
|
|
// pointer to video_receiver_controller_->CreateStream(). Calling the
|
|
// constructor while on the worker thread ensures that we don't call
|
|
// OnRtpPacket until the constructor is finished and the object is
|
|
// in a valid state, since OnRtpPacket runs on the same thread.
|
|
receive_stream = new FlexfecReceiveStreamImpl(
|
|
clock_, &video_receiver_controller_, config, recovered_packet_receiver,
|
|
call_stats_->AsRtcpRttStats(), module_process_thread_->process_thread());
|
|
|
|
RTC_DCHECK(receive_rtp_config_.find(config.remote_ssrc) ==
|
|
receive_rtp_config_.end());
|
|
receive_rtp_config_.emplace(config.remote_ssrc, ReceiveRtpConfig(config));
|
|
|
|
// TODO(brandtr): Store config in RtcEventLog here.
|
|
|
|
return receive_stream;
|
|
}
|
|
|
|
void Call::DestroyFlexfecReceiveStream(FlexfecReceiveStream* receive_stream) {
|
|
TRACE_EVENT0("webrtc", "Call::DestroyFlexfecReceiveStream");
|
|
RTC_DCHECK_RUN_ON(worker_thread_);
|
|
|
|
RTC_DCHECK(receive_stream != nullptr);
|
|
const FlexfecReceiveStream::Config& config = receive_stream->GetConfig();
|
|
uint32_t ssrc = config.remote_ssrc;
|
|
receive_rtp_config_.erase(ssrc);
|
|
|
|
// Remove all SSRCs pointing to the FlexfecReceiveStreamImpl to be
|
|
// destroyed.
|
|
receive_side_cc_.GetRemoteBitrateEstimator(UseSendSideBwe(config))
|
|
->RemoveStream(ssrc);
|
|
|
|
delete receive_stream;
|
|
}
|
|
|
|
void Call::AddAdaptationResource(rtc::scoped_refptr<Resource> resource) {
|
|
RTC_DCHECK_RUN_ON(worker_thread_);
|
|
adaptation_resource_forwarders_.push_back(
|
|
std::make_unique<ResourceVideoSendStreamForwarder>(resource));
|
|
const auto& resource_forwarder = adaptation_resource_forwarders_.back();
|
|
for (VideoSendStream* send_stream : video_send_streams_) {
|
|
resource_forwarder->OnCreateVideoSendStream(send_stream);
|
|
}
|
|
}
|
|
|
|
RtpTransportControllerSendInterface* Call::GetTransportControllerSend() {
|
|
return transport_send_ptr_;
|
|
}
|
|
|
|
Call::Stats Call::GetStats() const {
|
|
RTC_DCHECK_RUN_ON(worker_thread_);
|
|
|
|
Stats stats;
|
|
// TODO(srte): It is unclear if we only want to report queues if network is
|
|
// available.
|
|
stats.pacer_delay_ms =
|
|
aggregate_network_up_ ? transport_send_ptr_->GetPacerQueuingDelayMs() : 0;
|
|
|
|
stats.rtt_ms = call_stats_->LastProcessedRtt();
|
|
|
|
// Fetch available send/receive bitrates.
|
|
std::vector<unsigned int> ssrcs;
|
|
uint32_t recv_bandwidth = 0;
|
|
receive_side_cc_.GetRemoteBitrateEstimator(false)->LatestEstimate(
|
|
&ssrcs, &recv_bandwidth);
|
|
stats.recv_bandwidth_bps = recv_bandwidth;
|
|
stats.send_bandwidth_bps = last_bandwidth_bps_;
|
|
stats.max_padding_bitrate_bps = configured_max_padding_bitrate_bps_;
|
|
|
|
return stats;
|
|
}
|
|
|
|
void Call::SignalChannelNetworkState(MediaType media, NetworkState state) {
|
|
RTC_DCHECK_RUN_ON(worker_thread_);
|
|
switch (media) {
|
|
case MediaType::AUDIO:
|
|
audio_network_state_ = state;
|
|
break;
|
|
case MediaType::VIDEO:
|
|
video_network_state_ = state;
|
|
break;
|
|
case MediaType::ANY:
|
|
case MediaType::DATA:
|
|
RTC_NOTREACHED();
|
|
break;
|
|
}
|
|
|
|
UpdateAggregateNetworkState();
|
|
for (VideoReceiveStream2* video_receive_stream : video_receive_streams_) {
|
|
video_receive_stream->SignalNetworkState(video_network_state_);
|
|
}
|
|
}
|
|
|
|
void Call::OnAudioTransportOverheadChanged(int transport_overhead_per_packet) {
|
|
RTC_DCHECK_RUN_ON(worker_thread_);
|
|
for (auto& kv : audio_send_ssrcs_) {
|
|
kv.second->SetTransportOverhead(transport_overhead_per_packet);
|
|
}
|
|
}
|
|
|
|
void Call::UpdateAggregateNetworkState() {
|
|
RTC_DCHECK_RUN_ON(worker_thread_);
|
|
|
|
bool have_audio =
|
|
!audio_send_ssrcs_.empty() || !audio_receive_streams_.empty();
|
|
bool have_video =
|
|
!video_send_ssrcs_.empty() || !video_receive_streams_.empty();
|
|
|
|
bool aggregate_network_up =
|
|
((have_video && video_network_state_ == kNetworkUp) ||
|
|
(have_audio && audio_network_state_ == kNetworkUp));
|
|
|
|
if (aggregate_network_up != aggregate_network_up_) {
|
|
RTC_LOG(LS_INFO)
|
|
<< "UpdateAggregateNetworkState: aggregate_state change to "
|
|
<< (aggregate_network_up ? "up" : "down");
|
|
} else {
|
|
RTC_LOG(LS_VERBOSE)
|
|
<< "UpdateAggregateNetworkState: aggregate_state remains at "
|
|
<< (aggregate_network_up ? "up" : "down");
|
|
}
|
|
aggregate_network_up_ = aggregate_network_up;
|
|
|
|
transport_send_ptr_->OnNetworkAvailability(aggregate_network_up);
|
|
}
|
|
|
|
void Call::OnSentPacket(const rtc::SentPacket& sent_packet) {
|
|
video_send_delay_stats_->OnSentPacket(sent_packet.packet_id,
|
|
clock_->TimeInMilliseconds());
|
|
transport_send_ptr_->OnSentPacket(sent_packet);
|
|
}
|
|
|
|
void Call::OnStartRateUpdate(DataRate start_rate) {
|
|
RTC_DCHECK_RUN_ON(send_transport_queue());
|
|
bitrate_allocator_->UpdateStartRate(start_rate.bps<uint32_t>());
|
|
}
|
|
|
|
void Call::OnTargetTransferRate(TargetTransferRate msg) {
|
|
RTC_DCHECK_RUN_ON(send_transport_queue());
|
|
|
|
uint32_t target_bitrate_bps = msg.target_rate.bps();
|
|
// For controlling the rate of feedback messages.
|
|
receive_side_cc_.OnBitrateChanged(target_bitrate_bps);
|
|
bitrate_allocator_->OnNetworkEstimateChanged(msg);
|
|
|
|
worker_thread_->PostTask(
|
|
ToQueuedTask(task_safety_, [this, target_bitrate_bps]() {
|
|
RTC_DCHECK_RUN_ON(worker_thread_);
|
|
last_bandwidth_bps_ = target_bitrate_bps;
|
|
|
|
// Ignore updates if bitrate is zero (the aggregate network state is
|
|
// down) or if we're not sending video.
|
|
if (target_bitrate_bps == 0 || video_send_streams_.empty()) {
|
|
estimated_send_bitrate_kbps_counter_.ProcessAndPause();
|
|
pacer_bitrate_kbps_counter_.ProcessAndPause();
|
|
return;
|
|
}
|
|
|
|
estimated_send_bitrate_kbps_counter_.Add(target_bitrate_bps / 1000);
|
|
// Pacer bitrate may be higher than bitrate estimate if enforcing min
|
|
// bitrate.
|
|
uint32_t pacer_bitrate_bps =
|
|
std::max(target_bitrate_bps, min_allocated_send_bitrate_bps_);
|
|
pacer_bitrate_kbps_counter_.Add(pacer_bitrate_bps / 1000);
|
|
}));
|
|
}
|
|
|
|
void Call::OnAllocationLimitsChanged(BitrateAllocationLimits limits) {
|
|
RTC_DCHECK_RUN_ON(send_transport_queue());
|
|
|
|
transport_send_ptr_->SetAllocatedSendBitrateLimits(limits);
|
|
|
|
worker_thread_->PostTask(ToQueuedTask(task_safety_, [this, limits]() {
|
|
RTC_DCHECK_RUN_ON(worker_thread_);
|
|
min_allocated_send_bitrate_bps_ = limits.min_allocatable_rate.bps();
|
|
configured_max_padding_bitrate_bps_ = limits.max_padding_rate.bps();
|
|
}));
|
|
}
|
|
|
|
void Call::ConfigureSync(const std::string& sync_group) {
|
|
// Set sync only if there was no previous one.
|
|
if (sync_group.empty())
|
|
return;
|
|
|
|
AudioReceiveStream* sync_audio_stream = nullptr;
|
|
// Find existing audio stream.
|
|
const auto it = sync_stream_mapping_.find(sync_group);
|
|
if (it != sync_stream_mapping_.end()) {
|
|
sync_audio_stream = it->second;
|
|
} else {
|
|
// No configured audio stream, see if we can find one.
|
|
for (AudioReceiveStream* stream : audio_receive_streams_) {
|
|
if (stream->config().sync_group == sync_group) {
|
|
if (sync_audio_stream != nullptr) {
|
|
RTC_LOG(LS_WARNING)
|
|
<< "Attempting to sync more than one audio stream "
|
|
"within the same sync group. This is not "
|
|
"supported in the current implementation.";
|
|
break;
|
|
}
|
|
sync_audio_stream = stream;
|
|
}
|
|
}
|
|
}
|
|
if (sync_audio_stream)
|
|
sync_stream_mapping_[sync_group] = sync_audio_stream;
|
|
size_t num_synced_streams = 0;
|
|
for (VideoReceiveStream2* video_stream : video_receive_streams_) {
|
|
if (video_stream->config().sync_group != sync_group)
|
|
continue;
|
|
++num_synced_streams;
|
|
if (num_synced_streams > 1) {
|
|
// TODO(pbos): Support synchronizing more than one A/V pair.
|
|
// https://code.google.com/p/webrtc/issues/detail?id=4762
|
|
RTC_LOG(LS_WARNING)
|
|
<< "Attempting to sync more than one audio/video pair "
|
|
"within the same sync group. This is not supported in "
|
|
"the current implementation.";
|
|
}
|
|
// Only sync the first A/V pair within this sync group.
|
|
if (num_synced_streams == 1) {
|
|
// sync_audio_stream may be null and that's ok.
|
|
video_stream->SetSync(sync_audio_stream);
|
|
} else {
|
|
video_stream->SetSync(nullptr);
|
|
}
|
|
}
|
|
}
|
|
|
|
PacketReceiver::DeliveryStatus Call::DeliverRtcp(MediaType media_type,
|
|
const uint8_t* packet,
|
|
size_t length) {
|
|
TRACE_EVENT0("webrtc", "Call::DeliverRtcp");
|
|
// TODO(pbos): Make sure it's a valid packet.
|
|
// Return DELIVERY_UNKNOWN_SSRC if it can be determined that
|
|
// there's no receiver of the packet.
|
|
if (received_bytes_per_second_counter_.HasSample()) {
|
|
// First RTP packet has been received.
|
|
received_bytes_per_second_counter_.Add(static_cast<int>(length));
|
|
received_rtcp_bytes_per_second_counter_.Add(static_cast<int>(length));
|
|
}
|
|
bool rtcp_delivered = false;
|
|
if (media_type == MediaType::ANY || media_type == MediaType::VIDEO) {
|
|
for (VideoReceiveStream2* stream : video_receive_streams_) {
|
|
if (stream->DeliverRtcp(packet, length))
|
|
rtcp_delivered = true;
|
|
}
|
|
}
|
|
if (media_type == MediaType::ANY || media_type == MediaType::AUDIO) {
|
|
for (AudioReceiveStream* stream : audio_receive_streams_) {
|
|
stream->DeliverRtcp(packet, length);
|
|
rtcp_delivered = true;
|
|
}
|
|
}
|
|
if (media_type == MediaType::ANY || media_type == MediaType::VIDEO) {
|
|
for (VideoSendStream* stream : video_send_streams_) {
|
|
stream->DeliverRtcp(packet, length);
|
|
rtcp_delivered = true;
|
|
}
|
|
}
|
|
if (media_type == MediaType::ANY || media_type == MediaType::AUDIO) {
|
|
for (auto& kv : audio_send_ssrcs_) {
|
|
kv.second->DeliverRtcp(packet, length);
|
|
rtcp_delivered = true;
|
|
}
|
|
}
|
|
|
|
if (rtcp_delivered) {
|
|
event_log_->Log(std::make_unique<RtcEventRtcpPacketIncoming>(
|
|
rtc::MakeArrayView(packet, length)));
|
|
}
|
|
|
|
return rtcp_delivered ? DELIVERY_OK : DELIVERY_PACKET_ERROR;
|
|
}
|
|
|
|
PacketReceiver::DeliveryStatus Call::DeliverRtp(MediaType media_type,
|
|
rtc::CopyOnWriteBuffer packet,
|
|
int64_t packet_time_us) {
|
|
TRACE_EVENT0("webrtc", "Call::DeliverRtp");
|
|
|
|
RtpPacketReceived parsed_packet;
|
|
if (!parsed_packet.Parse(std::move(packet)))
|
|
return DELIVERY_PACKET_ERROR;
|
|
|
|
if (packet_time_us != -1) {
|
|
if (receive_time_calculator_) {
|
|
// Repair packet_time_us for clock resets by comparing a new read of
|
|
// the same clock (TimeUTCMicros) to a monotonic clock reading.
|
|
packet_time_us = receive_time_calculator_->ReconcileReceiveTimes(
|
|
packet_time_us, rtc::TimeUTCMicros(), clock_->TimeInMicroseconds());
|
|
}
|
|
parsed_packet.set_arrival_time_ms((packet_time_us + 500) / 1000);
|
|
} else {
|
|
parsed_packet.set_arrival_time_ms(clock_->TimeInMilliseconds());
|
|
}
|
|
|
|
// We might get RTP keep-alive packets in accordance with RFC6263 section 4.6.
|
|
// These are empty (zero length payload) RTP packets with an unsignaled
|
|
// payload type.
|
|
const bool is_keep_alive_packet = parsed_packet.payload_size() == 0;
|
|
|
|
RTC_DCHECK(media_type == MediaType::AUDIO || media_type == MediaType::VIDEO ||
|
|
is_keep_alive_packet);
|
|
|
|
auto it = receive_rtp_config_.find(parsed_packet.Ssrc());
|
|
if (it == receive_rtp_config_.end()) {
|
|
RTC_LOG(LS_ERROR) << "receive_rtp_config_ lookup failed for ssrc "
|
|
<< parsed_packet.Ssrc();
|
|
// Destruction of the receive stream, including deregistering from the
|
|
// RtpDemuxer, is not protected by the |worker_thread_|.
|
|
// But deregistering in the |receive_rtp_config_| map is. So by not passing
|
|
// the packet on to demuxing in this case, we prevent incoming packets to be
|
|
// passed on via the demuxer to a receive stream which is being torned down.
|
|
return DELIVERY_UNKNOWN_SSRC;
|
|
}
|
|
|
|
parsed_packet.IdentifyExtensions(it->second.extensions);
|
|
|
|
NotifyBweOfReceivedPacket(parsed_packet, media_type);
|
|
|
|
// RateCounters expect input parameter as int, save it as int,
|
|
// instead of converting each time it is passed to RateCounter::Add below.
|
|
int length = static_cast<int>(parsed_packet.size());
|
|
if (media_type == MediaType::AUDIO) {
|
|
if (audio_receiver_controller_.OnRtpPacket(parsed_packet)) {
|
|
received_bytes_per_second_counter_.Add(length);
|
|
received_audio_bytes_per_second_counter_.Add(length);
|
|
event_log_->Log(
|
|
std::make_unique<RtcEventRtpPacketIncoming>(parsed_packet));
|
|
const int64_t arrival_time_ms = parsed_packet.arrival_time_ms();
|
|
if (!first_received_rtp_audio_ms_) {
|
|
first_received_rtp_audio_ms_.emplace(arrival_time_ms);
|
|
}
|
|
last_received_rtp_audio_ms_.emplace(arrival_time_ms);
|
|
return DELIVERY_OK;
|
|
}
|
|
} else if (media_type == MediaType::VIDEO) {
|
|
parsed_packet.set_payload_type_frequency(kVideoPayloadTypeFrequency);
|
|
if (video_receiver_controller_.OnRtpPacket(parsed_packet)) {
|
|
received_bytes_per_second_counter_.Add(length);
|
|
received_video_bytes_per_second_counter_.Add(length);
|
|
event_log_->Log(
|
|
std::make_unique<RtcEventRtpPacketIncoming>(parsed_packet));
|
|
const int64_t arrival_time_ms = parsed_packet.arrival_time_ms();
|
|
if (!first_received_rtp_video_ms_) {
|
|
first_received_rtp_video_ms_.emplace(arrival_time_ms);
|
|
}
|
|
last_received_rtp_video_ms_.emplace(arrival_time_ms);
|
|
return DELIVERY_OK;
|
|
}
|
|
}
|
|
return DELIVERY_UNKNOWN_SSRC;
|
|
}
|
|
|
|
PacketReceiver::DeliveryStatus Call::DeliverPacket(
|
|
MediaType media_type,
|
|
rtc::CopyOnWriteBuffer packet,
|
|
int64_t packet_time_us) {
|
|
RTC_DCHECK_RUN_ON(worker_thread_);
|
|
|
|
if (IsRtcp(packet.cdata(), packet.size()))
|
|
return DeliverRtcp(media_type, packet.cdata(), packet.size());
|
|
|
|
return DeliverRtp(media_type, std::move(packet), packet_time_us);
|
|
}
|
|
|
|
void Call::OnRecoveredPacket(const uint8_t* packet, size_t length) {
|
|
RTC_DCHECK_RUN_ON(worker_thread_);
|
|
RtpPacketReceived parsed_packet;
|
|
if (!parsed_packet.Parse(packet, length))
|
|
return;
|
|
|
|
parsed_packet.set_recovered(true);
|
|
|
|
auto it = receive_rtp_config_.find(parsed_packet.Ssrc());
|
|
if (it == receive_rtp_config_.end()) {
|
|
RTC_LOG(LS_ERROR) << "receive_rtp_config_ lookup failed for ssrc "
|
|
<< parsed_packet.Ssrc();
|
|
// Destruction of the receive stream, including deregistering from the
|
|
// RtpDemuxer, is not protected by the |worker_thread_|.
|
|
// But deregistering in the |receive_rtp_config_| map is.
|
|
// So by not passing the packet on to demuxing in this case, we prevent
|
|
// incoming packets to be passed on via the demuxer to a receive stream
|
|
// which is being torn down.
|
|
return;
|
|
}
|
|
parsed_packet.IdentifyExtensions(it->second.extensions);
|
|
|
|
// TODO(brandtr): Update here when we support protecting audio packets too.
|
|
parsed_packet.set_payload_type_frequency(kVideoPayloadTypeFrequency);
|
|
video_receiver_controller_.OnRtpPacket(parsed_packet);
|
|
}
|
|
|
|
void Call::NotifyBweOfReceivedPacket(const RtpPacketReceived& packet,
|
|
MediaType media_type) {
|
|
auto it = receive_rtp_config_.find(packet.Ssrc());
|
|
bool use_send_side_bwe =
|
|
(it != receive_rtp_config_.end()) && it->second.use_send_side_bwe;
|
|
|
|
RTPHeader header;
|
|
packet.GetHeader(&header);
|
|
|
|
ReceivedPacket packet_msg;
|
|
packet_msg.size = DataSize::Bytes(packet.payload_size());
|
|
packet_msg.receive_time = Timestamp::Millis(packet.arrival_time_ms());
|
|
if (header.extension.hasAbsoluteSendTime) {
|
|
packet_msg.send_time = header.extension.GetAbsoluteSendTimestamp();
|
|
}
|
|
transport_send_ptr_->OnReceivedPacket(packet_msg);
|
|
|
|
if (!use_send_side_bwe && header.extension.hasTransportSequenceNumber) {
|
|
// Inconsistent configuration of send side BWE. Do nothing.
|
|
// TODO(nisse): Without this check, we may produce RTCP feedback
|
|
// packets even when not negotiated. But it would be cleaner to
|
|
// move the check down to RTCPSender::SendFeedbackPacket, which
|
|
// would also help the PacketRouter to select an appropriate rtp
|
|
// module in the case that some, but not all, have RTCP feedback
|
|
// enabled.
|
|
return;
|
|
}
|
|
// For audio, we only support send side BWE.
|
|
if (media_type == MediaType::VIDEO ||
|
|
(use_send_side_bwe && header.extension.hasTransportSequenceNumber)) {
|
|
receive_side_cc_.OnReceivedPacket(
|
|
packet.arrival_time_ms(), packet.payload_size() + packet.padding_size(),
|
|
header);
|
|
}
|
|
}
|
|
|
|
} // namespace internal
|
|
|
|
} // namespace webrtc
|