/* * Copyright (c) 2013 The WebRTC project authors. All Rights Reserved. * * Use of this source code is governed by a BSD-style license * that can be found in the LICENSE file in the root of the source * tree. An additional intellectual property rights grant can be found * in the file PATENTS. All contributing project authors may * be found in the AUTHORS file in the root of the source tree. */ #include "call/call.h" #include #include #include #include #include #include #include #include "absl/types/optional.h" #include "api/rtc_event_log/rtc_event_log.h" #include "api/transport/network_control.h" #include "audio/audio_receive_stream.h" #include "audio/audio_send_stream.h" #include "audio/audio_state.h" #include "call/adaptation/broadcast_resource_listener.h" #include "call/bitrate_allocator.h" #include "call/flexfec_receive_stream_impl.h" #include "call/receive_time_calculator.h" #include "call/rtp_stream_receiver_controller.h" #include "call/rtp_transport_controller_send.h" #include "logging/rtc_event_log/events/rtc_event_audio_receive_stream_config.h" #include "logging/rtc_event_log/events/rtc_event_rtcp_packet_incoming.h" #include "logging/rtc_event_log/events/rtc_event_rtp_packet_incoming.h" #include "logging/rtc_event_log/events/rtc_event_video_receive_stream_config.h" #include "logging/rtc_event_log/events/rtc_event_video_send_stream_config.h" #include "logging/rtc_event_log/rtc_stream_config.h" #include "modules/congestion_controller/include/receive_side_congestion_controller.h" #include "modules/rtp_rtcp/include/flexfec_receiver.h" #include "modules/rtp_rtcp/include/rtp_header_extension_map.h" #include "modules/rtp_rtcp/source/byte_io.h" #include "modules/rtp_rtcp/source/rtp_packet_received.h" #include "modules/rtp_rtcp/source/rtp_utility.h" #include "modules/utility/include/process_thread.h" #include "modules/video_coding/fec_controller_default.h" #include "rtc_base/checks.h" #include "rtc_base/constructor_magic.h" #include "rtc_base/location.h" #include "rtc_base/logging.h" #include "rtc_base/strings/string_builder.h" #include "rtc_base/synchronization/sequence_checker.h" #include "rtc_base/task_utils/pending_task_safety_flag.h" #include "rtc_base/thread_annotations.h" #include "rtc_base/time_utils.h" #include "rtc_base/trace_event.h" #include "system_wrappers/include/clock.h" #include "system_wrappers/include/cpu_info.h" #include "system_wrappers/include/field_trial.h" #include "system_wrappers/include/metrics.h" #include "video/call_stats2.h" #include "video/send_delay_stats.h" #include "video/stats_counter.h" #include "video/video_receive_stream2.h" #include "video/video_send_stream.h" namespace webrtc { namespace { bool SendPeriodicFeedback(const std::vector& extensions) { for (const auto& extension : extensions) { if (extension.uri == RtpExtension::kTransportSequenceNumberV2Uri) return false; } return true; } // TODO(nisse): This really begs for a shared context struct. bool UseSendSideBwe(const std::vector& extensions, bool transport_cc) { if (!transport_cc) return false; for (const auto& extension : extensions) { if (extension.uri == RtpExtension::kTransportSequenceNumberUri || extension.uri == RtpExtension::kTransportSequenceNumberV2Uri) return true; } return false; } bool UseSendSideBwe(const VideoReceiveStream::Config& config) { return UseSendSideBwe(config.rtp.extensions, config.rtp.transport_cc); } bool UseSendSideBwe(const AudioReceiveStream::Config& config) { return UseSendSideBwe(config.rtp.extensions, config.rtp.transport_cc); } bool UseSendSideBwe(const FlexfecReceiveStream::Config& config) { return UseSendSideBwe(config.rtp_header_extensions, config.transport_cc); } const int* FindKeyByValue(const std::map& m, int v) { for (const auto& kv : m) { if (kv.second == v) return &kv.first; } return nullptr; } std::unique_ptr CreateRtcLogStreamConfig( const VideoReceiveStream::Config& config) { auto rtclog_config = std::make_unique(); rtclog_config->remote_ssrc = config.rtp.remote_ssrc; rtclog_config->local_ssrc = config.rtp.local_ssrc; rtclog_config->rtx_ssrc = config.rtp.rtx_ssrc; rtclog_config->rtcp_mode = config.rtp.rtcp_mode; rtclog_config->rtp_extensions = config.rtp.extensions; for (const auto& d : config.decoders) { const int* search = FindKeyByValue(config.rtp.rtx_associated_payload_types, d.payload_type); rtclog_config->codecs.emplace_back(d.video_format.name, d.payload_type, search ? *search : 0); } return rtclog_config; } std::unique_ptr CreateRtcLogStreamConfig( const VideoSendStream::Config& config, size_t ssrc_index) { auto rtclog_config = std::make_unique(); rtclog_config->local_ssrc = config.rtp.ssrcs[ssrc_index]; if (ssrc_index < config.rtp.rtx.ssrcs.size()) { rtclog_config->rtx_ssrc = config.rtp.rtx.ssrcs[ssrc_index]; } rtclog_config->rtcp_mode = config.rtp.rtcp_mode; rtclog_config->rtp_extensions = config.rtp.extensions; rtclog_config->codecs.emplace_back(config.rtp.payload_name, config.rtp.payload_type, config.rtp.rtx.payload_type); return rtclog_config; } std::unique_ptr CreateRtcLogStreamConfig( const AudioReceiveStream::Config& config) { auto rtclog_config = std::make_unique(); rtclog_config->remote_ssrc = config.rtp.remote_ssrc; rtclog_config->local_ssrc = config.rtp.local_ssrc; rtclog_config->rtp_extensions = config.rtp.extensions; return rtclog_config; } bool IsRtcp(const uint8_t* packet, size_t length) { RtpUtility::RtpHeaderParser rtp_parser(packet, length); return rtp_parser.RTCP(); } TaskQueueBase* GetCurrentTaskQueueOrThread() { TaskQueueBase* current = TaskQueueBase::Current(); if (!current) current = rtc::ThreadManager::Instance()->CurrentThread(); return current; } } // namespace namespace internal { // Wraps an injected resource in a BroadcastResourceListener and handles adding // and removing adapter resources to individual VideoSendStreams. class ResourceVideoSendStreamForwarder { public: ResourceVideoSendStreamForwarder( rtc::scoped_refptr resource) : broadcast_resource_listener_(resource) { broadcast_resource_listener_.StartListening(); } ~ResourceVideoSendStreamForwarder() { RTC_DCHECK(adapter_resources_.empty()); broadcast_resource_listener_.StopListening(); } rtc::scoped_refptr Resource() const { return broadcast_resource_listener_.SourceResource(); } void OnCreateVideoSendStream(VideoSendStream* video_send_stream) { RTC_DCHECK(adapter_resources_.find(video_send_stream) == adapter_resources_.end()); auto adapter_resource = broadcast_resource_listener_.CreateAdapterResource(); video_send_stream->AddAdaptationResource(adapter_resource); adapter_resources_.insert( std::make_pair(video_send_stream, adapter_resource)); } void OnDestroyVideoSendStream(VideoSendStream* video_send_stream) { auto it = adapter_resources_.find(video_send_stream); RTC_DCHECK(it != adapter_resources_.end()); broadcast_resource_listener_.RemoveAdapterResource(it->second); adapter_resources_.erase(it); } private: BroadcastResourceListener broadcast_resource_listener_; std::map> adapter_resources_; }; class Call final : public webrtc::Call, public PacketReceiver, public RecoveredPacketReceiver, public TargetTransferRateObserver, public BitrateAllocator::LimitObserver { public: Call(Clock* clock, const Call::Config& config, std::unique_ptr transport_send, rtc::scoped_refptr module_process_thread, TaskQueueFactory* task_queue_factory); ~Call() override; // Implements webrtc::Call. PacketReceiver* Receiver() override; webrtc::AudioSendStream* CreateAudioSendStream( const webrtc::AudioSendStream::Config& config) override; void DestroyAudioSendStream(webrtc::AudioSendStream* send_stream) override; webrtc::AudioReceiveStream* CreateAudioReceiveStream( const webrtc::AudioReceiveStream::Config& config) override; void DestroyAudioReceiveStream( webrtc::AudioReceiveStream* receive_stream) override; webrtc::VideoSendStream* CreateVideoSendStream( webrtc::VideoSendStream::Config config, VideoEncoderConfig encoder_config) override; webrtc::VideoSendStream* CreateVideoSendStream( webrtc::VideoSendStream::Config config, VideoEncoderConfig encoder_config, std::unique_ptr fec_controller) override; void DestroyVideoSendStream(webrtc::VideoSendStream* send_stream) override; webrtc::VideoReceiveStream* CreateVideoReceiveStream( webrtc::VideoReceiveStream::Config configuration) override; void DestroyVideoReceiveStream( webrtc::VideoReceiveStream* receive_stream) override; FlexfecReceiveStream* CreateFlexfecReceiveStream( const FlexfecReceiveStream::Config& config) override; void DestroyFlexfecReceiveStream( FlexfecReceiveStream* receive_stream) override; void AddAdaptationResource(rtc::scoped_refptr resource) override; RtpTransportControllerSendInterface* GetTransportControllerSend() override; Stats GetStats() const override; // Implements PacketReceiver. DeliveryStatus DeliverPacket(MediaType media_type, rtc::CopyOnWriteBuffer packet, int64_t packet_time_us) override; // Implements RecoveredPacketReceiver. void OnRecoveredPacket(const uint8_t* packet, size_t length) override; void SignalChannelNetworkState(MediaType media, NetworkState state) override; void OnAudioTransportOverheadChanged( int transport_overhead_per_packet) override; void OnSentPacket(const rtc::SentPacket& sent_packet) override; // Implements TargetTransferRateObserver, void OnTargetTransferRate(TargetTransferRate msg) override; void OnStartRateUpdate(DataRate start_rate) override; // Implements BitrateAllocator::LimitObserver. void OnAllocationLimitsChanged(BitrateAllocationLimits limits) override; void SetClientBitratePreferences(const BitrateSettings& preferences) override; private: DeliveryStatus DeliverRtcp(MediaType media_type, const uint8_t* packet, size_t length) RTC_EXCLUSIVE_LOCKS_REQUIRED(worker_thread_); DeliveryStatus DeliverRtp(MediaType media_type, rtc::CopyOnWriteBuffer packet, int64_t packet_time_us) RTC_EXCLUSIVE_LOCKS_REQUIRED(worker_thread_); void ConfigureSync(const std::string& sync_group) RTC_EXCLUSIVE_LOCKS_REQUIRED(worker_thread_); void NotifyBweOfReceivedPacket(const RtpPacketReceived& packet, MediaType media_type) RTC_SHARED_LOCKS_REQUIRED(worker_thread_); void UpdateSendHistograms(Timestamp first_sent_packet) RTC_EXCLUSIVE_LOCKS_REQUIRED(worker_thread_); void UpdateReceiveHistograms(); void UpdateHistograms(); void UpdateAggregateNetworkState(); void RegisterRateObserver(); rtc::TaskQueue* send_transport_queue() const { return transport_send_ptr_->GetWorkerQueue(); } Clock* const clock_; TaskQueueFactory* const task_queue_factory_; TaskQueueBase* const worker_thread_; const int num_cpu_cores_; const rtc::scoped_refptr module_process_thread_; const std::unique_ptr call_stats_; const std::unique_ptr bitrate_allocator_; Call::Config config_; NetworkState audio_network_state_; NetworkState video_network_state_; bool aggregate_network_up_ RTC_GUARDED_BY(worker_thread_); // Audio, Video, and FlexFEC receive streams are owned by the client that // creates them. std::set audio_receive_streams_ RTC_GUARDED_BY(worker_thread_); std::set video_receive_streams_ RTC_GUARDED_BY(worker_thread_); std::map sync_stream_mapping_ RTC_GUARDED_BY(worker_thread_); // TODO(nisse): Should eventually be injected at creation, // with a single object in the bundled case. RtpStreamReceiverController audio_receiver_controller_; RtpStreamReceiverController video_receiver_controller_; // This extra map is used for receive processing which is // independent of media type. // TODO(nisse): In the RTP transport refactoring, we should have a // single mapping from ssrc to a more abstract receive stream, with // accessor methods for all configuration we need at this level. struct ReceiveRtpConfig { explicit ReceiveRtpConfig(const webrtc::AudioReceiveStream::Config& config) : extensions(config.rtp.extensions), use_send_side_bwe(UseSendSideBwe(config)) {} explicit ReceiveRtpConfig(const webrtc::VideoReceiveStream::Config& config) : extensions(config.rtp.extensions), use_send_side_bwe(UseSendSideBwe(config)) {} explicit ReceiveRtpConfig(const FlexfecReceiveStream::Config& config) : extensions(config.rtp_header_extensions), use_send_side_bwe(UseSendSideBwe(config)) {} // Registered RTP header extensions for each stream. Note that RTP header // extensions are negotiated per track ("m= line") in the SDP, but we have // no notion of tracks at the Call level. We therefore store the RTP header // extensions per SSRC instead, which leads to some storage overhead. const RtpHeaderExtensionMap extensions; // Set if both RTP extension the RTCP feedback message needed for // send side BWE are negotiated. const bool use_send_side_bwe; }; std::map receive_rtp_config_ RTC_GUARDED_BY(worker_thread_); // Audio and Video send streams are owned by the client that creates them. std::map audio_send_ssrcs_ RTC_GUARDED_BY(worker_thread_); std::map video_send_ssrcs_ RTC_GUARDED_BY(worker_thread_); std::set video_send_streams_ RTC_GUARDED_BY(worker_thread_); // Each forwarder wraps an adaptation resource that was added to the call. std::vector> adaptation_resource_forwarders_ RTC_GUARDED_BY(worker_thread_); using RtpStateMap = std::map; RtpStateMap suspended_audio_send_ssrcs_ RTC_GUARDED_BY(worker_thread_); RtpStateMap suspended_video_send_ssrcs_ RTC_GUARDED_BY(worker_thread_); using RtpPayloadStateMap = std::map; RtpPayloadStateMap suspended_video_payload_states_ RTC_GUARDED_BY(worker_thread_); webrtc::RtcEventLog* event_log_; // The following members are only accessed (exclusively) from one thread and // from the destructor, and therefore doesn't need any explicit // synchronization. RateCounter received_bytes_per_second_counter_; RateCounter received_audio_bytes_per_second_counter_; RateCounter received_video_bytes_per_second_counter_; RateCounter received_rtcp_bytes_per_second_counter_; absl::optional first_received_rtp_audio_ms_; absl::optional last_received_rtp_audio_ms_; absl::optional first_received_rtp_video_ms_; absl::optional last_received_rtp_video_ms_; uint32_t last_bandwidth_bps_ RTC_GUARDED_BY(worker_thread_); // TODO(holmer): Remove this lock once BitrateController no longer calls // OnNetworkChanged from multiple threads. uint32_t min_allocated_send_bitrate_bps_ RTC_GUARDED_BY(worker_thread_); uint32_t configured_max_padding_bitrate_bps_ RTC_GUARDED_BY(worker_thread_); AvgCounter estimated_send_bitrate_kbps_counter_ RTC_GUARDED_BY(worker_thread_); AvgCounter pacer_bitrate_kbps_counter_ RTC_GUARDED_BY(worker_thread_); ReceiveSideCongestionController receive_side_cc_; const std::unique_ptr receive_time_calculator_; const std::unique_ptr video_send_delay_stats_; const int64_t start_ms_; // Note that |task_safety_| needs to be at a greater scope than the task queue // owned by |transport_send_| since calls might arrive on the network thread // while Call is being deleted and the task queue is being torn down. ScopedTaskSafety task_safety_; // Caches transport_send_.get(), to avoid racing with destructor. // Note that this is declared before transport_send_ to ensure that it is not // invalidated until no more tasks can be running on the transport_send_ task // queue. RtpTransportControllerSendInterface* const transport_send_ptr_; // Declared last since it will issue callbacks from a task queue. Declaring it // last ensures that it is destroyed first and any running tasks are finished. std::unique_ptr transport_send_; bool is_target_rate_observer_registered_ RTC_GUARDED_BY(worker_thread_) = false; RTC_DISALLOW_COPY_AND_ASSIGN(Call); }; } // namespace internal std::string Call::Stats::ToString(int64_t time_ms) const { char buf[1024]; rtc::SimpleStringBuilder ss(buf); ss << "Call stats: " << time_ms << ", {"; ss << "send_bw_bps: " << send_bandwidth_bps << ", "; ss << "recv_bw_bps: " << recv_bandwidth_bps << ", "; ss << "max_pad_bps: " << max_padding_bitrate_bps << ", "; ss << "pacer_delay_ms: " << pacer_delay_ms << ", "; ss << "rtt_ms: " << rtt_ms; ss << '}'; return ss.str(); } Call* Call::Create(const Call::Config& config) { rtc::scoped_refptr call_thread = SharedModuleThread::Create(ProcessThread::Create("ModuleProcessThread"), nullptr); return Create(config, std::move(call_thread)); } Call* Call::Create(const Call::Config& config, rtc::scoped_refptr call_thread) { return Create(config, Clock::GetRealTimeClock(), std::move(call_thread), ProcessThread::Create("PacerThread")); } Call* Call::Create(const Call::Config& config, Clock* clock, rtc::scoped_refptr call_thread, std::unique_ptr pacer_thread) { RTC_DCHECK(config.task_queue_factory); return new internal::Call( clock, config, std::make_unique( clock, config.event_log, config.network_state_predictor_factory, config.network_controller_factory, config.bitrate_config, std::move(pacer_thread), config.task_queue_factory, config.trials), std::move(call_thread), config.task_queue_factory); } class SharedModuleThread::Impl { public: Impl(std::unique_ptr process_thread, std::function on_one_ref_remaining) : module_thread_(std::move(process_thread)), on_one_ref_remaining_(std::move(on_one_ref_remaining)) {} void EnsureStarted() { RTC_DCHECK_RUN_ON(&sequence_checker_); if (started_) return; started_ = true; module_thread_->Start(); } ProcessThread* process_thread() { RTC_DCHECK_RUN_ON(&sequence_checker_); return module_thread_.get(); } void AddRef() const { RTC_DCHECK_RUN_ON(&sequence_checker_); ++ref_count_; } rtc::RefCountReleaseStatus Release() const { RTC_DCHECK_RUN_ON(&sequence_checker_); --ref_count_; if (ref_count_ == 0) { module_thread_->Stop(); return rtc::RefCountReleaseStatus::kDroppedLastRef; } if (ref_count_ == 1 && on_one_ref_remaining_) { auto moved_fn = std::move(on_one_ref_remaining_); // NOTE: after this function returns, chances are that |this| has been // deleted - do not touch any member variables. // If the owner of the last reference implements a lambda that releases // that last reference inside of the callback (which is legal according // to this implementation), we will recursively enter Release() above, // call Stop() and release the last reference. moved_fn(); } return rtc::RefCountReleaseStatus::kOtherRefsRemained; } private: SequenceChecker sequence_checker_; mutable int ref_count_ RTC_GUARDED_BY(sequence_checker_) = 0; std::unique_ptr const module_thread_; std::function const on_one_ref_remaining_; bool started_ = false; }; SharedModuleThread::SharedModuleThread( std::unique_ptr process_thread, std::function on_one_ref_remaining) : impl_(std::make_unique(std::move(process_thread), std::move(on_one_ref_remaining))) {} SharedModuleThread::~SharedModuleThread() = default; // static rtc::scoped_refptr SharedModuleThread::Create( std::unique_ptr process_thread, std::function on_one_ref_remaining) { return new SharedModuleThread(std::move(process_thread), std::move(on_one_ref_remaining)); } void SharedModuleThread::EnsureStarted() { impl_->EnsureStarted(); } ProcessThread* SharedModuleThread::process_thread() { return impl_->process_thread(); } void SharedModuleThread::AddRef() const { impl_->AddRef(); } rtc::RefCountReleaseStatus SharedModuleThread::Release() const { auto ret = impl_->Release(); if (ret == rtc::RefCountReleaseStatus::kDroppedLastRef) delete this; return ret; } // This method here to avoid subclasses has to implement this method. // Call perf test will use Internal::Call::CreateVideoSendStream() to inject // FecController. VideoSendStream* Call::CreateVideoSendStream( VideoSendStream::Config config, VideoEncoderConfig encoder_config, std::unique_ptr fec_controller) { return nullptr; } namespace internal { Call::Call(Clock* clock, const Call::Config& config, std::unique_ptr transport_send, rtc::scoped_refptr module_process_thread, TaskQueueFactory* task_queue_factory) : clock_(clock), task_queue_factory_(task_queue_factory), worker_thread_(GetCurrentTaskQueueOrThread()), num_cpu_cores_(CpuInfo::DetectNumberOfCores()), module_process_thread_(std::move(module_process_thread)), call_stats_(new CallStats(clock_, worker_thread_)), bitrate_allocator_(new BitrateAllocator(this)), config_(config), audio_network_state_(kNetworkDown), video_network_state_(kNetworkDown), aggregate_network_up_(false), event_log_(config.event_log), received_bytes_per_second_counter_(clock_, nullptr, true), received_audio_bytes_per_second_counter_(clock_, nullptr, true), received_video_bytes_per_second_counter_(clock_, nullptr, true), received_rtcp_bytes_per_second_counter_(clock_, nullptr, true), last_bandwidth_bps_(0), min_allocated_send_bitrate_bps_(0), configured_max_padding_bitrate_bps_(0), estimated_send_bitrate_kbps_counter_(clock_, nullptr, true), pacer_bitrate_kbps_counter_(clock_, nullptr, true), receive_side_cc_(clock_, transport_send->packet_router()), receive_time_calculator_(ReceiveTimeCalculator::CreateFromFieldTrial()), video_send_delay_stats_(new SendDelayStats(clock_)), start_ms_(clock_->TimeInMilliseconds()), transport_send_ptr_(transport_send.get()), transport_send_(std::move(transport_send)) { RTC_DCHECK(config.event_log != nullptr); RTC_DCHECK(config.trials != nullptr); RTC_DCHECK(worker_thread_->IsCurrent()); call_stats_->RegisterStatsObserver(&receive_side_cc_); module_process_thread_->process_thread()->RegisterModule( receive_side_cc_.GetRemoteBitrateEstimator(true), RTC_FROM_HERE); module_process_thread_->process_thread()->RegisterModule(&receive_side_cc_, RTC_FROM_HERE); } Call::~Call() { RTC_DCHECK_RUN_ON(worker_thread_); RTC_CHECK(audio_send_ssrcs_.empty()); RTC_CHECK(video_send_ssrcs_.empty()); RTC_CHECK(video_send_streams_.empty()); RTC_CHECK(audio_receive_streams_.empty()); RTC_CHECK(video_receive_streams_.empty()); module_process_thread_->process_thread()->DeRegisterModule( receive_side_cc_.GetRemoteBitrateEstimator(true)); module_process_thread_->process_thread()->DeRegisterModule(&receive_side_cc_); call_stats_->DeregisterStatsObserver(&receive_side_cc_); absl::optional first_sent_packet_ms = transport_send_->GetFirstPacketTime(); // Only update histograms after process threads have been shut down, so that // they won't try to concurrently update stats. if (first_sent_packet_ms) { UpdateSendHistograms(*first_sent_packet_ms); } UpdateReceiveHistograms(); UpdateHistograms(); } void Call::RegisterRateObserver() { RTC_DCHECK_RUN_ON(worker_thread_); if (is_target_rate_observer_registered_) return; is_target_rate_observer_registered_ = true; // This call seems to kick off a number of things, so probably better left // off being kicked off on request rather than in the ctor. transport_send_ptr_->RegisterTargetTransferRateObserver(this); module_process_thread_->EnsureStarted(); } void Call::SetClientBitratePreferences(const BitrateSettings& preferences) { RTC_DCHECK_RUN_ON(worker_thread_); GetTransportControllerSend()->SetClientBitratePreferences(preferences); } void Call::UpdateHistograms() { RTC_HISTOGRAM_COUNTS_100000( "WebRTC.Call.LifetimeInSeconds", (clock_->TimeInMilliseconds() - start_ms_) / 1000); } // Called from the dtor. void Call::UpdateSendHistograms(Timestamp first_sent_packet) { int64_t elapsed_sec = (clock_->TimeInMilliseconds() - first_sent_packet.ms()) / 1000; if (elapsed_sec < metrics::kMinRunTimeInSeconds) return; const int kMinRequiredPeriodicSamples = 5; AggregatedStats send_bitrate_stats = estimated_send_bitrate_kbps_counter_.ProcessAndGetStats(); if (send_bitrate_stats.num_samples > kMinRequiredPeriodicSamples) { RTC_HISTOGRAM_COUNTS_100000("WebRTC.Call.EstimatedSendBitrateInKbps", send_bitrate_stats.average); RTC_LOG(LS_INFO) << "WebRTC.Call.EstimatedSendBitrateInKbps, " << send_bitrate_stats.ToString(); } AggregatedStats pacer_bitrate_stats = pacer_bitrate_kbps_counter_.ProcessAndGetStats(); if (pacer_bitrate_stats.num_samples > kMinRequiredPeriodicSamples) { RTC_HISTOGRAM_COUNTS_100000("WebRTC.Call.PacerBitrateInKbps", pacer_bitrate_stats.average); RTC_LOG(LS_INFO) << "WebRTC.Call.PacerBitrateInKbps, " << pacer_bitrate_stats.ToString(); } } void Call::UpdateReceiveHistograms() { if (first_received_rtp_audio_ms_) { RTC_HISTOGRAM_COUNTS_100000( "WebRTC.Call.TimeReceivingAudioRtpPacketsInSeconds", (*last_received_rtp_audio_ms_ - *first_received_rtp_audio_ms_) / 1000); } if (first_received_rtp_video_ms_) { RTC_HISTOGRAM_COUNTS_100000( "WebRTC.Call.TimeReceivingVideoRtpPacketsInSeconds", (*last_received_rtp_video_ms_ - *first_received_rtp_video_ms_) / 1000); } const int kMinRequiredPeriodicSamples = 5; AggregatedStats video_bytes_per_sec = received_video_bytes_per_second_counter_.GetStats(); if (video_bytes_per_sec.num_samples > kMinRequiredPeriodicSamples) { RTC_HISTOGRAM_COUNTS_100000("WebRTC.Call.VideoBitrateReceivedInKbps", video_bytes_per_sec.average * 8 / 1000); RTC_LOG(LS_INFO) << "WebRTC.Call.VideoBitrateReceivedInBps, " << video_bytes_per_sec.ToStringWithMultiplier(8); } AggregatedStats audio_bytes_per_sec = received_audio_bytes_per_second_counter_.GetStats(); if (audio_bytes_per_sec.num_samples > kMinRequiredPeriodicSamples) { RTC_HISTOGRAM_COUNTS_100000("WebRTC.Call.AudioBitrateReceivedInKbps", audio_bytes_per_sec.average * 8 / 1000); RTC_LOG(LS_INFO) << "WebRTC.Call.AudioBitrateReceivedInBps, " << audio_bytes_per_sec.ToStringWithMultiplier(8); } AggregatedStats rtcp_bytes_per_sec = received_rtcp_bytes_per_second_counter_.GetStats(); if (rtcp_bytes_per_sec.num_samples > kMinRequiredPeriodicSamples) { RTC_HISTOGRAM_COUNTS_100000("WebRTC.Call.RtcpBitrateReceivedInBps", rtcp_bytes_per_sec.average * 8); RTC_LOG(LS_INFO) << "WebRTC.Call.RtcpBitrateReceivedInBps, " << rtcp_bytes_per_sec.ToStringWithMultiplier(8); } AggregatedStats recv_bytes_per_sec = received_bytes_per_second_counter_.GetStats(); if (recv_bytes_per_sec.num_samples > kMinRequiredPeriodicSamples) { RTC_HISTOGRAM_COUNTS_100000("WebRTC.Call.BitrateReceivedInKbps", recv_bytes_per_sec.average * 8 / 1000); RTC_LOG(LS_INFO) << "WebRTC.Call.BitrateReceivedInBps, " << recv_bytes_per_sec.ToStringWithMultiplier(8); } } PacketReceiver* Call::Receiver() { RTC_DCHECK_RUN_ON(worker_thread_); return this; } webrtc::AudioSendStream* Call::CreateAudioSendStream( const webrtc::AudioSendStream::Config& config) { TRACE_EVENT0("webrtc", "Call::CreateAudioSendStream"); RTC_DCHECK_RUN_ON(worker_thread_); RegisterRateObserver(); // Stream config is logged in AudioSendStream::ConfigureStream, as it may // change during the stream's lifetime. absl::optional suspended_rtp_state; { const auto& iter = suspended_audio_send_ssrcs_.find(config.rtp.ssrc); if (iter != suspended_audio_send_ssrcs_.end()) { suspended_rtp_state.emplace(iter->second); } } AudioSendStream* send_stream = new AudioSendStream( clock_, config, config_.audio_state, task_queue_factory_, module_process_thread_->process_thread(), transport_send_ptr_, bitrate_allocator_.get(), event_log_, call_stats_->AsRtcpRttStats(), suspended_rtp_state); RTC_DCHECK(audio_send_ssrcs_.find(config.rtp.ssrc) == audio_send_ssrcs_.end()); audio_send_ssrcs_[config.rtp.ssrc] = send_stream; for (AudioReceiveStream* stream : audio_receive_streams_) { if (stream->config().rtp.local_ssrc == config.rtp.ssrc) { stream->AssociateSendStream(send_stream); } } UpdateAggregateNetworkState(); return send_stream; } void Call::DestroyAudioSendStream(webrtc::AudioSendStream* send_stream) { TRACE_EVENT0("webrtc", "Call::DestroyAudioSendStream"); RTC_DCHECK_RUN_ON(worker_thread_); RTC_DCHECK(send_stream != nullptr); send_stream->Stop(); const uint32_t ssrc = send_stream->GetConfig().rtp.ssrc; webrtc::internal::AudioSendStream* audio_send_stream = static_cast(send_stream); suspended_audio_send_ssrcs_[ssrc] = audio_send_stream->GetRtpState(); size_t num_deleted = audio_send_ssrcs_.erase(ssrc); RTC_DCHECK_EQ(1, num_deleted); for (AudioReceiveStream* stream : audio_receive_streams_) { if (stream->config().rtp.local_ssrc == ssrc) { stream->AssociateSendStream(nullptr); } } UpdateAggregateNetworkState(); delete send_stream; } webrtc::AudioReceiveStream* Call::CreateAudioReceiveStream( const webrtc::AudioReceiveStream::Config& config) { TRACE_EVENT0("webrtc", "Call::CreateAudioReceiveStream"); RTC_DCHECK_RUN_ON(worker_thread_); RegisterRateObserver(); event_log_->Log(std::make_unique( CreateRtcLogStreamConfig(config))); AudioReceiveStream* receive_stream = new AudioReceiveStream( clock_, &audio_receiver_controller_, transport_send_ptr_->packet_router(), module_process_thread_->process_thread(), config_.neteq_factory, config, config_.audio_state, event_log_); receive_rtp_config_.emplace(config.rtp.remote_ssrc, ReceiveRtpConfig(config)); audio_receive_streams_.insert(receive_stream); ConfigureSync(config.sync_group); auto it = audio_send_ssrcs_.find(config.rtp.local_ssrc); if (it != audio_send_ssrcs_.end()) { receive_stream->AssociateSendStream(it->second); } UpdateAggregateNetworkState(); return receive_stream; } void Call::DestroyAudioReceiveStream( webrtc::AudioReceiveStream* receive_stream) { TRACE_EVENT0("webrtc", "Call::DestroyAudioReceiveStream"); RTC_DCHECK_RUN_ON(worker_thread_); RTC_DCHECK(receive_stream != nullptr); webrtc::internal::AudioReceiveStream* audio_receive_stream = static_cast(receive_stream); const AudioReceiveStream::Config& config = audio_receive_stream->config(); uint32_t ssrc = config.rtp.remote_ssrc; receive_side_cc_.GetRemoteBitrateEstimator(UseSendSideBwe(config)) ->RemoveStream(ssrc); audio_receive_streams_.erase(audio_receive_stream); const std::string& sync_group = audio_receive_stream->config().sync_group; const auto it = sync_stream_mapping_.find(sync_group); if (it != sync_stream_mapping_.end() && it->second == audio_receive_stream) { sync_stream_mapping_.erase(it); ConfigureSync(sync_group); } receive_rtp_config_.erase(ssrc); UpdateAggregateNetworkState(); delete audio_receive_stream; } // This method can be used for Call tests with external fec controller factory. webrtc::VideoSendStream* Call::CreateVideoSendStream( webrtc::VideoSendStream::Config config, VideoEncoderConfig encoder_config, std::unique_ptr fec_controller) { TRACE_EVENT0("webrtc", "Call::CreateVideoSendStream"); RTC_DCHECK_RUN_ON(worker_thread_); RegisterRateObserver(); video_send_delay_stats_->AddSsrcs(config); for (size_t ssrc_index = 0; ssrc_index < config.rtp.ssrcs.size(); ++ssrc_index) { event_log_->Log(std::make_unique( CreateRtcLogStreamConfig(config, ssrc_index))); } // TODO(mflodman): Base the start bitrate on a current bandwidth estimate, if // the call has already started. // Copy ssrcs from |config| since |config| is moved. std::vector ssrcs = config.rtp.ssrcs; VideoSendStream* send_stream = new VideoSendStream( clock_, num_cpu_cores_, module_process_thread_->process_thread(), task_queue_factory_, call_stats_->AsRtcpRttStats(), transport_send_ptr_, bitrate_allocator_.get(), video_send_delay_stats_.get(), event_log_, std::move(config), std::move(encoder_config), suspended_video_send_ssrcs_, suspended_video_payload_states_, std::move(fec_controller)); for (uint32_t ssrc : ssrcs) { RTC_DCHECK(video_send_ssrcs_.find(ssrc) == video_send_ssrcs_.end()); video_send_ssrcs_[ssrc] = send_stream; } video_send_streams_.insert(send_stream); // Forward resources that were previously added to the call to the new stream. for (const auto& resource_forwarder : adaptation_resource_forwarders_) { resource_forwarder->OnCreateVideoSendStream(send_stream); } UpdateAggregateNetworkState(); return send_stream; } webrtc::VideoSendStream* Call::CreateVideoSendStream( webrtc::VideoSendStream::Config config, VideoEncoderConfig encoder_config) { if (config_.fec_controller_factory) { RTC_LOG(LS_INFO) << "External FEC Controller will be used."; } std::unique_ptr fec_controller = config_.fec_controller_factory ? config_.fec_controller_factory->CreateFecController() : std::make_unique(clock_); return CreateVideoSendStream(std::move(config), std::move(encoder_config), std::move(fec_controller)); } void Call::DestroyVideoSendStream(webrtc::VideoSendStream* send_stream) { TRACE_EVENT0("webrtc", "Call::DestroyVideoSendStream"); RTC_DCHECK(send_stream != nullptr); RTC_DCHECK_RUN_ON(worker_thread_); send_stream->Stop(); VideoSendStream* send_stream_impl = nullptr; auto it = video_send_ssrcs_.begin(); while (it != video_send_ssrcs_.end()) { if (it->second == static_cast(send_stream)) { send_stream_impl = it->second; video_send_ssrcs_.erase(it++); } else { ++it; } } // Stop forwarding resources to the stream being destroyed. for (const auto& resource_forwarder : adaptation_resource_forwarders_) { resource_forwarder->OnDestroyVideoSendStream(send_stream_impl); } video_send_streams_.erase(send_stream_impl); RTC_CHECK(send_stream_impl != nullptr); VideoSendStream::RtpStateMap rtp_states; VideoSendStream::RtpPayloadStateMap rtp_payload_states; send_stream_impl->StopPermanentlyAndGetRtpStates(&rtp_states, &rtp_payload_states); for (const auto& kv : rtp_states) { suspended_video_send_ssrcs_[kv.first] = kv.second; } for (const auto& kv : rtp_payload_states) { suspended_video_payload_states_[kv.first] = kv.second; } UpdateAggregateNetworkState(); delete send_stream_impl; } webrtc::VideoReceiveStream* Call::CreateVideoReceiveStream( webrtc::VideoReceiveStream::Config configuration) { TRACE_EVENT0("webrtc", "Call::CreateVideoReceiveStream"); RTC_DCHECK_RUN_ON(worker_thread_); receive_side_cc_.SetSendPeriodicFeedback( SendPeriodicFeedback(configuration.rtp.extensions)); RegisterRateObserver(); TaskQueueBase* current = GetCurrentTaskQueueOrThread(); RTC_CHECK(current); VideoReceiveStream2* receive_stream = new VideoReceiveStream2( task_queue_factory_, current, &video_receiver_controller_, num_cpu_cores_, transport_send_ptr_->packet_router(), std::move(configuration), module_process_thread_->process_thread(), call_stats_.get(), clock_, new VCMTiming(clock_)); const webrtc::VideoReceiveStream::Config& config = receive_stream->config(); if (config.rtp.rtx_ssrc) { // We record identical config for the rtx stream as for the main // stream. Since the transport_send_cc negotiation is per payload // type, we may get an incorrect value for the rtx stream, but // that is unlikely to matter in practice. receive_rtp_config_.emplace(config.rtp.rtx_ssrc, ReceiveRtpConfig(config)); } receive_rtp_config_.emplace(config.rtp.remote_ssrc, ReceiveRtpConfig(config)); video_receive_streams_.insert(receive_stream); ConfigureSync(config.sync_group); receive_stream->SignalNetworkState(video_network_state_); UpdateAggregateNetworkState(); event_log_->Log(std::make_unique( CreateRtcLogStreamConfig(config))); return receive_stream; } void Call::DestroyVideoReceiveStream( webrtc::VideoReceiveStream* receive_stream) { TRACE_EVENT0("webrtc", "Call::DestroyVideoReceiveStream"); RTC_DCHECK_RUN_ON(worker_thread_); RTC_DCHECK(receive_stream != nullptr); VideoReceiveStream2* receive_stream_impl = static_cast(receive_stream); const VideoReceiveStream::Config& config = receive_stream_impl->config(); // Remove all ssrcs pointing to a receive stream. As RTX retransmits on a // separate SSRC there can be either one or two. receive_rtp_config_.erase(config.rtp.remote_ssrc); if (config.rtp.rtx_ssrc) { receive_rtp_config_.erase(config.rtp.rtx_ssrc); } video_receive_streams_.erase(receive_stream_impl); ConfigureSync(config.sync_group); receive_side_cc_.GetRemoteBitrateEstimator(UseSendSideBwe(config)) ->RemoveStream(config.rtp.remote_ssrc); UpdateAggregateNetworkState(); delete receive_stream_impl; } FlexfecReceiveStream* Call::CreateFlexfecReceiveStream( const FlexfecReceiveStream::Config& config) { TRACE_EVENT0("webrtc", "Call::CreateFlexfecReceiveStream"); RTC_DCHECK_RUN_ON(worker_thread_); RecoveredPacketReceiver* recovered_packet_receiver = this; FlexfecReceiveStreamImpl* receive_stream; // Unlike the video and audio receive streams, FlexfecReceiveStream implements // RtpPacketSinkInterface itself, and hence its constructor passes its |this| // pointer to video_receiver_controller_->CreateStream(). Calling the // constructor while on the worker thread ensures that we don't call // OnRtpPacket until the constructor is finished and the object is // in a valid state, since OnRtpPacket runs on the same thread. receive_stream = new FlexfecReceiveStreamImpl( clock_, &video_receiver_controller_, config, recovered_packet_receiver, call_stats_->AsRtcpRttStats(), module_process_thread_->process_thread()); RTC_DCHECK(receive_rtp_config_.find(config.remote_ssrc) == receive_rtp_config_.end()); receive_rtp_config_.emplace(config.remote_ssrc, ReceiveRtpConfig(config)); // TODO(brandtr): Store config in RtcEventLog here. return receive_stream; } void Call::DestroyFlexfecReceiveStream(FlexfecReceiveStream* receive_stream) { TRACE_EVENT0("webrtc", "Call::DestroyFlexfecReceiveStream"); RTC_DCHECK_RUN_ON(worker_thread_); RTC_DCHECK(receive_stream != nullptr); const FlexfecReceiveStream::Config& config = receive_stream->GetConfig(); uint32_t ssrc = config.remote_ssrc; receive_rtp_config_.erase(ssrc); // Remove all SSRCs pointing to the FlexfecReceiveStreamImpl to be // destroyed. receive_side_cc_.GetRemoteBitrateEstimator(UseSendSideBwe(config)) ->RemoveStream(ssrc); delete receive_stream; } void Call::AddAdaptationResource(rtc::scoped_refptr resource) { RTC_DCHECK_RUN_ON(worker_thread_); adaptation_resource_forwarders_.push_back( std::make_unique(resource)); const auto& resource_forwarder = adaptation_resource_forwarders_.back(); for (VideoSendStream* send_stream : video_send_streams_) { resource_forwarder->OnCreateVideoSendStream(send_stream); } } RtpTransportControllerSendInterface* Call::GetTransportControllerSend() { return transport_send_ptr_; } Call::Stats Call::GetStats() const { RTC_DCHECK_RUN_ON(worker_thread_); Stats stats; // TODO(srte): It is unclear if we only want to report queues if network is // available. stats.pacer_delay_ms = aggregate_network_up_ ? transport_send_ptr_->GetPacerQueuingDelayMs() : 0; stats.rtt_ms = call_stats_->LastProcessedRtt(); // Fetch available send/receive bitrates. std::vector ssrcs; uint32_t recv_bandwidth = 0; receive_side_cc_.GetRemoteBitrateEstimator(false)->LatestEstimate( &ssrcs, &recv_bandwidth); stats.recv_bandwidth_bps = recv_bandwidth; stats.send_bandwidth_bps = last_bandwidth_bps_; stats.max_padding_bitrate_bps = configured_max_padding_bitrate_bps_; return stats; } void Call::SignalChannelNetworkState(MediaType media, NetworkState state) { RTC_DCHECK_RUN_ON(worker_thread_); switch (media) { case MediaType::AUDIO: audio_network_state_ = state; break; case MediaType::VIDEO: video_network_state_ = state; break; case MediaType::ANY: case MediaType::DATA: RTC_NOTREACHED(); break; } UpdateAggregateNetworkState(); for (VideoReceiveStream2* video_receive_stream : video_receive_streams_) { video_receive_stream->SignalNetworkState(video_network_state_); } } void Call::OnAudioTransportOverheadChanged(int transport_overhead_per_packet) { RTC_DCHECK_RUN_ON(worker_thread_); for (auto& kv : audio_send_ssrcs_) { kv.second->SetTransportOverhead(transport_overhead_per_packet); } } void Call::UpdateAggregateNetworkState() { RTC_DCHECK_RUN_ON(worker_thread_); bool have_audio = !audio_send_ssrcs_.empty() || !audio_receive_streams_.empty(); bool have_video = !video_send_ssrcs_.empty() || !video_receive_streams_.empty(); bool aggregate_network_up = ((have_video && video_network_state_ == kNetworkUp) || (have_audio && audio_network_state_ == kNetworkUp)); if (aggregate_network_up != aggregate_network_up_) { RTC_LOG(LS_INFO) << "UpdateAggregateNetworkState: aggregate_state change to " << (aggregate_network_up ? "up" : "down"); } else { RTC_LOG(LS_VERBOSE) << "UpdateAggregateNetworkState: aggregate_state remains at " << (aggregate_network_up ? "up" : "down"); } aggregate_network_up_ = aggregate_network_up; transport_send_ptr_->OnNetworkAvailability(aggregate_network_up); } void Call::OnSentPacket(const rtc::SentPacket& sent_packet) { video_send_delay_stats_->OnSentPacket(sent_packet.packet_id, clock_->TimeInMilliseconds()); transport_send_ptr_->OnSentPacket(sent_packet); } void Call::OnStartRateUpdate(DataRate start_rate) { RTC_DCHECK_RUN_ON(send_transport_queue()); bitrate_allocator_->UpdateStartRate(start_rate.bps()); } void Call::OnTargetTransferRate(TargetTransferRate msg) { RTC_DCHECK_RUN_ON(send_transport_queue()); uint32_t target_bitrate_bps = msg.target_rate.bps(); // For controlling the rate of feedback messages. receive_side_cc_.OnBitrateChanged(target_bitrate_bps); bitrate_allocator_->OnNetworkEstimateChanged(msg); worker_thread_->PostTask( ToQueuedTask(task_safety_, [this, target_bitrate_bps]() { RTC_DCHECK_RUN_ON(worker_thread_); last_bandwidth_bps_ = target_bitrate_bps; // Ignore updates if bitrate is zero (the aggregate network state is // down) or if we're not sending video. if (target_bitrate_bps == 0 || video_send_streams_.empty()) { estimated_send_bitrate_kbps_counter_.ProcessAndPause(); pacer_bitrate_kbps_counter_.ProcessAndPause(); return; } estimated_send_bitrate_kbps_counter_.Add(target_bitrate_bps / 1000); // Pacer bitrate may be higher than bitrate estimate if enforcing min // bitrate. uint32_t pacer_bitrate_bps = std::max(target_bitrate_bps, min_allocated_send_bitrate_bps_); pacer_bitrate_kbps_counter_.Add(pacer_bitrate_bps / 1000); })); } void Call::OnAllocationLimitsChanged(BitrateAllocationLimits limits) { RTC_DCHECK_RUN_ON(send_transport_queue()); transport_send_ptr_->SetAllocatedSendBitrateLimits(limits); worker_thread_->PostTask(ToQueuedTask(task_safety_, [this, limits]() { RTC_DCHECK_RUN_ON(worker_thread_); min_allocated_send_bitrate_bps_ = limits.min_allocatable_rate.bps(); configured_max_padding_bitrate_bps_ = limits.max_padding_rate.bps(); })); } void Call::ConfigureSync(const std::string& sync_group) { // Set sync only if there was no previous one. if (sync_group.empty()) return; AudioReceiveStream* sync_audio_stream = nullptr; // Find existing audio stream. const auto it = sync_stream_mapping_.find(sync_group); if (it != sync_stream_mapping_.end()) { sync_audio_stream = it->second; } else { // No configured audio stream, see if we can find one. for (AudioReceiveStream* stream : audio_receive_streams_) { if (stream->config().sync_group == sync_group) { if (sync_audio_stream != nullptr) { RTC_LOG(LS_WARNING) << "Attempting to sync more than one audio stream " "within the same sync group. This is not " "supported in the current implementation."; break; } sync_audio_stream = stream; } } } if (sync_audio_stream) sync_stream_mapping_[sync_group] = sync_audio_stream; size_t num_synced_streams = 0; for (VideoReceiveStream2* video_stream : video_receive_streams_) { if (video_stream->config().sync_group != sync_group) continue; ++num_synced_streams; if (num_synced_streams > 1) { // TODO(pbos): Support synchronizing more than one A/V pair. // https://code.google.com/p/webrtc/issues/detail?id=4762 RTC_LOG(LS_WARNING) << "Attempting to sync more than one audio/video pair " "within the same sync group. This is not supported in " "the current implementation."; } // Only sync the first A/V pair within this sync group. if (num_synced_streams == 1) { // sync_audio_stream may be null and that's ok. video_stream->SetSync(sync_audio_stream); } else { video_stream->SetSync(nullptr); } } } PacketReceiver::DeliveryStatus Call::DeliverRtcp(MediaType media_type, const uint8_t* packet, size_t length) { TRACE_EVENT0("webrtc", "Call::DeliverRtcp"); // TODO(pbos): Make sure it's a valid packet. // Return DELIVERY_UNKNOWN_SSRC if it can be determined that // there's no receiver of the packet. if (received_bytes_per_second_counter_.HasSample()) { // First RTP packet has been received. received_bytes_per_second_counter_.Add(static_cast(length)); received_rtcp_bytes_per_second_counter_.Add(static_cast(length)); } bool rtcp_delivered = false; if (media_type == MediaType::ANY || media_type == MediaType::VIDEO) { for (VideoReceiveStream2* stream : video_receive_streams_) { if (stream->DeliverRtcp(packet, length)) rtcp_delivered = true; } } if (media_type == MediaType::ANY || media_type == MediaType::AUDIO) { for (AudioReceiveStream* stream : audio_receive_streams_) { stream->DeliverRtcp(packet, length); rtcp_delivered = true; } } if (media_type == MediaType::ANY || media_type == MediaType::VIDEO) { for (VideoSendStream* stream : video_send_streams_) { stream->DeliverRtcp(packet, length); rtcp_delivered = true; } } if (media_type == MediaType::ANY || media_type == MediaType::AUDIO) { for (auto& kv : audio_send_ssrcs_) { kv.second->DeliverRtcp(packet, length); rtcp_delivered = true; } } if (rtcp_delivered) { event_log_->Log(std::make_unique( rtc::MakeArrayView(packet, length))); } return rtcp_delivered ? DELIVERY_OK : DELIVERY_PACKET_ERROR; } PacketReceiver::DeliveryStatus Call::DeliverRtp(MediaType media_type, rtc::CopyOnWriteBuffer packet, int64_t packet_time_us) { TRACE_EVENT0("webrtc", "Call::DeliverRtp"); RtpPacketReceived parsed_packet; if (!parsed_packet.Parse(std::move(packet))) return DELIVERY_PACKET_ERROR; if (packet_time_us != -1) { if (receive_time_calculator_) { // Repair packet_time_us for clock resets by comparing a new read of // the same clock (TimeUTCMicros) to a monotonic clock reading. packet_time_us = receive_time_calculator_->ReconcileReceiveTimes( packet_time_us, rtc::TimeUTCMicros(), clock_->TimeInMicroseconds()); } parsed_packet.set_arrival_time_ms((packet_time_us + 500) / 1000); } else { parsed_packet.set_arrival_time_ms(clock_->TimeInMilliseconds()); } // We might get RTP keep-alive packets in accordance with RFC6263 section 4.6. // These are empty (zero length payload) RTP packets with an unsignaled // payload type. const bool is_keep_alive_packet = parsed_packet.payload_size() == 0; RTC_DCHECK(media_type == MediaType::AUDIO || media_type == MediaType::VIDEO || is_keep_alive_packet); auto it = receive_rtp_config_.find(parsed_packet.Ssrc()); if (it == receive_rtp_config_.end()) { RTC_LOG(LS_ERROR) << "receive_rtp_config_ lookup failed for ssrc " << parsed_packet.Ssrc(); // Destruction of the receive stream, including deregistering from the // RtpDemuxer, is not protected by the |worker_thread_|. // But deregistering in the |receive_rtp_config_| map is. So by not passing // the packet on to demuxing in this case, we prevent incoming packets to be // passed on via the demuxer to a receive stream which is being torned down. return DELIVERY_UNKNOWN_SSRC; } parsed_packet.IdentifyExtensions(it->second.extensions); NotifyBweOfReceivedPacket(parsed_packet, media_type); // RateCounters expect input parameter as int, save it as int, // instead of converting each time it is passed to RateCounter::Add below. int length = static_cast(parsed_packet.size()); if (media_type == MediaType::AUDIO) { if (audio_receiver_controller_.OnRtpPacket(parsed_packet)) { received_bytes_per_second_counter_.Add(length); received_audio_bytes_per_second_counter_.Add(length); event_log_->Log( std::make_unique(parsed_packet)); const int64_t arrival_time_ms = parsed_packet.arrival_time_ms(); if (!first_received_rtp_audio_ms_) { first_received_rtp_audio_ms_.emplace(arrival_time_ms); } last_received_rtp_audio_ms_.emplace(arrival_time_ms); return DELIVERY_OK; } } else if (media_type == MediaType::VIDEO) { parsed_packet.set_payload_type_frequency(kVideoPayloadTypeFrequency); if (video_receiver_controller_.OnRtpPacket(parsed_packet)) { received_bytes_per_second_counter_.Add(length); received_video_bytes_per_second_counter_.Add(length); event_log_->Log( std::make_unique(parsed_packet)); const int64_t arrival_time_ms = parsed_packet.arrival_time_ms(); if (!first_received_rtp_video_ms_) { first_received_rtp_video_ms_.emplace(arrival_time_ms); } last_received_rtp_video_ms_.emplace(arrival_time_ms); return DELIVERY_OK; } } return DELIVERY_UNKNOWN_SSRC; } PacketReceiver::DeliveryStatus Call::DeliverPacket( MediaType media_type, rtc::CopyOnWriteBuffer packet, int64_t packet_time_us) { RTC_DCHECK_RUN_ON(worker_thread_); if (IsRtcp(packet.cdata(), packet.size())) return DeliverRtcp(media_type, packet.cdata(), packet.size()); return DeliverRtp(media_type, std::move(packet), packet_time_us); } void Call::OnRecoveredPacket(const uint8_t* packet, size_t length) { RTC_DCHECK_RUN_ON(worker_thread_); RtpPacketReceived parsed_packet; if (!parsed_packet.Parse(packet, length)) return; parsed_packet.set_recovered(true); auto it = receive_rtp_config_.find(parsed_packet.Ssrc()); if (it == receive_rtp_config_.end()) { RTC_LOG(LS_ERROR) << "receive_rtp_config_ lookup failed for ssrc " << parsed_packet.Ssrc(); // Destruction of the receive stream, including deregistering from the // RtpDemuxer, is not protected by the |worker_thread_|. // But deregistering in the |receive_rtp_config_| map is. // So by not passing the packet on to demuxing in this case, we prevent // incoming packets to be passed on via the demuxer to a receive stream // which is being torn down. return; } parsed_packet.IdentifyExtensions(it->second.extensions); // TODO(brandtr): Update here when we support protecting audio packets too. parsed_packet.set_payload_type_frequency(kVideoPayloadTypeFrequency); video_receiver_controller_.OnRtpPacket(parsed_packet); } void Call::NotifyBweOfReceivedPacket(const RtpPacketReceived& packet, MediaType media_type) { auto it = receive_rtp_config_.find(packet.Ssrc()); bool use_send_side_bwe = (it != receive_rtp_config_.end()) && it->second.use_send_side_bwe; RTPHeader header; packet.GetHeader(&header); ReceivedPacket packet_msg; packet_msg.size = DataSize::Bytes(packet.payload_size()); packet_msg.receive_time = Timestamp::Millis(packet.arrival_time_ms()); if (header.extension.hasAbsoluteSendTime) { packet_msg.send_time = header.extension.GetAbsoluteSendTimestamp(); } transport_send_ptr_->OnReceivedPacket(packet_msg); if (!use_send_side_bwe && header.extension.hasTransportSequenceNumber) { // Inconsistent configuration of send side BWE. Do nothing. // TODO(nisse): Without this check, we may produce RTCP feedback // packets even when not negotiated. But it would be cleaner to // move the check down to RTCPSender::SendFeedbackPacket, which // would also help the PacketRouter to select an appropriate rtp // module in the case that some, but not all, have RTCP feedback // enabled. return; } // For audio, we only support send side BWE. if (media_type == MediaType::VIDEO || (use_send_side_bwe && header.extension.hasTransportSequenceNumber)) { receive_side_cc_.OnReceivedPacket( packet.arrival_time_ms(), packet.payload_size() + packet.padding_size(), header); } } } // namespace internal } // namespace webrtc