mirror of
https://github.com/mollyim/webrtc.git
synced 2025-05-13 22:00:47 +01:00

Call is instantiated on what we traditionally call the 'worker thread' in PeerConnection terms. Call statistics are however gathered, processed and reported in a number of different ways, which results in a lot of locking, which is also unpredictable due to the those actions themselves contending with other parts of the system. Designating the worker thread as the general owner of the stats, helps us keeps things regular and avoids loading unrelated task queues/threads with reporting things like histograms or locking up due to a call to GetStats(). This is a reland of remaining changes from https://webrtc-review.googlesource.com/c/src/+/172847: This applies the changes from the above CL to the forked files and switches call.cc over to using the forked implementation. Bug: webrtc:11489 Change-Id: I93ad560500806ddd0e6df1448b1bcf5a1aae7583 Reviewed-on: https://webrtc-review.googlesource.com/c/src/+/174000 Reviewed-by: Mirko Bonadei <mbonadei@webrtc.org> Reviewed-by: Magnus Flodman <mflodman@webrtc.org> Reviewed-by: Danil Chapovalov <danilchap@webrtc.org> Commit-Queue: Tommi <tommi@webrtc.org> Cr-Commit-Position: refs/heads/master@{#31186}
266 lines
10 KiB
C++
266 lines
10 KiB
C++
/*
|
|
* Copyright (c) 2012 The WebRTC project authors. All Rights Reserved.
|
|
*
|
|
* Use of this source code is governed by a BSD-style license
|
|
* that can be found in the LICENSE file in the root of the source
|
|
* tree. An additional intellectual property rights grant can be found
|
|
* in the file PATENTS. All contributing project authors may
|
|
* be found in the AUTHORS file in the root of the source tree.
|
|
*/
|
|
|
|
#include "modules/video_coding/generic_decoder.h"
|
|
|
|
#include <stddef.h>
|
|
|
|
#include <algorithm>
|
|
|
|
#include "api/video/video_timing.h"
|
|
#include "modules/video_coding/include/video_error_codes.h"
|
|
#include "rtc_base/checks.h"
|
|
#include "rtc_base/logging.h"
|
|
#include "rtc_base/thread.h"
|
|
#include "rtc_base/time_utils.h"
|
|
#include "rtc_base/trace_event.h"
|
|
#include "system_wrappers/include/clock.h"
|
|
#include "system_wrappers/include/field_trial.h"
|
|
|
|
namespace webrtc {
|
|
|
|
VCMDecodedFrameCallback::VCMDecodedFrameCallback(VCMTiming* timing,
|
|
Clock* clock)
|
|
: _clock(clock),
|
|
_timing(timing),
|
|
_timestampMap(kDecoderFrameMemoryLength),
|
|
_extra_decode_time("t", absl::nullopt) {
|
|
ntp_offset_ =
|
|
_clock->CurrentNtpInMilliseconds() - _clock->TimeInMilliseconds();
|
|
|
|
ParseFieldTrial({&_extra_decode_time},
|
|
field_trial::FindFullName("WebRTC-SlowDownDecoder"));
|
|
}
|
|
|
|
VCMDecodedFrameCallback::~VCMDecodedFrameCallback() {}
|
|
|
|
void VCMDecodedFrameCallback::SetUserReceiveCallback(
|
|
VCMReceiveCallback* receiveCallback) {
|
|
RTC_DCHECK(construction_thread_.IsCurrent());
|
|
RTC_DCHECK((!_receiveCallback && receiveCallback) ||
|
|
(_receiveCallback && !receiveCallback));
|
|
_receiveCallback = receiveCallback;
|
|
}
|
|
|
|
VCMReceiveCallback* VCMDecodedFrameCallback::UserReceiveCallback() {
|
|
// Called on the decode thread via VCMCodecDataBase::GetDecoder.
|
|
// The callback must always have been set before this happens.
|
|
RTC_DCHECK(_receiveCallback);
|
|
return _receiveCallback;
|
|
}
|
|
|
|
int32_t VCMDecodedFrameCallback::Decoded(VideoFrame& decodedImage) {
|
|
// This function may be called on the decode TaskQueue, but may also be called
|
|
// on an OS provided queue such as on iOS (see e.g. b/153465112).
|
|
return Decoded(decodedImage, -1);
|
|
}
|
|
|
|
int32_t VCMDecodedFrameCallback::Decoded(VideoFrame& decodedImage,
|
|
int64_t decode_time_ms) {
|
|
Decoded(decodedImage,
|
|
decode_time_ms >= 0 ? absl::optional<int32_t>(decode_time_ms)
|
|
: absl::nullopt,
|
|
absl::nullopt);
|
|
return WEBRTC_VIDEO_CODEC_OK;
|
|
}
|
|
|
|
void VCMDecodedFrameCallback::Decoded(VideoFrame& decodedImage,
|
|
absl::optional<int32_t> decode_time_ms,
|
|
absl::optional<uint8_t> qp) {
|
|
// Wait some extra time to simulate a slow decoder.
|
|
if (_extra_decode_time) {
|
|
rtc::Thread::SleepMs(_extra_decode_time->ms());
|
|
}
|
|
|
|
RTC_DCHECK(_receiveCallback) << "Callback must not be null at this point";
|
|
TRACE_EVENT_INSTANT1("webrtc", "VCMDecodedFrameCallback::Decoded",
|
|
"timestamp", decodedImage.timestamp());
|
|
// TODO(holmer): We should improve this so that we can handle multiple
|
|
// callbacks from one call to Decode().
|
|
VCMFrameInformation* frameInfo;
|
|
{
|
|
rtc::CritScope cs(&lock_);
|
|
frameInfo = _timestampMap.Pop(decodedImage.timestamp());
|
|
}
|
|
|
|
if (frameInfo == NULL) {
|
|
RTC_LOG(LS_WARNING) << "Too many frames backed up in the decoder, dropping "
|
|
"this one.";
|
|
_receiveCallback->OnDroppedFrames(1);
|
|
return;
|
|
}
|
|
|
|
decodedImage.set_ntp_time_ms(frameInfo->ntp_time_ms);
|
|
decodedImage.set_packet_infos(frameInfo->packet_infos);
|
|
decodedImage.set_rotation(frameInfo->rotation);
|
|
|
|
const Timestamp now = _clock->CurrentTime();
|
|
RTC_DCHECK(frameInfo->decodeStart);
|
|
if (!decode_time_ms) {
|
|
decode_time_ms = (now - *frameInfo->decodeStart).ms();
|
|
}
|
|
_timing->StopDecodeTimer(*decode_time_ms, now.ms());
|
|
decodedImage.set_processing_time({*frameInfo->decodeStart, now});
|
|
|
|
// Report timing information.
|
|
TimingFrameInfo timing_frame_info;
|
|
if (frameInfo->timing.flags != VideoSendTiming::kInvalid) {
|
|
int64_t capture_time_ms = decodedImage.ntp_time_ms() - ntp_offset_;
|
|
// Convert remote timestamps to local time from ntp timestamps.
|
|
frameInfo->timing.encode_start_ms -= ntp_offset_;
|
|
frameInfo->timing.encode_finish_ms -= ntp_offset_;
|
|
frameInfo->timing.packetization_finish_ms -= ntp_offset_;
|
|
frameInfo->timing.pacer_exit_ms -= ntp_offset_;
|
|
frameInfo->timing.network_timestamp_ms -= ntp_offset_;
|
|
frameInfo->timing.network2_timestamp_ms -= ntp_offset_;
|
|
|
|
int64_t sender_delta_ms = 0;
|
|
if (decodedImage.ntp_time_ms() < 0) {
|
|
// Sender clock is not estimated yet. Make sure that sender times are all
|
|
// negative to indicate that. Yet they still should be relatively correct.
|
|
sender_delta_ms =
|
|
std::max({capture_time_ms, frameInfo->timing.encode_start_ms,
|
|
frameInfo->timing.encode_finish_ms,
|
|
frameInfo->timing.packetization_finish_ms,
|
|
frameInfo->timing.pacer_exit_ms,
|
|
frameInfo->timing.network_timestamp_ms,
|
|
frameInfo->timing.network2_timestamp_ms}) +
|
|
1;
|
|
}
|
|
|
|
timing_frame_info.capture_time_ms = capture_time_ms - sender_delta_ms;
|
|
timing_frame_info.encode_start_ms =
|
|
frameInfo->timing.encode_start_ms - sender_delta_ms;
|
|
timing_frame_info.encode_finish_ms =
|
|
frameInfo->timing.encode_finish_ms - sender_delta_ms;
|
|
timing_frame_info.packetization_finish_ms =
|
|
frameInfo->timing.packetization_finish_ms - sender_delta_ms;
|
|
timing_frame_info.pacer_exit_ms =
|
|
frameInfo->timing.pacer_exit_ms - sender_delta_ms;
|
|
timing_frame_info.network_timestamp_ms =
|
|
frameInfo->timing.network_timestamp_ms - sender_delta_ms;
|
|
timing_frame_info.network2_timestamp_ms =
|
|
frameInfo->timing.network2_timestamp_ms - sender_delta_ms;
|
|
}
|
|
|
|
timing_frame_info.flags = frameInfo->timing.flags;
|
|
timing_frame_info.decode_start_ms = frameInfo->decodeStart->ms();
|
|
timing_frame_info.decode_finish_ms = now.ms();
|
|
timing_frame_info.render_time_ms = frameInfo->renderTimeMs;
|
|
timing_frame_info.rtp_timestamp = decodedImage.timestamp();
|
|
timing_frame_info.receive_start_ms = frameInfo->timing.receive_start_ms;
|
|
timing_frame_info.receive_finish_ms = frameInfo->timing.receive_finish_ms;
|
|
_timing->SetTimingFrameInfo(timing_frame_info);
|
|
|
|
decodedImage.set_timestamp_us(frameInfo->renderTimeMs *
|
|
rtc::kNumMicrosecsPerMillisec);
|
|
_receiveCallback->FrameToRender(decodedImage, qp, *decode_time_ms,
|
|
frameInfo->content_type);
|
|
}
|
|
|
|
void VCMDecodedFrameCallback::OnDecoderImplementationName(
|
|
const char* implementation_name) {
|
|
_receiveCallback->OnDecoderImplementationName(implementation_name);
|
|
}
|
|
|
|
void VCMDecodedFrameCallback::Map(uint32_t timestamp,
|
|
VCMFrameInformation* frameInfo) {
|
|
rtc::CritScope cs(&lock_);
|
|
_timestampMap.Add(timestamp, frameInfo);
|
|
}
|
|
|
|
int32_t VCMDecodedFrameCallback::Pop(uint32_t timestamp) {
|
|
rtc::CritScope cs(&lock_);
|
|
if (_timestampMap.Pop(timestamp) == NULL) {
|
|
return VCM_GENERAL_ERROR;
|
|
}
|
|
_receiveCallback->OnDroppedFrames(1);
|
|
return VCM_OK;
|
|
}
|
|
|
|
VCMGenericDecoder::VCMGenericDecoder(std::unique_ptr<VideoDecoder> decoder)
|
|
: VCMGenericDecoder(decoder.release(), false /* isExternal */) {}
|
|
|
|
VCMGenericDecoder::VCMGenericDecoder(VideoDecoder* decoder, bool isExternal)
|
|
: _callback(NULL),
|
|
_frameInfos(),
|
|
_nextFrameInfoIdx(0),
|
|
decoder_(decoder),
|
|
_codecType(kVideoCodecGeneric),
|
|
_isExternal(isExternal),
|
|
_last_keyframe_content_type(VideoContentType::UNSPECIFIED) {
|
|
RTC_DCHECK(decoder_);
|
|
}
|
|
|
|
VCMGenericDecoder::~VCMGenericDecoder() {
|
|
decoder_->Release();
|
|
if (_isExternal)
|
|
decoder_.release();
|
|
RTC_DCHECK(_isExternal || decoder_);
|
|
}
|
|
|
|
int32_t VCMGenericDecoder::InitDecode(const VideoCodec* settings,
|
|
int32_t numberOfCores) {
|
|
TRACE_EVENT0("webrtc", "VCMGenericDecoder::InitDecode");
|
|
_codecType = settings->codecType;
|
|
|
|
return decoder_->InitDecode(settings, numberOfCores);
|
|
}
|
|
|
|
int32_t VCMGenericDecoder::Decode(const VCMEncodedFrame& frame, Timestamp now) {
|
|
TRACE_EVENT1("webrtc", "VCMGenericDecoder::Decode", "timestamp",
|
|
frame.Timestamp());
|
|
_frameInfos[_nextFrameInfoIdx].decodeStart = now;
|
|
_frameInfos[_nextFrameInfoIdx].renderTimeMs = frame.RenderTimeMs();
|
|
_frameInfos[_nextFrameInfoIdx].rotation = frame.rotation();
|
|
_frameInfos[_nextFrameInfoIdx].timing = frame.video_timing();
|
|
_frameInfos[_nextFrameInfoIdx].ntp_time_ms =
|
|
frame.EncodedImage().ntp_time_ms_;
|
|
_frameInfos[_nextFrameInfoIdx].packet_infos = frame.PacketInfos();
|
|
|
|
// Set correctly only for key frames. Thus, use latest key frame
|
|
// content type. If the corresponding key frame was lost, decode will fail
|
|
// and content type will be ignored.
|
|
if (frame.FrameType() == VideoFrameType::kVideoFrameKey) {
|
|
_frameInfos[_nextFrameInfoIdx].content_type = frame.contentType();
|
|
_last_keyframe_content_type = frame.contentType();
|
|
} else {
|
|
_frameInfos[_nextFrameInfoIdx].content_type = _last_keyframe_content_type;
|
|
}
|
|
_callback->Map(frame.Timestamp(), &_frameInfos[_nextFrameInfoIdx]);
|
|
|
|
_nextFrameInfoIdx = (_nextFrameInfoIdx + 1) % kDecoderFrameMemoryLength;
|
|
int32_t ret = decoder_->Decode(frame.EncodedImage(), frame.MissingFrame(),
|
|
frame.RenderTimeMs());
|
|
|
|
_callback->OnDecoderImplementationName(decoder_->ImplementationName());
|
|
if (ret < WEBRTC_VIDEO_CODEC_OK) {
|
|
RTC_LOG(LS_WARNING) << "Failed to decode frame with timestamp "
|
|
<< frame.Timestamp() << ", error code: " << ret;
|
|
_callback->Pop(frame.Timestamp());
|
|
return ret;
|
|
} else if (ret == WEBRTC_VIDEO_CODEC_NO_OUTPUT) {
|
|
// No output
|
|
_callback->Pop(frame.Timestamp());
|
|
}
|
|
return ret;
|
|
}
|
|
|
|
int32_t VCMGenericDecoder::RegisterDecodeCompleteCallback(
|
|
VCMDecodedFrameCallback* callback) {
|
|
_callback = callback;
|
|
return decoder_->RegisterDecodeCompleteCallback(callback);
|
|
}
|
|
|
|
bool VCMGenericDecoder::PrefersLateDecoding() const {
|
|
return decoder_->PrefersLateDecoding();
|
|
}
|
|
|
|
} // namespace webrtc
|