/* * Copyright (c) 2012 The WebRTC project authors. All Rights Reserved. * * Use of this source code is governed by a BSD-style license * that can be found in the LICENSE file in the root of the source * tree. An additional intellectual property rights grant can be found * in the file PATENTS. All contributing project authors may * be found in the AUTHORS file in the root of the source tree. */ #ifndef COMMON_TYPES_H_ #define COMMON_TYPES_H_ #include #include #include #include #include "api/array_view.h" #include "api/optional.h" // TODO(sprang): Remove this include when all usage includes it directly. #include "api/video/video_bitrate_allocation.h" #include "rtc_base/checks.h" #include "rtc_base/deprecation.h" #include "typedefs.h" // NOLINT(build/include) #if defined(_MSC_VER) // Disable "new behavior: elements of array will be default initialized" // warning. Affects OverUseDetectorOptions. #pragma warning(disable : 4351) #endif #ifndef NULL #define NULL 0 #endif #define RTP_PAYLOAD_NAME_SIZE 32u #if defined(WEBRTC_WIN) || defined(WIN32) // Compares two strings without regard to case. #define STR_CASE_CMP(s1, s2) ::_stricmp(s1, s2) // Compares characters of two strings without regard to case. #define STR_NCASE_CMP(s1, s2, n) ::_strnicmp(s1, s2, n) #else #define STR_CASE_CMP(s1, s2) ::strcasecmp(s1, s2) #define STR_NCASE_CMP(s1, s2, n) ::strncasecmp(s1, s2, n) #endif namespace webrtc { enum FrameType { kEmptyFrame = 0, kAudioFrameSpeech = 1, kAudioFrameCN = 2, kVideoFrameKey = 3, kVideoFrameDelta = 4, }; // Statistics for an RTCP channel struct RtcpStatistics { RtcpStatistics() : fraction_lost(0), packets_lost(0), extended_highest_sequence_number(0), jitter(0) {} uint8_t fraction_lost; union { int32_t packets_lost; // Defined as a 24 bit signed integer in RTCP RTC_DEPRECATED uint32_t cumulative_lost; }; union { uint32_t extended_highest_sequence_number; RTC_DEPRECATED uint32_t extended_max_sequence_number; }; uint32_t jitter; }; class RtcpStatisticsCallback { public: virtual ~RtcpStatisticsCallback() {} virtual void StatisticsUpdated(const RtcpStatistics& statistics, uint32_t ssrc) = 0; virtual void CNameChanged(const char* cname, uint32_t ssrc) = 0; }; // Statistics for RTCP packet types. struct RtcpPacketTypeCounter { RtcpPacketTypeCounter() : first_packet_time_ms(-1), nack_packets(0), fir_packets(0), pli_packets(0), nack_requests(0), unique_nack_requests(0) {} void Add(const RtcpPacketTypeCounter& other) { nack_packets += other.nack_packets; fir_packets += other.fir_packets; pli_packets += other.pli_packets; nack_requests += other.nack_requests; unique_nack_requests += other.unique_nack_requests; if (other.first_packet_time_ms != -1 && (other.first_packet_time_ms < first_packet_time_ms || first_packet_time_ms == -1)) { // Use oldest time. first_packet_time_ms = other.first_packet_time_ms; } } void Subtract(const RtcpPacketTypeCounter& other) { nack_packets -= other.nack_packets; fir_packets -= other.fir_packets; pli_packets -= other.pli_packets; nack_requests -= other.nack_requests; unique_nack_requests -= other.unique_nack_requests; if (other.first_packet_time_ms != -1 && (other.first_packet_time_ms > first_packet_time_ms || first_packet_time_ms == -1)) { // Use youngest time. first_packet_time_ms = other.first_packet_time_ms; } } int64_t TimeSinceFirstPacketInMs(int64_t now_ms) const { return (first_packet_time_ms == -1) ? -1 : (now_ms - first_packet_time_ms); } int UniqueNackRequestsInPercent() const { if (nack_requests == 0) { return 0; } return static_cast((unique_nack_requests * 100.0f / nack_requests) + 0.5f); } int64_t first_packet_time_ms; // Time when first packet is sent/received. uint32_t nack_packets; // Number of RTCP NACK packets. uint32_t fir_packets; // Number of RTCP FIR packets. uint32_t pli_packets; // Number of RTCP PLI packets. uint32_t nack_requests; // Number of NACKed RTP packets. uint32_t unique_nack_requests; // Number of unique NACKed RTP packets. }; class RtcpPacketTypeCounterObserver { public: virtual ~RtcpPacketTypeCounterObserver() {} virtual void RtcpPacketTypesCounterUpdated( uint32_t ssrc, const RtcpPacketTypeCounter& packet_counter) = 0; }; // Callback, used to notify an observer whenever new rates have been estimated. class BitrateStatisticsObserver { public: virtual ~BitrateStatisticsObserver() {} virtual void Notify(uint32_t total_bitrate_bps, uint32_t retransmit_bitrate_bps, uint32_t ssrc) = 0; }; struct FrameCounts { FrameCounts() : key_frames(0), delta_frames(0) {} int key_frames; int delta_frames; }; // Callback, used to notify an observer whenever frame counts have been updated. class FrameCountObserver { public: virtual ~FrameCountObserver() {} virtual void FrameCountUpdated(const FrameCounts& frame_counts, uint32_t ssrc) = 0; }; // Callback, used to notify an observer whenever the send-side delay is updated. class SendSideDelayObserver { public: virtual ~SendSideDelayObserver() {} virtual void SendSideDelayUpdated(int avg_delay_ms, int max_delay_ms, uint32_t ssrc) = 0; }; // Callback, used to notify an observer whenever a packet is sent to the // transport. // TODO(asapersson): This class will remove the need for SendSideDelayObserver. // Remove SendSideDelayObserver once possible. class SendPacketObserver { public: virtual ~SendPacketObserver() {} virtual void OnSendPacket(uint16_t packet_id, int64_t capture_time_ms, uint32_t ssrc) = 0; }; // Callback, used to notify an observer when the overhead per packet // has changed. class OverheadObserver { public: virtual ~OverheadObserver() = default; virtual void OnOverheadChanged(size_t overhead_bytes_per_packet) = 0; }; // ================================================================== // Voice specific types // ================================================================== // Each codec supported can be described by this structure. struct CodecInst { int pltype; char plname[RTP_PAYLOAD_NAME_SIZE]; int plfreq; int pacsize; size_t channels; int rate; // bits/sec unlike {start,min,max}Bitrate elsewhere in this file! bool operator==(const CodecInst& other) const { return pltype == other.pltype && (STR_CASE_CMP(plname, other.plname) == 0) && plfreq == other.plfreq && pacsize == other.pacsize && channels == other.channels && rate == other.rate; } bool operator!=(const CodecInst& other) const { return !(*this == other); } }; // RTP enum { kRtpCsrcSize = 15 }; // RFC 3550 page 13 // NETEQ statistics. struct NetworkStatistics { // current jitter buffer size in ms uint16_t currentBufferSize; // preferred (optimal) buffer size in ms uint16_t preferredBufferSize; // adding extra delay due to "peaky jitter" bool jitterPeaksFound; // Stats below correspond to similarly-named fields in the WebRTC stats spec. // https://w3c.github.io/webrtc-stats/#dom-rtcmediastreamtrackstats uint64_t totalSamplesReceived; uint64_t concealedSamples; uint64_t concealmentEvents; uint64_t jitterBufferDelayMs; // Stats below DO NOT correspond directly to anything in the WebRTC stats // Loss rate (network + late); fraction between 0 and 1, scaled to Q14. uint16_t currentPacketLossRate; // Late loss rate; fraction between 0 and 1, scaled to Q14. union { RTC_DEPRECATED uint16_t currentDiscardRate; }; // fraction (of original stream) of synthesized audio inserted through // expansion (in Q14) uint16_t currentExpandRate; // fraction (of original stream) of synthesized speech inserted through // expansion (in Q14) uint16_t currentSpeechExpandRate; // fraction of synthesized speech inserted through pre-emptive expansion // (in Q14) uint16_t currentPreemptiveRate; // fraction of data removed through acceleration (in Q14) uint16_t currentAccelerateRate; // fraction of data coming from secondary decoding (in Q14) uint16_t currentSecondaryDecodedRate; // Fraction of secondary data, including FEC and RED, that is discarded (in // Q14). Discarding of secondary data can be caused by the reception of the // primary data, obsoleting the secondary data. It can also be caused by early // or late arrival of secondary data. uint16_t currentSecondaryDiscardedRate; // clock-drift in parts-per-million (negative or positive) int32_t clockDriftPPM; // average packet waiting time in the jitter buffer (ms) int meanWaitingTimeMs; // median packet waiting time in the jitter buffer (ms) int medianWaitingTimeMs; // min packet waiting time in the jitter buffer (ms) int minWaitingTimeMs; // max packet waiting time in the jitter buffer (ms) int maxWaitingTimeMs; // added samples in off mode due to packet loss size_t addedSamples; }; // Statistics for calls to AudioCodingModule::PlayoutData10Ms(). struct AudioDecodingCallStats { AudioDecodingCallStats() : calls_to_silence_generator(0), calls_to_neteq(0), decoded_normal(0), decoded_plc(0), decoded_cng(0), decoded_plc_cng(0), decoded_muted_output(0) {} int calls_to_silence_generator; // Number of calls where silence generated, // and NetEq was disengaged from decoding. int calls_to_neteq; // Number of calls to NetEq. int decoded_normal; // Number of calls where audio RTP packet decoded. int decoded_plc; // Number of calls resulted in PLC. int decoded_cng; // Number of calls where comfort noise generated due to DTX. int decoded_plc_cng; // Number of calls resulted where PLC faded to CNG. int decoded_muted_output; // Number of calls returning a muted state output. }; // ================================================================== // Video specific types // ================================================================== // TODO(nisse): Delete, and switch to fourcc values everywhere? // Supported video types. enum class VideoType { kUnknown, kI420, kIYUV, kRGB24, kABGR, kARGB, kARGB4444, kRGB565, kARGB1555, kYUY2, kYV12, kUYVY, kMJPEG, kNV21, kNV12, kBGRA, }; // Video codec enum VideoCodecComplexity { kComplexityNormal = 0, kComplexityHigh = 1, kComplexityHigher = 2, kComplexityMax = 3 }; // VP8 specific struct VideoCodecVP8 { bool operator==(const VideoCodecVP8& other) const; bool operator!=(const VideoCodecVP8& other) const { return !(*this == other); } VideoCodecComplexity complexity; unsigned char numberOfTemporalLayers; bool denoisingOn; bool automaticResizeOn; bool frameDroppingOn; int keyFrameInterval; }; enum class InterLayerPredMode { kOn, // Allow inter-layer prediction for all frames. // Frame of low spatial layer can be used for // prediction of next spatial layer frame. kOff, // Encoder produces independent spatial layers. kOnKeyPic // Allow inter-layer prediction only for frames // within key picture. }; // VP9 specific. struct VideoCodecVP9 { bool operator==(const VideoCodecVP9& other) const; bool operator!=(const VideoCodecVP9& other) const { return !(*this == other); } VideoCodecComplexity complexity; unsigned char numberOfTemporalLayers; bool denoisingOn; bool frameDroppingOn; int keyFrameInterval; bool adaptiveQpMode; bool automaticResizeOn; unsigned char numberOfSpatialLayers; bool flexibleMode; InterLayerPredMode interLayerPred; }; // TODO(magjed): Move this and other H264 related classes out to their own file. namespace H264 { enum Profile { kProfileConstrainedBaseline, kProfileBaseline, kProfileMain, kProfileConstrainedHigh, kProfileHigh, }; } // namespace H264 // H264 specific. struct VideoCodecH264 { bool operator==(const VideoCodecH264& other) const; bool operator!=(const VideoCodecH264& other) const { return !(*this == other); } bool frameDroppingOn; int keyFrameInterval; // These are NULL/0 if not externally negotiated. const uint8_t* spsData; size_t spsLen; const uint8_t* ppsData; size_t ppsLen; H264::Profile profile; }; // Video codec types enum VideoCodecType { kVideoCodecVP8, kVideoCodecVP9, kVideoCodecH264, kVideoCodecI420, kVideoCodecRED, kVideoCodecULPFEC, kVideoCodecFlexfec, kVideoCodecGeneric, kVideoCodecMultiplex, kVideoCodecUnknown }; // Translates from name of codec to codec type and vice versa. const char* CodecTypeToPayloadString(VideoCodecType type); VideoCodecType PayloadStringToCodecType(const std::string& name); union VideoCodecUnion { VideoCodecVP8 VP8; VideoCodecVP9 VP9; VideoCodecH264 H264; }; struct SpatialLayer { bool operator==(const SpatialLayer& other) const; bool operator!=(const SpatialLayer& other) const { return !(*this == other); } unsigned short width; unsigned short height; unsigned char numberOfTemporalLayers; unsigned int maxBitrate; // kilobits/sec. unsigned int targetBitrate; // kilobits/sec. unsigned int minBitrate; // kilobits/sec. unsigned int qpMax; // minimum quality bool active; // encoded and sent. }; // Simulcast is when the same stream is encoded multiple times with different // settings such as resolution. typedef SpatialLayer SimulcastStream; enum VideoCodecMode { kRealtimeVideo, kScreensharing }; // Common video codec properties class VideoCodec { public: VideoCodec(); // Public variables. TODO(hta): Make them private with accessors. VideoCodecType codecType; unsigned char plType; unsigned short width; unsigned short height; unsigned int startBitrate; // kilobits/sec. unsigned int maxBitrate; // kilobits/sec. unsigned int minBitrate; // kilobits/sec. unsigned int targetBitrate; // kilobits/sec. uint32_t maxFramerate; // This enables/disables encoding and sending when there aren't multiple // simulcast streams,by allocating 0 bitrate if inactive. bool active; unsigned int qpMax; unsigned char numberOfSimulcastStreams; SimulcastStream simulcastStream[kMaxSimulcastStreams]; SpatialLayer spatialLayers[kMaxSpatialLayers]; VideoCodecMode mode; bool expect_encode_from_texture; // Timing frames configuration. There is delay of delay_ms between two // consequent timing frames, excluding outliers. Frame is always made a // timing frame if it's at least outlier_ratio in percent of "ideal" average // frame given bitrate and framerate, i.e. if it's bigger than // |outlier_ratio / 100.0 * bitrate_bps / fps| in bits. This way, timing // frames will not be sent too often usually. Yet large frames will always // have timing information for debug purposes because they are more likely to // cause extra delays. struct TimingFrameTriggerThresholds { int64_t delay_ms; uint16_t outlier_ratio_percent; } timing_frame_thresholds; bool operator==(const VideoCodec& other) const = delete; bool operator!=(const VideoCodec& other) const = delete; // Accessors for codec specific information. // There is a const version of each that returns a reference, // and a non-const version that returns a pointer, in order // to allow modification of the parameters. VideoCodecVP8* VP8(); const VideoCodecVP8& VP8() const; VideoCodecVP9* VP9(); const VideoCodecVP9& VP9() const; VideoCodecH264* H264(); const VideoCodecH264& H264() const; private: // TODO(hta): Consider replacing the union with a pointer type. // This will allow removing the VideoCodec* types from this file. VideoCodecUnion codec_specific_; }; // TODO(sprang): Remove this when downstream projects have been updated. using BitrateAllocation = VideoBitrateAllocation; // Bandwidth over-use detector options. These are used to drive // experimentation with bandwidth estimation parameters. // See modules/remote_bitrate_estimator/overuse_detector.h // TODO(terelius): This is only used in overuse_estimator.cc, and only in the // default constructed state. Can we move the relevant variables into that // class and delete this? See also disabled warning at line 27 struct OverUseDetectorOptions { OverUseDetectorOptions() : initial_slope(8.0 / 512.0), initial_offset(0), initial_e(), initial_process_noise(), initial_avg_noise(0.0), initial_var_noise(50) { initial_e[0][0] = 100; initial_e[1][1] = 1e-1; initial_e[0][1] = initial_e[1][0] = 0; initial_process_noise[0] = 1e-13; initial_process_noise[1] = 1e-3; } double initial_slope; double initial_offset; double initial_e[2][2]; double initial_process_noise[2]; double initial_avg_noise; double initial_var_noise; }; // This structure will have the information about when packet is actually // received by socket. struct PacketTime { PacketTime() : timestamp(-1), not_before(-1) {} PacketTime(int64_t timestamp, int64_t not_before) : timestamp(timestamp), not_before(not_before) {} int64_t timestamp; // Receive time after socket delivers the data. int64_t not_before; // Earliest possible time the data could have arrived, // indicating the potential error in the |timestamp| // value,in case the system is busy. // For example, the time of the last select() call. // If unknown, this value will be set to zero. }; // Minimum and maximum playout delay values from capture to render. // These are best effort values. // // A value < 0 indicates no change from previous valid value. // // min = max = 0 indicates that the receiver should try and render // frame as soon as possible. // // min = x, max = y indicates that the receiver is free to adapt // in the range (x, y) based on network jitter. // // Note: Given that this gets embedded in a union, it is up-to the owner to // initialize these values. struct PlayoutDelay { int min_ms; int max_ms; }; } // namespace webrtc #endif // COMMON_TYPES_H_