/* * Copyright (c) 2020 The WebRTC project authors. All Rights Reserved. * * Use of this source code is governed by a BSD-style license * that can be found in the LICENSE file in the root of the source * tree. An additional intellectual property rights grant can be found * in the file PATENTS. All contributing project authors may * be found in the AUTHORS file in the root of the source tree. */ #include "modules/video_coding/codecs/av1/libaom_av1_encoder.h" #include #include #include #include #include #include "absl/algorithm/container.h" #include "absl/base/macros.h" #include "absl/base/nullability.h" #include "absl/strings/match.h" #include "absl/types/optional.h" #include "api/environment/environment.h" #include "api/field_trials_view.h" #include "api/scoped_refptr.h" #include "api/transport/field_trial_based_config.h" #include "api/video/encoded_image.h" #include "api/video/i420_buffer.h" #include "api/video/video_frame.h" #include "api/video_codecs/scalability_mode.h" #include "api/video_codecs/video_codec.h" #include "api/video_codecs/video_encoder.h" #include "modules/video_coding/include/video_codec_interface.h" #include "modules/video_coding/include/video_error_codes.h" #include "modules/video_coding/svc/create_scalability_structure.h" #include "modules/video_coding/svc/scalable_video_controller.h" #include "modules/video_coding/svc/scalable_video_controller_no_layering.h" #include "rtc_base/checks.h" #include "rtc_base/experiments/encoder_info_settings.h" #include "rtc_base/logging.h" #include "third_party/libaom/source/libaom/aom/aom_codec.h" #include "third_party/libaom/source/libaom/aom/aom_encoder.h" #include "third_party/libaom/source/libaom/aom/aomcx.h" #if (defined(WEBRTC_ARCH_ARM) || defined(WEBRTC_ARCH_ARM64)) && \ (defined(WEBRTC_ANDROID) || defined(WEBRTC_IOS)) #define MOBILE_ARM #endif #define SET_ENCODER_PARAM_OR_RETURN_ERROR(param_id, param_value) \ do { \ if (!SetEncoderControlParameters(param_id, param_value)) { \ return WEBRTC_VIDEO_CODEC_ERROR; \ } \ } while (0) namespace webrtc { namespace { // Encoder configuration parameters constexpr int kQpMin = 10; constexpr int kUsageProfile = AOM_USAGE_REALTIME; constexpr int kMinQindex = 145; // Min qindex threshold for QP scaling. constexpr int kMaxQindex = 205; // Max qindex threshold for QP scaling. constexpr int kBitDepth = 8; constexpr int kLagInFrames = 0; // No look ahead. constexpr int kRtpTicksPerSecond = 90000; constexpr double kMinimumFrameRate = 1.0; aom_superblock_size_t GetSuperblockSize(int width, int height, int threads) { int resolution = width * height; if (threads >= 4 && resolution >= 960 * 540 && resolution < 1920 * 1080) return AOM_SUPERBLOCK_SIZE_64X64; else return AOM_SUPERBLOCK_SIZE_DYNAMIC; } class LibaomAv1Encoder final : public VideoEncoder { public: LibaomAv1Encoder(const absl::optional& aux_config, const FieldTrialsView& trials); ~LibaomAv1Encoder(); int InitEncode(const VideoCodec* codec_settings, const Settings& settings) override; int32_t RegisterEncodeCompleteCallback( EncodedImageCallback* encoded_image_callback) override; int32_t Release() override; int32_t Encode(const VideoFrame& frame, const std::vector* frame_types) override; void SetRates(const RateControlParameters& parameters) override; EncoderInfo GetEncoderInfo() const override; private: template bool SetEncoderControlParameters(int param_id, P param_value); // Get value to be used for encoder cpu_speed setting int GetCpuSpeed(int width, int height); // Determine number of encoder threads to use. int NumberOfThreads(int width, int height, int number_of_cores); bool SvcEnabled() const { return svc_params_.has_value(); } // Fills svc_params_ memeber value. Returns false on error. bool SetSvcParams(ScalableVideoController::StreamLayersConfig svc_config); // Configures the encoder with layer for the next frame. void SetSvcLayerId( const ScalableVideoController::LayerFrameConfig& layer_frame); // Configures the encoder which buffers next frame updates and can reference. void SetSvcRefFrameConfig( const ScalableVideoController::LayerFrameConfig& layer_frame); // If pixel format doesn't match, then reallocate. void MaybeRewrapImgWithFormat(const aom_img_fmt_t fmt); std::unique_ptr svc_controller_; absl::optional scalability_mode_; bool inited_; bool rates_configured_; absl::optional svc_params_; VideoCodec encoder_settings_; absl::optional aux_config_; aom_image_t* frame_for_encode_; aom_codec_ctx_t ctx_; aom_codec_enc_cfg_t cfg_; EncodedImageCallback* encoded_image_callback_; int64_t timestamp_; const LibaomAv1EncoderInfoSettings encoder_info_override_; // TODO(webrtc:15225): Kill switch for disabling frame dropping. Remove it // after frame dropping is fully rolled out. bool disable_frame_dropping_; int max_consec_frame_drop_; }; int32_t VerifyCodecSettings(const VideoCodec& codec_settings) { if (codec_settings.width < 1) { return WEBRTC_VIDEO_CODEC_ERR_PARAMETER; } if (codec_settings.height < 1) { return WEBRTC_VIDEO_CODEC_ERR_PARAMETER; } // maxBitrate == 0 represents an unspecified maxBitRate. if (codec_settings.maxBitrate > 0 && codec_settings.minBitrate > codec_settings.maxBitrate) { return WEBRTC_VIDEO_CODEC_ERR_PARAMETER; } if (codec_settings.maxBitrate > 0 && codec_settings.startBitrate > codec_settings.maxBitrate) { return WEBRTC_VIDEO_CODEC_ERR_PARAMETER; } if (codec_settings.startBitrate < codec_settings.minBitrate) { return WEBRTC_VIDEO_CODEC_ERR_PARAMETER; } if (codec_settings.maxFramerate < 1) { return WEBRTC_VIDEO_CODEC_ERR_PARAMETER; } if (codec_settings.qpMax < kQpMin || codec_settings.qpMax > 63) { return WEBRTC_VIDEO_CODEC_ERR_PARAMETER; } return WEBRTC_VIDEO_CODEC_OK; } int GetMaxConsecutiveFrameDrop(const FieldTrialsView& field_trials) { webrtc::FieldTrialParameter maxdrop("maxdrop", 0); webrtc::ParseFieldTrial( {&maxdrop}, field_trials.Lookup("WebRTC-LibaomAv1Encoder-MaxConsecFrameDrop")); return maxdrop; } LibaomAv1Encoder::LibaomAv1Encoder( const absl::optional& aux_config, const FieldTrialsView& trials) : inited_(false), rates_configured_(false), aux_config_(aux_config), frame_for_encode_(nullptr), encoded_image_callback_(nullptr), timestamp_(0), disable_frame_dropping_(absl::StartsWith( trials.Lookup("WebRTC-LibaomAv1Encoder-DisableFrameDropping"), "Enabled")), max_consec_frame_drop_(GetMaxConsecutiveFrameDrop(trials)) {} LibaomAv1Encoder::~LibaomAv1Encoder() { Release(); } int LibaomAv1Encoder::InitEncode(const VideoCodec* codec_settings, const Settings& settings) { if (codec_settings == nullptr) { RTC_LOG(LS_WARNING) << "No codec settings provided to " "LibaomAv1Encoder."; return WEBRTC_VIDEO_CODEC_ERR_PARAMETER; } if (settings.number_of_cores < 1) { return WEBRTC_VIDEO_CODEC_ERR_PARAMETER; } if (inited_) { RTC_LOG(LS_WARNING) << "Initing LibaomAv1Encoder without first releasing."; Release(); } encoder_settings_ = *codec_settings; // Sanity checks for encoder configuration. const int32_t result = VerifyCodecSettings(encoder_settings_); if (result < 0) { RTC_LOG(LS_WARNING) << "Incorrect codec settings provided to " "LibaomAv1Encoder."; return result; } if (encoder_settings_.numberOfSimulcastStreams > 1) { RTC_LOG(LS_WARNING) << "Simulcast is not implemented by LibaomAv1Encoder."; return result; } scalability_mode_ = encoder_settings_.GetScalabilityMode(); if (!scalability_mode_.has_value()) { RTC_LOG(LS_WARNING) << "Scalability mode is not set, using 'L1T1'."; scalability_mode_ = ScalabilityMode::kL1T1; } svc_controller_ = CreateScalabilityStructure(*scalability_mode_); if (svc_controller_ == nullptr) { RTC_LOG(LS_WARNING) << "Failed to set scalability mode " << static_cast(*scalability_mode_); return WEBRTC_VIDEO_CODEC_ERR_PARAMETER; } if (!SetSvcParams(svc_controller_->StreamConfig())) { return WEBRTC_VIDEO_CODEC_ERROR; } // Initialize encoder configuration structure with default values aom_codec_err_t ret = aom_codec_enc_config_default(aom_codec_av1_cx(), &cfg_, kUsageProfile); if (ret != AOM_CODEC_OK) { RTC_LOG(LS_WARNING) << "LibaomAv1Encoder::EncodeInit returned " << ret << " on aom_codec_enc_config_default."; return WEBRTC_VIDEO_CODEC_ERROR; } // Overwrite default config with input encoder settings & RTC-relevant values. cfg_.g_w = encoder_settings_.width; cfg_.g_h = encoder_settings_.height; cfg_.g_threads = NumberOfThreads(cfg_.g_w, cfg_.g_h, settings.number_of_cores); cfg_.g_timebase.num = 1; cfg_.g_timebase.den = kRtpTicksPerSecond; cfg_.rc_target_bitrate = encoder_settings_.startBitrate; // kilobits/sec. cfg_.rc_dropframe_thresh = (!disable_frame_dropping_ && encoder_settings_.GetFrameDropEnabled()) ? 30 : 0; cfg_.g_input_bit_depth = kBitDepth; cfg_.kf_mode = AOM_KF_DISABLED; cfg_.rc_min_quantizer = kQpMin; cfg_.rc_max_quantizer = encoder_settings_.qpMax; cfg_.rc_undershoot_pct = 50; cfg_.rc_overshoot_pct = 50; cfg_.rc_buf_initial_sz = 600; cfg_.rc_buf_optimal_sz = 600; cfg_.rc_buf_sz = 1000; cfg_.g_usage = kUsageProfile; cfg_.g_error_resilient = 0; // Low-latency settings. cfg_.rc_end_usage = AOM_CBR; // Constant Bit Rate (CBR) mode cfg_.g_pass = AOM_RC_ONE_PASS; // One-pass rate control cfg_.g_lag_in_frames = kLagInFrames; // No look ahead when lag equals 0. if (frame_for_encode_ != nullptr) { aom_img_free(frame_for_encode_); frame_for_encode_ = nullptr; } // Flag options: AOM_CODEC_USE_PSNR and AOM_CODEC_USE_HIGHBITDEPTH aom_codec_flags_t flags = 0; // Initialize an encoder instance. ret = aom_codec_enc_init(&ctx_, aom_codec_av1_cx(), &cfg_, flags); if (ret != AOM_CODEC_OK) { RTC_LOG(LS_WARNING) << "LibaomAv1Encoder::EncodeInit returned " << ret << " on aom_codec_enc_init."; return WEBRTC_VIDEO_CODEC_ERROR; } inited_ = true; // Set control parameters SET_ENCODER_PARAM_OR_RETURN_ERROR(AOME_SET_CPUUSED, GetCpuSpeed(cfg_.g_w, cfg_.g_h)); SET_ENCODER_PARAM_OR_RETURN_ERROR(AV1E_SET_ENABLE_CDEF, 1); SET_ENCODER_PARAM_OR_RETURN_ERROR(AV1E_SET_ENABLE_TPL_MODEL, 0); SET_ENCODER_PARAM_OR_RETURN_ERROR(AV1E_SET_DELTAQ_MODE, 0); SET_ENCODER_PARAM_OR_RETURN_ERROR(AV1E_SET_ENABLE_ORDER_HINT, 0); SET_ENCODER_PARAM_OR_RETURN_ERROR(AV1E_SET_AQ_MODE, 3); SET_ENCODER_PARAM_OR_RETURN_ERROR(AOME_SET_MAX_INTRA_BITRATE_PCT, 300); SET_ENCODER_PARAM_OR_RETURN_ERROR(AV1E_SET_COEFF_COST_UPD_FREQ, 3); SET_ENCODER_PARAM_OR_RETURN_ERROR(AV1E_SET_MODE_COST_UPD_FREQ, 3); SET_ENCODER_PARAM_OR_RETURN_ERROR(AV1E_SET_MV_COST_UPD_FREQ, 3); if (codec_settings->mode == VideoCodecMode::kScreensharing) { SET_ENCODER_PARAM_OR_RETURN_ERROR(AV1E_SET_TUNE_CONTENT, AOM_CONTENT_SCREEN); SET_ENCODER_PARAM_OR_RETURN_ERROR(AV1E_SET_ENABLE_PALETTE, 1); } else { SET_ENCODER_PARAM_OR_RETURN_ERROR(AV1E_SET_ENABLE_PALETTE, 0); } if (codec_settings->mode == VideoCodecMode::kRealtimeVideo && encoder_settings_.GetFrameDropEnabled() && max_consec_frame_drop_ > 0) { SET_ENCODER_PARAM_OR_RETURN_ERROR(AV1E_SET_MAX_CONSEC_FRAME_DROP_CBR, max_consec_frame_drop_); } if (cfg_.g_threads == 8) { // Values passed to AV1E_SET_TILE_ROWS and AV1E_SET_TILE_COLUMNS are log2() // based. // Use 4 tile columns x 2 tile rows for 8 threads. SET_ENCODER_PARAM_OR_RETURN_ERROR(AV1E_SET_TILE_ROWS, 1); SET_ENCODER_PARAM_OR_RETURN_ERROR(AV1E_SET_TILE_COLUMNS, 2); } else if (cfg_.g_threads == 4) { // Use 2 tile columns x 2 tile rows for 4 threads. SET_ENCODER_PARAM_OR_RETURN_ERROR(AV1E_SET_TILE_ROWS, 1); SET_ENCODER_PARAM_OR_RETURN_ERROR(AV1E_SET_TILE_COLUMNS, 1); } else { SET_ENCODER_PARAM_OR_RETURN_ERROR(AV1E_SET_TILE_COLUMNS, static_cast(log2(cfg_.g_threads))); } SET_ENCODER_PARAM_OR_RETURN_ERROR(AV1E_SET_ROW_MT, 1); SET_ENCODER_PARAM_OR_RETURN_ERROR(AV1E_SET_ENABLE_OBMC, 0); SET_ENCODER_PARAM_OR_RETURN_ERROR(AV1E_SET_NOISE_SENSITIVITY, 0); SET_ENCODER_PARAM_OR_RETURN_ERROR(AV1E_SET_ENABLE_WARPED_MOTION, 0); SET_ENCODER_PARAM_OR_RETURN_ERROR(AV1E_SET_ENABLE_GLOBAL_MOTION, 0); SET_ENCODER_PARAM_OR_RETURN_ERROR(AV1E_SET_ENABLE_REF_FRAME_MVS, 0); SET_ENCODER_PARAM_OR_RETURN_ERROR( AV1E_SET_SUPERBLOCK_SIZE, GetSuperblockSize(cfg_.g_w, cfg_.g_h, cfg_.g_threads)); SET_ENCODER_PARAM_OR_RETURN_ERROR(AV1E_SET_ENABLE_CFL_INTRA, 0); SET_ENCODER_PARAM_OR_RETURN_ERROR(AV1E_SET_ENABLE_SMOOTH_INTRA, 0); SET_ENCODER_PARAM_OR_RETURN_ERROR(AV1E_SET_ENABLE_ANGLE_DELTA, 0); SET_ENCODER_PARAM_OR_RETURN_ERROR(AV1E_SET_ENABLE_FILTER_INTRA, 0); SET_ENCODER_PARAM_OR_RETURN_ERROR(AV1E_SET_INTRA_DEFAULT_TX_ONLY, 1); SET_ENCODER_PARAM_OR_RETURN_ERROR(AV1E_SET_DISABLE_TRELLIS_QUANT, 1); SET_ENCODER_PARAM_OR_RETURN_ERROR(AV1E_SET_ENABLE_DIST_WTD_COMP, 0); SET_ENCODER_PARAM_OR_RETURN_ERROR(AV1E_SET_ENABLE_DIFF_WTD_COMP, 0); SET_ENCODER_PARAM_OR_RETURN_ERROR(AV1E_SET_ENABLE_DUAL_FILTER, 0); SET_ENCODER_PARAM_OR_RETURN_ERROR(AV1E_SET_ENABLE_INTERINTRA_COMP, 0); SET_ENCODER_PARAM_OR_RETURN_ERROR(AV1E_SET_ENABLE_INTERINTRA_WEDGE, 0); SET_ENCODER_PARAM_OR_RETURN_ERROR(AV1E_SET_ENABLE_INTRA_EDGE_FILTER, 0); SET_ENCODER_PARAM_OR_RETURN_ERROR(AV1E_SET_ENABLE_INTRABC, 0); SET_ENCODER_PARAM_OR_RETURN_ERROR(AV1E_SET_ENABLE_MASKED_COMP, 0); SET_ENCODER_PARAM_OR_RETURN_ERROR(AV1E_SET_ENABLE_PAETH_INTRA, 0); SET_ENCODER_PARAM_OR_RETURN_ERROR(AV1E_SET_ENABLE_QM, 0); SET_ENCODER_PARAM_OR_RETURN_ERROR(AV1E_SET_ENABLE_RECT_PARTITIONS, 0); SET_ENCODER_PARAM_OR_RETURN_ERROR(AV1E_SET_ENABLE_RESTORATION, 0); SET_ENCODER_PARAM_OR_RETURN_ERROR(AV1E_SET_ENABLE_SMOOTH_INTERINTRA, 0); SET_ENCODER_PARAM_OR_RETURN_ERROR(AV1E_SET_ENABLE_TX64, 0); SET_ENCODER_PARAM_OR_RETURN_ERROR(AV1E_SET_MAX_REFERENCE_FRAMES, 3); return WEBRTC_VIDEO_CODEC_OK; } template bool LibaomAv1Encoder::SetEncoderControlParameters(int param_id, P param_value) { aom_codec_err_t error_code = aom_codec_control(&ctx_, param_id, param_value); if (error_code != AOM_CODEC_OK) { RTC_LOG(LS_WARNING) << "LibaomAv1Encoder::SetEncoderControlParameters returned " << error_code << " on id: " << param_id << "."; } return error_code == AOM_CODEC_OK; } // Only positive speeds, range for real-time coding currently is: 6 - 10. // Speed 11 is used for screen sharing. // Lower means slower/better quality, higher means fastest/lower quality. int LibaomAv1Encoder::GetCpuSpeed(int width, int height) { if (aux_config_) { if (auto it = aux_config_->max_pixel_count_to_cpu_speed.lower_bound(width * height); it != aux_config_->max_pixel_count_to_cpu_speed.end()) { return it->second; } return 10; } else { if (encoder_settings_.mode == VideoCodecMode::kScreensharing) { return 11; } // For smaller resolutions, use lower speed setting (get some coding gain at // the cost of increased encoding complexity). switch (encoder_settings_.GetVideoEncoderComplexity()) { case VideoCodecComplexity::kComplexityHigh: if (width * height <= 320 * 180) return 8; else if (width * height <= 640 * 360) return 9; else return 10; case VideoCodecComplexity::kComplexityHigher: if (width * height <= 320 * 180) return 7; else if (width * height <= 640 * 360) return 8; else if (width * height <= 1280 * 720) return 9; else return 10; case VideoCodecComplexity::kComplexityMax: if (width * height <= 320 * 180) return 6; else if (width * height <= 640 * 360) return 7; else if (width * height <= 1280 * 720) return 8; else return 9; default: return 10; } } } int LibaomAv1Encoder::NumberOfThreads(int width, int height, int number_of_cores) { // Keep the number of encoder threads equal to the possible number of // column/row tiles, which is (1, 2, 4, 8). See comments below for // AV1E_SET_TILE_COLUMNS/ROWS. if (width * height > 1280 * 720 && number_of_cores > 8) { return 8; } else if (width * height >= 640 * 360 && number_of_cores > 4) { return 4; } else if (width * height >= 320 * 180 && number_of_cores > 2) { return 2; } else { // Use 2 threads for low res on ARM. #ifdef MOBILE_ARM if (width * height >= 320 * 180 && number_of_cores > 2) { return 2; } #endif // 1 thread less than VGA. return 1; } } bool LibaomAv1Encoder::SetSvcParams( ScalableVideoController::StreamLayersConfig svc_config) { bool svc_enabled = svc_config.num_spatial_layers > 1 || svc_config.num_temporal_layers > 1; if (!svc_enabled) { svc_params_ = absl::nullopt; return true; } if (svc_config.num_spatial_layers < 1 || svc_config.num_spatial_layers > 4) { RTC_LOG(LS_WARNING) << "Av1 supports up to 4 spatial layers. " << svc_config.num_spatial_layers << " configured."; return false; } if (svc_config.num_temporal_layers < 1 || svc_config.num_temporal_layers > 8) { RTC_LOG(LS_WARNING) << "Av1 supports up to 8 temporal layers. " << svc_config.num_temporal_layers << " configured."; return false; } aom_svc_params_t& svc_params = svc_params_.emplace(); svc_params.number_spatial_layers = svc_config.num_spatial_layers; svc_params.number_temporal_layers = svc_config.num_temporal_layers; int num_layers = svc_config.num_spatial_layers * svc_config.num_temporal_layers; for (int i = 0; i < num_layers; ++i) { svc_params.min_quantizers[i] = kQpMin; svc_params.max_quantizers[i] = encoder_settings_.qpMax; } // Assume each temporal layer doubles framerate. for (int tid = 0; tid < svc_config.num_temporal_layers; ++tid) { svc_params.framerate_factor[tid] = 1 << (svc_config.num_temporal_layers - tid - 1); } for (int sid = 0; sid < svc_config.num_spatial_layers; ++sid) { svc_params.scaling_factor_num[sid] = svc_config.scaling_factor_num[sid]; svc_params.scaling_factor_den[sid] = svc_config.scaling_factor_den[sid]; } // svc_params.layer_target_bitrate is set in SetRates() before svc_params is // passed to SetEncoderControlParameters(AV1E_SET_SVC_PARAMS). return true; } void LibaomAv1Encoder::SetSvcLayerId( const ScalableVideoController::LayerFrameConfig& layer_frame) { aom_svc_layer_id_t layer_id = {}; layer_id.spatial_layer_id = layer_frame.SpatialId(); layer_id.temporal_layer_id = layer_frame.TemporalId(); SetEncoderControlParameters(AV1E_SET_SVC_LAYER_ID, &layer_id); } void LibaomAv1Encoder::SetSvcRefFrameConfig( const ScalableVideoController::LayerFrameConfig& layer_frame) { // Buffer name to use for each layer_frame.buffers position. In particular // when there are 2 buffers are referenced, prefer name them last and golden, // because av1 bitstream format has dedicated fields for these two names. // See last_frame_idx and golden_frame_idx in the av1 spec // https://aomediacodec.github.io/av1-spec/av1-spec.pdf static constexpr int kPreferedSlotName[] = {0, // Last 3, // Golden 1, 2, 4, 5, 6}; static constexpr int kAv1NumBuffers = 8; aom_svc_ref_frame_config_t ref_frame_config = {}; RTC_CHECK_LE(layer_frame.Buffers().size(), ABSL_ARRAYSIZE(kPreferedSlotName)); for (size_t i = 0; i < layer_frame.Buffers().size(); ++i) { const CodecBufferUsage& buffer = layer_frame.Buffers()[i]; int slot_name = kPreferedSlotName[i]; RTC_CHECK_GE(buffer.id, 0); RTC_CHECK_LT(buffer.id, kAv1NumBuffers); ref_frame_config.ref_idx[slot_name] = buffer.id; if (buffer.referenced) { ref_frame_config.reference[slot_name] = 1; } if (buffer.updated) { ref_frame_config.refresh[buffer.id] = 1; } } SetEncoderControlParameters(AV1E_SET_SVC_REF_FRAME_CONFIG, &ref_frame_config); } int32_t LibaomAv1Encoder::RegisterEncodeCompleteCallback( EncodedImageCallback* encoded_image_callback) { encoded_image_callback_ = encoded_image_callback; return WEBRTC_VIDEO_CODEC_OK; } int32_t LibaomAv1Encoder::Release() { if (frame_for_encode_ != nullptr) { aom_img_free(frame_for_encode_); frame_for_encode_ = nullptr; } if (inited_) { if (aom_codec_destroy(&ctx_)) { return WEBRTC_VIDEO_CODEC_MEMORY; } inited_ = false; } rates_configured_ = false; return WEBRTC_VIDEO_CODEC_OK; } void LibaomAv1Encoder::MaybeRewrapImgWithFormat(const aom_img_fmt_t fmt) { if (!frame_for_encode_) { frame_for_encode_ = aom_img_wrap(nullptr, fmt, cfg_.g_w, cfg_.g_h, 1, nullptr); } else if (frame_for_encode_->fmt != fmt) { RTC_LOG(LS_INFO) << "Switching AV1 encoder pixel format to " << (fmt == AOM_IMG_FMT_NV12 ? "NV12" : "I420"); aom_img_free(frame_for_encode_); frame_for_encode_ = aom_img_wrap(nullptr, fmt, cfg_.g_w, cfg_.g_h, 1, nullptr); } // else no-op since the image is already in the right format. } int32_t LibaomAv1Encoder::Encode( const VideoFrame& frame, const std::vector* frame_types) { if (!inited_ || encoded_image_callback_ == nullptr || !rates_configured_) { return WEBRTC_VIDEO_CODEC_UNINITIALIZED; } bool keyframe_required = frame_types != nullptr && absl::c_linear_search(*frame_types, VideoFrameType::kVideoFrameKey); std::vector layer_frames = svc_controller_->NextFrameConfig(keyframe_required); if (layer_frames.empty()) { RTC_LOG(LS_ERROR) << "SVCController returned no configuration for a frame."; return WEBRTC_VIDEO_CODEC_ERROR; } rtc::scoped_refptr buffer = frame.video_frame_buffer(); absl::InlinedVector supported_formats = {VideoFrameBuffer::Type::kI420, VideoFrameBuffer::Type::kNV12}; rtc::scoped_refptr mapped_buffer; if (buffer->type() != VideoFrameBuffer::Type::kNative) { // `buffer` is already mapped. mapped_buffer = buffer; } else { // Attempt to map to one of the supported formats. mapped_buffer = buffer->GetMappedFrameBuffer(supported_formats); } // Convert input frame to I420, if needed. if (!mapped_buffer || (absl::c_find(supported_formats, mapped_buffer->type()) == supported_formats.end() && mapped_buffer->type() != VideoFrameBuffer::Type::kI420A)) { rtc::scoped_refptr converted_buffer(buffer->ToI420()); if (!converted_buffer) { RTC_LOG(LS_ERROR) << "Failed to convert " << VideoFrameBufferTypeToString( frame.video_frame_buffer()->type()) << " image to I420. Can't encode frame."; return WEBRTC_VIDEO_CODEC_ENCODER_FAILURE; } RTC_CHECK(converted_buffer->type() == VideoFrameBuffer::Type::kI420 || converted_buffer->type() == VideoFrameBuffer::Type::kI420A); mapped_buffer = converted_buffer; } switch (mapped_buffer->type()) { case VideoFrameBuffer::Type::kI420: case VideoFrameBuffer::Type::kI420A: { // Set frame_for_encode_ data pointers and strides. MaybeRewrapImgWithFormat(AOM_IMG_FMT_I420); auto i420_buffer = mapped_buffer->GetI420(); RTC_DCHECK(i420_buffer); frame_for_encode_->planes[AOM_PLANE_Y] = const_cast(i420_buffer->DataY()); frame_for_encode_->planes[AOM_PLANE_U] = const_cast(i420_buffer->DataU()); frame_for_encode_->planes[AOM_PLANE_V] = const_cast(i420_buffer->DataV()); frame_for_encode_->stride[AOM_PLANE_Y] = i420_buffer->StrideY(); frame_for_encode_->stride[AOM_PLANE_U] = i420_buffer->StrideU(); frame_for_encode_->stride[AOM_PLANE_V] = i420_buffer->StrideV(); break; } case VideoFrameBuffer::Type::kNV12: { MaybeRewrapImgWithFormat(AOM_IMG_FMT_NV12); const NV12BufferInterface* nv12_buffer = mapped_buffer->GetNV12(); RTC_DCHECK(nv12_buffer); frame_for_encode_->planes[AOM_PLANE_Y] = const_cast(nv12_buffer->DataY()); frame_for_encode_->planes[AOM_PLANE_U] = const_cast(nv12_buffer->DataUV()); frame_for_encode_->planes[AOM_PLANE_V] = nullptr; frame_for_encode_->stride[AOM_PLANE_Y] = nv12_buffer->StrideY(); frame_for_encode_->stride[AOM_PLANE_U] = nv12_buffer->StrideUV(); frame_for_encode_->stride[AOM_PLANE_V] = 0; break; } default: return WEBRTC_VIDEO_CODEC_ENCODER_FAILURE; } const uint32_t duration = kRtpTicksPerSecond / static_cast(encoder_settings_.maxFramerate); timestamp_ += duration; const size_t num_spatial_layers = svc_params_ ? svc_params_->number_spatial_layers : 1; auto next_layer_frame = layer_frames.begin(); for (size_t i = 0; i < num_spatial_layers; ++i) { // The libaom AV1 encoder requires that `aom_codec_encode` is called for // every spatial layer, even if the configured bitrate for that layer is // zero. For zero bitrate spatial layers no frames will be produced. absl::optional non_encoded_layer_frame; ScalableVideoController::LayerFrameConfig* layer_frame; if (next_layer_frame != layer_frames.end() && next_layer_frame->SpatialId() == static_cast(i)) { layer_frame = &*next_layer_frame; ++next_layer_frame; } else { // For layers that are not encoded only the spatial id matters. non_encoded_layer_frame.emplace().S(i); layer_frame = &*non_encoded_layer_frame; } const bool end_of_picture = (next_layer_frame == layer_frames.end()); aom_enc_frame_flags_t flags = layer_frame->IsKeyframe() ? AOM_EFLAG_FORCE_KF : 0; if (SvcEnabled()) { SetSvcLayerId(*layer_frame); SetSvcRefFrameConfig(*layer_frame); } // Encode a frame. The presentation timestamp `pts` should not use real // timestamps from frames or the wall clock, as that can cause the rate // controller to misbehave. aom_codec_err_t ret = aom_codec_encode(&ctx_, frame_for_encode_, timestamp_, duration, flags); if (ret != AOM_CODEC_OK) { RTC_LOG(LS_WARNING) << "LibaomAv1Encoder::Encode returned " << ret << " on aom_codec_encode."; return WEBRTC_VIDEO_CODEC_ERROR; } if (non_encoded_layer_frame) { continue; } // Get encoded image data. EncodedImage encoded_image; aom_codec_iter_t iter = nullptr; int data_pkt_count = 0; while (const aom_codec_cx_pkt_t* pkt = aom_codec_get_cx_data(&ctx_, &iter)) { if (pkt->kind == AOM_CODEC_CX_FRAME_PKT && pkt->data.frame.sz > 0) { if (data_pkt_count > 0) { RTC_LOG(LS_WARNING) << "LibaomAv1Encoder::Encoder returned more than " "one data packet for an input video frame."; Release(); } encoded_image.SetEncodedData(EncodedImageBuffer::Create( /*data=*/static_cast(pkt->data.frame.buf), /*size=*/pkt->data.frame.sz)); if ((pkt->data.frame.flags & AOM_EFLAG_FORCE_KF) != 0) { layer_frame->Keyframe(); } encoded_image._frameType = layer_frame->IsKeyframe() ? VideoFrameType::kVideoFrameKey : VideoFrameType::kVideoFrameDelta; encoded_image.SetRtpTimestamp(frame.rtp_timestamp()); encoded_image.SetCaptureTimeIdentifier(frame.capture_time_identifier()); encoded_image.capture_time_ms_ = frame.render_time_ms(); encoded_image.rotation_ = frame.rotation(); encoded_image.content_type_ = VideoContentType::UNSPECIFIED; // If encoded image width/height info are added to aom_codec_cx_pkt_t, // use those values in lieu of the values in frame. if (svc_params_) { int n = svc_params_->scaling_factor_num[layer_frame->SpatialId()]; int d = svc_params_->scaling_factor_den[layer_frame->SpatialId()]; encoded_image._encodedWidth = cfg_.g_w * n / d; encoded_image._encodedHeight = cfg_.g_h * n / d; encoded_image.SetSpatialIndex(layer_frame->SpatialId()); encoded_image.SetTemporalIndex(layer_frame->TemporalId()); } else { encoded_image._encodedWidth = cfg_.g_w; encoded_image._encodedHeight = cfg_.g_h; } encoded_image.timing_.flags = VideoSendTiming::kInvalid; int qp = -1; SET_ENCODER_PARAM_OR_RETURN_ERROR(AOME_GET_LAST_QUANTIZER, &qp); encoded_image.qp_ = qp; encoded_image.SetColorSpace(frame.color_space()); ++data_pkt_count; } } // Deliver encoded image data. if (encoded_image.size() > 0) { CodecSpecificInfo codec_specific_info; codec_specific_info.codecType = kVideoCodecAV1; codec_specific_info.end_of_picture = end_of_picture; codec_specific_info.scalability_mode = scalability_mode_; bool is_keyframe = layer_frame->IsKeyframe(); codec_specific_info.generic_frame_info = svc_controller_->OnEncodeDone(*layer_frame); if (is_keyframe && codec_specific_info.generic_frame_info) { codec_specific_info.template_structure = svc_controller_->DependencyStructure(); auto& resolutions = codec_specific_info.template_structure->resolutions; if (SvcEnabled()) { resolutions.resize(svc_params_->number_spatial_layers); for (int sid = 0; sid < svc_params_->number_spatial_layers; ++sid) { int n = svc_params_->scaling_factor_num[sid]; int d = svc_params_->scaling_factor_den[sid]; resolutions[sid] = RenderResolution(cfg_.g_w * n / d, cfg_.g_h * n / d); } } else { resolutions = {RenderResolution(cfg_.g_w, cfg_.g_h)}; } } encoded_image_callback_->OnEncodedImage(encoded_image, &codec_specific_info); } } return WEBRTC_VIDEO_CODEC_OK; } void LibaomAv1Encoder::SetRates(const RateControlParameters& parameters) { if (!inited_) { RTC_LOG(LS_WARNING) << "SetRates() while encoder is not initialized"; return; } if (parameters.framerate_fps < kMinimumFrameRate) { RTC_LOG(LS_WARNING) << "Unsupported framerate (must be >= " << kMinimumFrameRate << " ): " << parameters.framerate_fps; return; } if (parameters.bitrate.get_sum_bps() == 0) { RTC_LOG(LS_WARNING) << "Attempt to set target bit rate to zero"; return; } // The bitrates caluclated internally in libaom when `AV1E_SET_SVC_PARAMS` is // called depends on the currently configured `rc_target_bitrate`. If the // total target bitrate is not updated first a division by zero could happen. svc_controller_->OnRatesUpdated(parameters.bitrate); cfg_.rc_target_bitrate = parameters.bitrate.get_sum_kbps(); aom_codec_err_t error_code = aom_codec_enc_config_set(&ctx_, &cfg_); if (error_code != AOM_CODEC_OK) { RTC_LOG(LS_WARNING) << "Error configuring encoder, error code: " << error_code; } if (SvcEnabled()) { for (int sid = 0; sid < svc_params_->number_spatial_layers; ++sid) { // libaom bitrate for spatial id S and temporal id T means bitrate // of frames with spatial_id=S and temporal_id<=T // while `parameters.bitrate` provdies bitrate of frames with // spatial_id=S and temporal_id=T int accumulated_bitrate_bps = 0; for (int tid = 0; tid < svc_params_->number_temporal_layers; ++tid) { int layer_index = sid * svc_params_->number_temporal_layers + tid; accumulated_bitrate_bps += parameters.bitrate.GetBitrate(sid, tid); // `svc_params_->layer_target_bitrate` expects bitrate in kbps. svc_params_->layer_target_bitrate[layer_index] = accumulated_bitrate_bps / 1000; } } SetEncoderControlParameters(AV1E_SET_SVC_PARAMS, &*svc_params_); } rates_configured_ = true; // Set frame rate to closest integer value. encoder_settings_.maxFramerate = static_cast(parameters.framerate_fps + 0.5); } VideoEncoder::EncoderInfo LibaomAv1Encoder::GetEncoderInfo() const { EncoderInfo info; info.supports_native_handle = false; info.implementation_name = "libaom"; info.has_trusted_rate_controller = true; info.is_hardware_accelerated = false; info.scaling_settings = (inited_ && !encoder_settings_.AV1().automatic_resize_on) ? VideoEncoder::ScalingSettings::kOff : VideoEncoder::ScalingSettings(kMinQindex, kMaxQindex); info.preferred_pixel_formats = {VideoFrameBuffer::Type::kI420, VideoFrameBuffer::Type::kNV12}; if (SvcEnabled()) { for (int sid = 0; sid < svc_params_->number_spatial_layers; ++sid) { info.fps_allocation[sid].resize(svc_params_->number_temporal_layers); for (int tid = 0; tid < svc_params_->number_temporal_layers; ++tid) { info.fps_allocation[sid][tid] = EncoderInfo::kMaxFramerateFraction / svc_params_->framerate_factor[tid]; } } } if (!encoder_info_override_.resolution_bitrate_limits().empty()) { info.resolution_bitrate_limits = encoder_info_override_.resolution_bitrate_limits(); } return info; } } // namespace absl::Nonnull> CreateLibaomAv1Encoder( const Environment& env, LibaomAv1EncoderSettings settings) { if (settings.max_pixel_count_to_cpu_speed.empty()) { return std::make_unique(absl::nullopt, env.field_trials()); } else { return std::make_unique(settings, env.field_trials()); } } std::unique_ptr CreateLibaomAv1Encoder() { return std::make_unique(absl::nullopt, FieldTrialBasedConfig()); } std::unique_ptr CreateLibaomAv1Encoder( const LibaomAv1EncoderAuxConfig& aux_config) { return std::make_unique(aux_config, FieldTrialBasedConfig()); } } // namespace webrtc