mirror of
https://github.com/mollyim/webrtc.git
synced 2025-05-13 13:50:40 +01:00

Bug: webrtc:342905193 No-Try: True Change-Id: Icc968be43b8830038ea9a1f5f604307220457807 Reviewed-on: https://webrtc-review.googlesource.com/c/src/+/361021 Auto-Submit: Florent Castelli <orphis@webrtc.org> Reviewed-by: Harald Alvestrand <hta@webrtc.org> Commit-Queue: Florent Castelli <orphis@webrtc.org> Cr-Commit-Position: refs/heads/main@{#42911}
103 lines
4.6 KiB
C++
103 lines
4.6 KiB
C++
/*
|
|
* Copyright (c) 2013 The WebRTC project authors. All Rights Reserved.
|
|
*
|
|
* Use of this source code is governed by a BSD-style license
|
|
* that can be found in the LICENSE file in the root of the source
|
|
* tree. An additional intellectual property rights grant can be found
|
|
* in the file PATENTS. All contributing project authors may
|
|
* be found in the AUTHORS file in the root of the source tree.
|
|
*/
|
|
|
|
#ifndef MODULES_AUDIO_DEVICE_FINE_AUDIO_BUFFER_H_
|
|
#define MODULES_AUDIO_DEVICE_FINE_AUDIO_BUFFER_H_
|
|
|
|
#include <cstddef>
|
|
#include <cstdint>
|
|
#include <optional>
|
|
|
|
#include "api/array_view.h"
|
|
#include "rtc_base/buffer.h"
|
|
|
|
namespace webrtc {
|
|
|
|
class AudioDeviceBuffer;
|
|
|
|
// FineAudioBuffer takes an AudioDeviceBuffer (ADB) which deals with 16-bit PCM
|
|
// audio samples corresponding to 10ms of data. It then allows for this data
|
|
// to be pulled in a finer or coarser granularity. I.e. interacting with this
|
|
// class instead of directly with the AudioDeviceBuffer one can ask for any
|
|
// number of audio data samples. This class also ensures that audio data can be
|
|
// delivered to the ADB in 10ms chunks when the size of the provided audio
|
|
// buffers differs from 10ms.
|
|
// As an example: calling DeliverRecordedData() with 5ms buffers will deliver
|
|
// accumulated 10ms worth of data to the ADB every second call.
|
|
class FineAudioBuffer {
|
|
public:
|
|
// `device_buffer` is a buffer that provides 10ms of audio data.
|
|
FineAudioBuffer(AudioDeviceBuffer* audio_device_buffer);
|
|
~FineAudioBuffer();
|
|
|
|
// Clears buffers and counters dealing with playout and/or recording.
|
|
void ResetPlayout();
|
|
void ResetRecord();
|
|
|
|
// Utility methods which returns true if valid parameters are acquired at
|
|
// constructions.
|
|
bool IsReadyForPlayout() const;
|
|
bool IsReadyForRecord() const;
|
|
|
|
// Copies audio samples into `audio_buffer` where number of requested
|
|
// elements is specified by `audio_buffer.size()`. The producer will always
|
|
// fill up the audio buffer and if no audio exists, the buffer will contain
|
|
// silence instead. The provided delay estimate in `playout_delay_ms` should
|
|
// contain an estimate of the latency between when an audio frame is read from
|
|
// WebRTC and when it is played out on the speaker.
|
|
void GetPlayoutData(rtc::ArrayView<int16_t> audio_buffer,
|
|
int playout_delay_ms);
|
|
|
|
// Consumes the audio data in `audio_buffer` and sends it to the WebRTC layer
|
|
// in chunks of 10ms. The sum of the provided delay estimate in
|
|
// `record_delay_ms` and the latest `playout_delay_ms` in GetPlayoutData()
|
|
// are given to the AEC in the audio processing module.
|
|
// They can be fixed values on most platforms and they are ignored if an
|
|
// external (hardware/built-in) AEC is used.
|
|
// Example: buffer size is 5ms => call #1 stores 5ms of data, call #2 stores
|
|
// 5ms of data and sends a total of 10ms to WebRTC and clears the internal
|
|
// cache. Call #3 restarts the scheme above.
|
|
void DeliverRecordedData(rtc::ArrayView<const int16_t> audio_buffer,
|
|
int record_delay_ms) {
|
|
DeliverRecordedData(audio_buffer, record_delay_ms, std::nullopt);
|
|
}
|
|
void DeliverRecordedData(rtc::ArrayView<const int16_t> audio_buffer,
|
|
int record_delay_ms,
|
|
std::optional<int64_t> capture_time_ns);
|
|
|
|
private:
|
|
// Device buffer that works with 10ms chunks of data both for playout and
|
|
// for recording. I.e., the WebRTC side will always be asked for audio to be
|
|
// played out in 10ms chunks and recorded audio will be sent to WebRTC in
|
|
// 10ms chunks as well. This raw pointer is owned by the constructor of this
|
|
// class and the owner must ensure that the pointer is valid during the life-
|
|
// time of this object.
|
|
AudioDeviceBuffer* const audio_device_buffer_;
|
|
// Number of audio samples per channel per 10ms. Set once at construction
|
|
// based on parameters in `audio_device_buffer`.
|
|
const size_t playout_samples_per_channel_10ms_;
|
|
const size_t record_samples_per_channel_10ms_;
|
|
// Number of audio channels. Set once at construction based on parameters in
|
|
// `audio_device_buffer`.
|
|
const size_t playout_channels_;
|
|
const size_t record_channels_;
|
|
// Storage for output samples from which a consumer can read audio buffers
|
|
// in any size using GetPlayoutData().
|
|
rtc::BufferT<int16_t> playout_buffer_;
|
|
// Storage for input samples that are about to be delivered to the WebRTC
|
|
// ADB or remains from the last successful delivery of a 10ms audio buffer.
|
|
rtc::BufferT<int16_t> record_buffer_;
|
|
// Contains latest delay estimate given to GetPlayoutData().
|
|
int playout_delay_ms_ = 0;
|
|
};
|
|
|
|
} // namespace webrtc
|
|
|
|
#endif // MODULES_AUDIO_DEVICE_FINE_AUDIO_BUFFER_H_
|