mirror of
https://github.com/mollyim/webrtc.git
synced 2025-05-16 15:20:42 +01:00

Some of the macros in format_macros.h follow the C standard and try to fill holes in it (on Windows). But this one has no direct equivalent in the standard and is just mimicking the naming convention. That's not nice. References: https://devblogs.microsoft.com/cppblog/c99-library-support-in-visual-studio-2013/ https://stackoverflow.com/a/2524673 Change-Id: I53f3faca2976a5b5d4b04a67ffb56ae0f4e930b2 Bug: webrtc:10852 Reviewed-on: https://webrtc-review.googlesource.com/c/src/+/147862 Commit-Queue: Oleh Prypin <oprypin@webrtc.org> Reviewed-by: Karl Wiberg <kwiberg@webrtc.org> Reviewed-by: Niels Moller <nisse@webrtc.org> Reviewed-by: Mirko Bonadei <mbonadei@webrtc.org> Cr-Commit-Position: refs/heads/master@{#28794}
163 lines
5.9 KiB
C++
163 lines
5.9 KiB
C++
/*
|
|
* Copyright (c) 2014 The WebRTC project authors. All Rights Reserved.
|
|
*
|
|
* Use of this source code is governed by a BSD-style license
|
|
* that can be found in the LICENSE file in the root of the source
|
|
* tree. An additional intellectual property rights grant can be found
|
|
* in the file PATENTS. All contributing project authors may
|
|
* be found in the AUTHORS file in the root of the source tree.
|
|
*/
|
|
|
|
#include "common_audio/audio_converter.h"
|
|
|
|
#include <algorithm>
|
|
#include <cmath>
|
|
#include <memory>
|
|
#include <vector>
|
|
|
|
#include "common_audio/channel_buffer.h"
|
|
#include "common_audio/resampler/push_sinc_resampler.h"
|
|
#include "rtc_base/arraysize.h"
|
|
#include "rtc_base/format_macros.h"
|
|
#include "test/gtest.h"
|
|
|
|
namespace webrtc {
|
|
|
|
typedef std::unique_ptr<ChannelBuffer<float>> ScopedBuffer;
|
|
|
|
// Sets the signal value to increase by |data| with every sample.
|
|
ScopedBuffer CreateBuffer(const std::vector<float>& data, size_t frames) {
|
|
const size_t num_channels = data.size();
|
|
ScopedBuffer sb(new ChannelBuffer<float>(frames, num_channels));
|
|
for (size_t i = 0; i < num_channels; ++i)
|
|
for (size_t j = 0; j < frames; ++j)
|
|
sb->channels()[i][j] = data[i] * j;
|
|
return sb;
|
|
}
|
|
|
|
void VerifyParams(const ChannelBuffer<float>& ref,
|
|
const ChannelBuffer<float>& test) {
|
|
EXPECT_EQ(ref.num_channels(), test.num_channels());
|
|
EXPECT_EQ(ref.num_frames(), test.num_frames());
|
|
}
|
|
|
|
// Computes the best SNR based on the error between |ref_frame| and
|
|
// |test_frame|. It searches around |expected_delay| in samples between the
|
|
// signals to compensate for the resampling delay.
|
|
float ComputeSNR(const ChannelBuffer<float>& ref,
|
|
const ChannelBuffer<float>& test,
|
|
size_t expected_delay) {
|
|
VerifyParams(ref, test);
|
|
float best_snr = 0;
|
|
size_t best_delay = 0;
|
|
|
|
// Search within one sample of the expected delay.
|
|
for (size_t delay = std::max(expected_delay, static_cast<size_t>(1)) - 1;
|
|
delay <= std::min(expected_delay + 1, ref.num_frames()); ++delay) {
|
|
float mse = 0;
|
|
float variance = 0;
|
|
float mean = 0;
|
|
for (size_t i = 0; i < ref.num_channels(); ++i) {
|
|
for (size_t j = 0; j < ref.num_frames() - delay; ++j) {
|
|
float error = ref.channels()[i][j] - test.channels()[i][j + delay];
|
|
mse += error * error;
|
|
variance += ref.channels()[i][j] * ref.channels()[i][j];
|
|
mean += ref.channels()[i][j];
|
|
}
|
|
}
|
|
|
|
const size_t length = ref.num_channels() * (ref.num_frames() - delay);
|
|
mse /= length;
|
|
variance /= length;
|
|
mean /= length;
|
|
variance -= mean * mean;
|
|
float snr = 100; // We assign 100 dB to the zero-error case.
|
|
if (mse > 0)
|
|
snr = 10 * std::log10(variance / mse);
|
|
if (snr > best_snr) {
|
|
best_snr = snr;
|
|
best_delay = delay;
|
|
}
|
|
}
|
|
printf("SNR=%.1f dB at delay=%" RTC_PRIuS "\n", best_snr, best_delay);
|
|
return best_snr;
|
|
}
|
|
|
|
// Sets the source to a linearly increasing signal for which we can easily
|
|
// generate a reference. Runs the AudioConverter and ensures the output has
|
|
// sufficiently high SNR relative to the reference.
|
|
void RunAudioConverterTest(size_t src_channels,
|
|
int src_sample_rate_hz,
|
|
size_t dst_channels,
|
|
int dst_sample_rate_hz) {
|
|
const float kSrcLeft = 0.0002f;
|
|
const float kSrcRight = 0.0001f;
|
|
const float resampling_factor =
|
|
(1.f * src_sample_rate_hz) / dst_sample_rate_hz;
|
|
const float dst_left = resampling_factor * kSrcLeft;
|
|
const float dst_right = resampling_factor * kSrcRight;
|
|
const float dst_mono = (dst_left + dst_right) / 2;
|
|
const size_t src_frames = static_cast<size_t>(src_sample_rate_hz / 100);
|
|
const size_t dst_frames = static_cast<size_t>(dst_sample_rate_hz / 100);
|
|
|
|
std::vector<float> src_data(1, kSrcLeft);
|
|
if (src_channels == 2)
|
|
src_data.push_back(kSrcRight);
|
|
ScopedBuffer src_buffer = CreateBuffer(src_data, src_frames);
|
|
|
|
std::vector<float> dst_data(1, 0);
|
|
std::vector<float> ref_data;
|
|
if (dst_channels == 1) {
|
|
if (src_channels == 1)
|
|
ref_data.push_back(dst_left);
|
|
else
|
|
ref_data.push_back(dst_mono);
|
|
} else {
|
|
dst_data.push_back(0);
|
|
ref_data.push_back(dst_left);
|
|
if (src_channels == 1)
|
|
ref_data.push_back(dst_left);
|
|
else
|
|
ref_data.push_back(dst_right);
|
|
}
|
|
ScopedBuffer dst_buffer = CreateBuffer(dst_data, dst_frames);
|
|
ScopedBuffer ref_buffer = CreateBuffer(ref_data, dst_frames);
|
|
|
|
// The sinc resampler has a known delay, which we compute here.
|
|
const size_t delay_frames =
|
|
src_sample_rate_hz == dst_sample_rate_hz
|
|
? 0
|
|
: static_cast<size_t>(
|
|
PushSincResampler::AlgorithmicDelaySeconds(src_sample_rate_hz) *
|
|
dst_sample_rate_hz);
|
|
// SNR reported on the same line later.
|
|
printf("(%" RTC_PRIuS ", %d Hz) -> (%" RTC_PRIuS ", %d Hz) ", src_channels,
|
|
src_sample_rate_hz, dst_channels, dst_sample_rate_hz);
|
|
|
|
std::unique_ptr<AudioConverter> converter = AudioConverter::Create(
|
|
src_channels, src_frames, dst_channels, dst_frames);
|
|
converter->Convert(src_buffer->channels(), src_buffer->size(),
|
|
dst_buffer->channels(), dst_buffer->size());
|
|
|
|
EXPECT_LT(43.f,
|
|
ComputeSNR(*ref_buffer.get(), *dst_buffer.get(), delay_frames));
|
|
}
|
|
|
|
TEST(AudioConverterTest, ConversionsPassSNRThreshold) {
|
|
const int kSampleRates[] = {8000, 16000, 32000, 44100, 48000};
|
|
const size_t kChannels[] = {1, 2};
|
|
for (size_t src_rate = 0; src_rate < arraysize(kSampleRates); ++src_rate) {
|
|
for (size_t dst_rate = 0; dst_rate < arraysize(kSampleRates); ++dst_rate) {
|
|
for (size_t src_channel = 0; src_channel < arraysize(kChannels);
|
|
++src_channel) {
|
|
for (size_t dst_channel = 0; dst_channel < arraysize(kChannels);
|
|
++dst_channel) {
|
|
RunAudioConverterTest(kChannels[src_channel], kSampleRates[src_rate],
|
|
kChannels[dst_channel], kSampleRates[dst_rate]);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
} // namespace webrtc
|