webrtc/modules/audio_processing/audio_frame_view_unittest.cc
Tommi 6056976709 Updates to AudioFrameView and VectorFloatFrame
Using DeinterleavedView<> simplifies these two classes, so now the
classes are arguably thin wrappers on top of DeinterleavedView<> and
AudioFrameView<> can be replaced with DeinterleavedView<>.

The changes are:
* Make VectorFloatFrame not use a vector of vectors but rather
  just hold a one dimensional vector of samples and leaves the mapping
  into the buffer up to DeinterleavedView<>.
* Remove the `channel_ptrs_` vector which was required due to an
  issue with AudioFrameView.
* AudioFrameView is now a wrapper over DeinterleavedView<>. The most
  important change is to remove the `audio_samples_` pointer, which
  pointed into an externally owned pointer array (in addition to
  the array that holds the samples themselves). Now AudioFrameView
  can be initialized without requiring such a long-lived array.

Bug: chromium:335805780
Change-Id: I8f3c23c0ac4b5a337f68e9161fc3a97271f4e87d
Reviewed-on: https://webrtc-review.googlesource.com/c/src/+/352504
Commit-Queue: Tomas Gunnarsson <tommi@webrtc.org>
Reviewed-by: Per Åhgren <peah@webrtc.org>
Cr-Commit-Position: refs/heads/main@{#42498}
2024-06-17 12:13:40 +00:00

91 lines
3.4 KiB
C++

/*
* Copyright 2018 The WebRTC Project Authors. All rights reserved.
*
* Use of this source code is governed by a BSD-style license
* that can be found in the LICENSE file in the root of the source
* tree. An additional intellectual property rights grant can be found
* in the file PATENTS. All contributing project authors may
* be found in the AUTHORS file in the root of the source tree.
*/
#include "modules/audio_processing/include/audio_frame_view.h"
#include <array>
#include "common_audio/channel_buffer.h"
#include "modules/audio_processing/audio_buffer.h"
#include "rtc_base/arraysize.h"
#include "test/gtest.h"
namespace webrtc {
TEST(AudioFrameTest, ConstructFromAudioBuffer) {
constexpr int kSampleRateHz = 48000;
constexpr int kNumChannels = 2;
constexpr float kFloatConstant = 1272.f;
constexpr float kIntConstant = 17252;
const StreamConfig stream_config(kSampleRateHz, kNumChannels);
AudioBuffer buffer(
stream_config.sample_rate_hz(), stream_config.num_channels(),
stream_config.sample_rate_hz(), stream_config.num_channels(),
stream_config.sample_rate_hz(), stream_config.num_channels());
AudioFrameView<float> non_const_view(buffer.channels(), buffer.num_channels(),
buffer.num_frames());
// Modification is allowed.
non_const_view.channel(0)[0] = kFloatConstant;
EXPECT_EQ(buffer.channels()[0][0], kFloatConstant);
AudioFrameView<const float> const_view(
buffer.channels(), buffer.num_channels(), buffer.num_frames());
// Modification is not allowed.
// const_view.channel(0)[0] = kFloatConstant;
// Assignment is allowed.
AudioFrameView<const float> other_const_view = non_const_view;
static_cast<void>(other_const_view);
// But not the other way. The following will fail:
// non_const_view = other_const_view;
AudioFrameView<float> non_const_float_view(
buffer.channels(), buffer.num_channels(), buffer.num_frames());
non_const_float_view.channel(0)[0] = kIntConstant;
EXPECT_EQ(buffer.channels()[0][0], kIntConstant);
}
TEST(AudioFrameTest, ConstructFromChannelBuffer) {
ChannelBuffer<float> buffer(480, 2);
AudioFrameView<float> view(buffer.channels(), buffer.num_channels(),
buffer.num_frames());
EXPECT_EQ(view.num_channels(), 2);
EXPECT_EQ(view.samples_per_channel(), 480);
}
TEST(AudioFrameTest, ToDeinterleavedView) {
ChannelBuffer<float> buffer(480, 2);
AudioFrameView<float> view(buffer.channels(), buffer.num_channels(),
buffer.num_frames());
DeinterleavedView<float> non_const_view = view.view();
DeinterleavedView<const float> const_view =
static_cast<const AudioFrameView<float>&>(view).view();
ASSERT_EQ(non_const_view.num_channels(), 2u);
ASSERT_EQ(const_view.num_channels(), 2u);
for (size_t i = 0; i < non_const_view.num_channels(); ++i) {
EXPECT_EQ(non_const_view[i].data(), const_view[i].data());
EXPECT_EQ(non_const_view[i].data(), view.channel(i).data());
}
}
TEST(AudioFrameTest, FromDeinterleavedView) {
std::array<float, 480 * 2> buffer;
DeinterleavedView<float> view(buffer.data(), buffer.size() / 2u, 2u);
AudioFrameView<float> frame_view(view);
EXPECT_EQ(static_cast<size_t>(frame_view.num_channels()),
view.num_channels());
EXPECT_EQ(frame_view[0], view[0]);
EXPECT_EQ(frame_view[1], view[1]);
}
} // namespace webrtc