| // Copyright 2021 The Chromium Authors |
| // Use of this source code is governed by a BSD-style license that can be |
| // found in the LICENSE file. |
| |
| #include "media/webrtc/audio_processor.h" |
| |
| #include <stddef.h> |
| #include <stdint.h> |
| |
| #include <optional> |
| #include <string> |
| |
| #include "base/files/file_path.h" |
| #include "base/files/file_util.h" |
| #include "base/logging.h" |
| #include "base/memory/aligned_memory.h" |
| #include "base/path_service.h" |
| #include "base/strings/stringprintf.h" |
| #include "base/test/mock_callback.h" |
| #include "base/test/scoped_feature_list.h" |
| #include "base/test/task_environment.h" |
| #include "base/test/test_file_util.h" |
| #include "base/time/time.h" |
| #include "build/build_config.h" |
| #include "build/chromecast_buildflags.h" |
| #include "media/base/audio_bus.h" |
| #include "media/base/audio_parameters.h" |
| #include "media/base/audio_processing.h" |
| #include "media/webrtc/constants.h" |
| #include "media/webrtc/webrtc_features.h" |
| #include "testing/gtest/include/gtest/gtest.h" |
| #include "third_party/webrtc/api/make_ref_counted.h" |
| #include "third_party/webrtc/modules/audio_processing/include/mock_audio_processing.h" |
| |
| using ::testing::_; |
| using ::testing::AnyNumber; |
| using ::testing::AtLeast; |
| using ::testing::Return; |
| |
| namespace media { |
| namespace { |
| |
| // Test all sample rates observed in UMA metric WebRTC.AudioInputSampleRate. |
| static const int kSupportedSampleRates[] = {8000, 11025, 16000, 22050, 24000, |
| 32000, 44100, 48000, 96000, 192000}; |
| |
| using MockProcessedCaptureCallback = |
| base::MockRepeatingCallback<void(const media::AudioBus& audio_bus, |
| base::TimeTicks audio_capture_time, |
| std::optional<double> new_volume)>; |
| |
| AudioProcessor::LogCallback LogCallbackForTesting() { |
| return base::BindRepeating( |
| [](base::StringPiece message) { VLOG(1) << (message); }); |
| } |
| |
| // The number of packets used for testing. |
| const int kNumberOfPacketsForTest = 100; |
| |
| void ReadDataFromSpeechFile(char* data, int length) { |
| base::FilePath file; |
| CHECK(base::PathService::Get(base::DIR_SRC_TEST_DATA_ROOT, &file)); |
| file = file.Append(FILE_PATH_LITERAL("media")) |
| .Append(FILE_PATH_LITERAL("test")) |
| .Append(FILE_PATH_LITERAL("data")) |
| .Append(FILE_PATH_LITERAL("speech_16b_stereo_48kHz.raw")); |
| DCHECK(base::PathExists(file)); |
| int64_t data_file_size64 = 0; |
| DCHECK(base::GetFileSize(file, &data_file_size64)); |
| EXPECT_EQ(length, base::ReadFile(file, data, length)); |
| DCHECK(data_file_size64 > length); |
| } |
| |
| void DisableDefaultSettings(AudioProcessingSettings& settings) { |
| settings.echo_cancellation = false; |
| settings.noise_suppression = false; |
| settings.transient_noise_suppression = false; |
| settings.automatic_gain_control = false; |
| settings.high_pass_filter = false; |
| settings.multi_channel_capture_processing = false; |
| settings.stereo_mirroring = false; |
| settings.force_apm_creation = false; |
| } |
| |
| } // namespace |
| |
| class AudioProcessorTest : public ::testing::Test { |
| public: |
| AudioProcessorTest() |
| : params_(media::AudioParameters::AUDIO_PCM_LOW_LATENCY, |
| ChannelLayoutConfig::Stereo(), |
| 48000, |
| 480) {} |
| |
| protected: |
| // Helper method to save duplicated code. |
| static void ProcessDataAndVerifyFormat( |
| AudioProcessor& audio_processor, |
| MockProcessedCaptureCallback& mock_capture_callback) { |
| // Read the audio data from a file. |
| const media::AudioParameters& input_params = audio_processor.input_format(); |
| const media::AudioParameters& output_params = |
| audio_processor.output_format(); |
| const int packet_size = |
| input_params.frames_per_buffer() * 2 * input_params.channels(); |
| const size_t length = packet_size * kNumberOfPacketsForTest; |
| auto capture_data = std::make_unique<char[]>(length); |
| ReadDataFromSpeechFile(capture_data.get(), static_cast<int>(length)); |
| const int16_t* data_ptr = |
| reinterpret_cast<const int16_t*>(capture_data.get()); |
| std::unique_ptr<media::AudioBus> data_bus = media::AudioBus::Create( |
| input_params.channels(), input_params.frames_per_buffer()); |
| |
| const base::TimeTicks input_capture_time = base::TimeTicks::Now(); |
| int num_preferred_channels = -1; |
| for (int i = 0; i < kNumberOfPacketsForTest; ++i) { |
| data_bus->FromInterleaved<media::SignedInt16SampleTypeTraits>( |
| data_ptr, data_bus->frames()); |
| |
| // 1. Provide playout audio, if echo cancellation is enabled. |
| const bool is_aec_enabled = |
| audio_processor.has_webrtc_audio_processing() && |
| (*audio_processor.GetAudioProcessingModuleConfigForTesting()) |
| .echo_canceller.enabled; |
| if (is_aec_enabled) { |
| audio_processor.OnPlayoutData(*data_bus, input_params.sample_rate(), |
| base::Milliseconds(10)); |
| } |
| |
| // 2. Set up expectations and process captured audio. |
| EXPECT_CALL(mock_capture_callback, Run(_, _, _)) |
| .WillRepeatedly([&](const media::AudioBus& processed_audio, |
| base::TimeTicks audio_capture_time, |
| std::optional<double> new_volume) { |
| EXPECT_EQ(processed_audio.channels(), output_params.channels()); |
| EXPECT_EQ(processed_audio.frames(), |
| output_params.frames_per_buffer()); |
| EXPECT_EQ(audio_capture_time, input_capture_time); |
| }); |
| audio_processor.ProcessCapturedAudio(*data_bus, input_capture_time, |
| num_preferred_channels, 1.0, false); |
| |
| data_ptr += input_params.frames_per_buffer() * input_params.channels(); |
| |
| // Test different values of num_preferred_channels. |
| if (++num_preferred_channels > 5) { |
| num_preferred_channels = 0; |
| } |
| } |
| } |
| |
| // TODO(bugs.webrtc.org/7494): Remove/reduce duplication with |
| // `CreateWebRtcAudioProcessingModuleTest.CheckDefaultAudioProcessingConfig`. |
| void VerifyDefaultComponents(AudioProcessor& audio_processor) { |
| ASSERT_TRUE(audio_processor.has_webrtc_audio_processing()); |
| const webrtc::AudioProcessing::Config config = |
| *audio_processor.GetAudioProcessingModuleConfigForTesting(); |
| |
| EXPECT_TRUE(config.high_pass_filter.enabled); |
| EXPECT_FALSE(config.pre_amplifier.enabled); |
| EXPECT_TRUE(config.echo_canceller.enabled); |
| |
| #if BUILDFLAG(IS_WIN) || BUILDFLAG(IS_MAC) || BUILDFLAG(IS_LINUX) |
| EXPECT_FALSE(config.gain_controller1.enabled); |
| EXPECT_TRUE(config.gain_controller2.enabled); |
| #elif BUILDFLAG(IS_CHROMEOS) || BUILDFLAG(IS_FUCHSIA) |
| EXPECT_FALSE(config.gain_controller1.enabled); |
| EXPECT_TRUE(config.gain_controller2.enabled); |
| #elif BUILDFLAG(IS_CASTOS) || BUILDFLAG(IS_CAST_ANDROID) |
| EXPECT_TRUE(config.gain_controller1.enabled); |
| EXPECT_FALSE(config.gain_controller2.enabled); |
| #elif BUILDFLAG(IS_ANDROID) || BUILDFLAG(IS_IOS) |
| EXPECT_FALSE(config.gain_controller1.enabled); |
| EXPECT_TRUE(config.gain_controller2.enabled); |
| #else |
| GTEST_FAIL() << "Undefined expectation."; |
| #endif |
| |
| EXPECT_TRUE(config.noise_suppression.enabled); |
| EXPECT_EQ(config.noise_suppression.level, |
| webrtc::AudioProcessing::Config::NoiseSuppression::kHigh); |
| |
| #if BUILDFLAG(IS_WIN) || BUILDFLAG(IS_MAC) || BUILDFLAG(IS_LINUX) || \ |
| BUILDFLAG(IS_CHROMEOS) || BUILDFLAG(IS_FUCHSIA) |
| EXPECT_FALSE(config.echo_canceller.mobile_mode); |
| EXPECT_FALSE(config.transient_suppression.enabled); |
| #elif BUILDFLAG(IS_ANDROID) || BUILDFLAG(IS_IOS) |
| // Android and iOS use echo cancellation optimized for mobiles, and does not |
| // support keytap suppression. |
| EXPECT_TRUE(config.echo_canceller.mobile_mode); |
| EXPECT_FALSE(config.transient_suppression.enabled); |
| #else |
| EXPECT_FALSE(config.echo_canceller.mobile_mode); |
| EXPECT_TRUE(config.transient_suppression.enabled); |
| #endif |
| } |
| |
| media::AudioParameters params_; |
| MockProcessedCaptureCallback mock_capture_callback_; |
| // Necessary for working with WebRTC task queues. |
| base::test::TaskEnvironment task_environment_; |
| }; |
| |
| struct AudioProcessorTestMultichannelAndFormat |
| : public AudioProcessorTest, |
| public ::testing::WithParamInterface<std::tuple<bool, bool>> { |
| AudioParameters GetProcessorOutputParams( |
| const AudioParameters& params, |
| const AudioProcessingSettings& settings) { |
| const bool use_input_format_for_output = std::get<1>(GetParam()); |
| return use_input_format_for_output |
| ? params |
| : AudioProcessor::GetDefaultOutputFormat(params, settings); |
| } |
| |
| static std::string PrintTestName( |
| const testing::TestParamInfo<ParamType>& info) { |
| auto [multichannel, input_format_for_output] = info.param; |
| return base::StringPrintf("MultichannelApm%sSameInputOutputFormat%s", |
| multichannel ? "True" : "False", |
| input_format_for_output ? "True" : "False"); |
| } |
| }; |
| |
| INSTANTIATE_TEST_SUITE_P( |
| /*no prefix*/, |
| AudioProcessorTestMultichannelAndFormat, |
| ::testing::Combine(::testing::Bool(), ::testing::Bool()), |
| &AudioProcessorTestMultichannelAndFormat::PrintTestName); |
| |
| // Test crashing with ASAN on Android. crbug.com/468762 |
| #if BUILDFLAG(IS_ANDROID) && defined(ADDRESS_SANITIZER) |
| #define MAYBE_WithAudioProcessing DISABLED_WithAudioProcessing |
| #else |
| #define MAYBE_WithAudioProcessing WithAudioProcessing |
| #endif |
| TEST_P(AudioProcessorTestMultichannelAndFormat, MAYBE_WithAudioProcessing) { |
| AudioProcessingSettings settings{.multi_channel_capture_processing = |
| std::get<0>(GetParam())}; |
| std::unique_ptr<AudioProcessor> audio_processor = AudioProcessor::Create( |
| mock_capture_callback_.Get(), LogCallbackForTesting(), settings, params_, |
| GetProcessorOutputParams(params_, settings)); |
| EXPECT_TRUE(audio_processor->has_webrtc_audio_processing()); |
| VerifyDefaultComponents(*audio_processor); |
| |
| ProcessDataAndVerifyFormat(*audio_processor, mock_capture_callback_); |
| } |
| |
| TEST_F(AudioProcessorTest, TurnOffDefaultConstraints) { |
| AudioProcessingSettings settings; |
| // Turn off the default settings and pass it to AudioProcessor. |
| DisableDefaultSettings(settings); |
| std::unique_ptr<AudioProcessor> audio_processor = AudioProcessor::Create( |
| mock_capture_callback_.Get(), LogCallbackForTesting(), settings, params_, |
| AudioProcessor::GetDefaultOutputFormat(params_, settings)); |
| EXPECT_FALSE(audio_processor->has_webrtc_audio_processing()); |
| |
| EXPECT_EQ(audio_processor->output_format().sample_rate(), |
| params_.sample_rate()); |
| EXPECT_EQ(audio_processor->output_format().channels(), params_.channels()); |
| EXPECT_EQ(audio_processor->output_format().frames_per_buffer(), |
| params_.sample_rate() / 100); |
| |
| ProcessDataAndVerifyFormat(*audio_processor, mock_capture_callback_); |
| } |
| |
| // Test crashing with ASAN on Android. crbug.com/468762 |
| #if BUILDFLAG(IS_ANDROID) && defined(ADDRESS_SANITIZER) |
| #define MAYBE_TestAllSampleRates DISABLED_TestAllSampleRates |
| #else |
| #define MAYBE_TestAllSampleRates TestAllSampleRates |
| #endif |
| TEST_P(AudioProcessorTestMultichannelAndFormat, MAYBE_TestAllSampleRates) { |
| AudioProcessingSettings settings{.multi_channel_capture_processing = |
| std::get<0>(GetParam())}; |
| |
| for (int sample_rate : kSupportedSampleRates) { |
| SCOPED_TRACE(testing::Message() << "sample_rate=" << sample_rate); |
| int buffer_size = sample_rate / 100; |
| media::AudioParameters params(media::AudioParameters::AUDIO_PCM_LOW_LATENCY, |
| ChannelLayoutConfig::Stereo(), sample_rate, |
| buffer_size); |
| std::unique_ptr<AudioProcessor> audio_processor = AudioProcessor::Create( |
| mock_capture_callback_.Get(), LogCallbackForTesting(), settings, params, |
| GetProcessorOutputParams(params, settings)); |
| EXPECT_TRUE(audio_processor->has_webrtc_audio_processing()); |
| VerifyDefaultComponents(*audio_processor); |
| |
| ProcessDataAndVerifyFormat(*audio_processor, mock_capture_callback_); |
| } |
| } |
| |
| TEST_F(AudioProcessorTest, StartStopAecDump) { |
| auto dir = base::CreateUniqueTempDirectoryScopedToTest(); |
| base::FilePath temp_file_path; |
| ASSERT_TRUE(base::CreateTemporaryFileInDir(dir, &temp_file_path)); |
| { |
| AudioProcessingSettings settings; |
| std::unique_ptr<AudioProcessor> audio_processor = AudioProcessor::Create( |
| mock_capture_callback_.Get(), LogCallbackForTesting(), settings, |
| params_, AudioProcessor::GetDefaultOutputFormat(params_, settings)); |
| |
| // Start and stop recording. |
| audio_processor->OnStartDump(base::File( |
| temp_file_path, base::File::FLAG_WRITE | base::File::FLAG_OPEN)); |
| audio_processor->OnStopDump(); |
| |
| // Start and stop a second recording. |
| audio_processor->OnStartDump(base::File( |
| temp_file_path, base::File::FLAG_WRITE | base::File::FLAG_OPEN)); |
| audio_processor->OnStopDump(); |
| } |
| |
| // Check that dump file is non-empty after audio processor has been |
| // destroyed. Note that this test fails when compiling WebRTC |
| // without protobuf support, rtc_enable_protobuf=false. |
| std::string output; |
| ASSERT_TRUE(base::ReadFileToString(temp_file_path, &output)); |
| EXPECT_FALSE(output.empty()); |
| // The temporary file is deleted when temp_directory exits scope. |
| } |
| |
| TEST_F(AudioProcessorTest, StartAecDumpDuringOngoingAecDump) { |
| auto temp_dir = base::CreateUniqueTempDirectoryScopedToTest(); |
| base::FilePath temp_file_path_a; |
| ASSERT_TRUE(base::CreateTemporaryFileInDir(temp_dir, &temp_file_path_a)); |
| |
| base::FilePath temp_file_path_b; |
| ASSERT_TRUE(base::CreateTemporaryFileInDir(temp_dir, &temp_file_path_b)); |
| { |
| AudioProcessingSettings settings; |
| std::unique_ptr<AudioProcessor> audio_processor = AudioProcessor::Create( |
| mock_capture_callback_.Get(), LogCallbackForTesting(), settings, |
| params_, AudioProcessor::GetDefaultOutputFormat(params_, settings)); |
| |
| // Start a recording. |
| audio_processor->OnStartDump(base::File( |
| temp_file_path_a, base::File::FLAG_WRITE | base::File::FLAG_OPEN)); |
| |
| // Start another recording without stopping the previous one. |
| audio_processor->OnStartDump(base::File( |
| temp_file_path_b, base::File::FLAG_WRITE | base::File::FLAG_OPEN)); |
| audio_processor->OnStopDump(); |
| } |
| |
| // Check that dump files are non-empty after audio processor has been |
| // destroyed. Note that this test fails when compiling WebRTC |
| // without protobuf support, rtc_enable_protobuf=false. |
| std::string output; |
| ASSERT_TRUE(base::ReadFileToString(temp_file_path_a, &output)); |
| EXPECT_FALSE(output.empty()); |
| ASSERT_TRUE(base::ReadFileToString(temp_file_path_b, &output)); |
| EXPECT_FALSE(output.empty()); |
| // The temporary files are deleted when temp_directory exits scope. |
| } |
| |
| TEST_P(AudioProcessorTestMultichannelAndFormat, TestStereoAudio) { |
| const bool use_multichannel_processing = std::get<0>(GetParam()); |
| |
| // Construct a stereo audio bus and fill the left channel with content. |
| std::unique_ptr<media::AudioBus> data_bus = |
| media::AudioBus::Create(params_.channels(), params_.frames_per_buffer()); |
| data_bus->Zero(); |
| for (int i = 0; i < data_bus->frames(); ++i) { |
| data_bus->channel(0)[i] = (i % 11) * 0.1f - 0.5f; |
| } |
| |
| // Test without and with audio processing enabled. |
| constexpr bool kUseApmValues[] = |
| #if BUILDFLAG(IS_IOS) |
| // TODO(https://crbug.com/1417474): `false` fails on ios-blink platform |
| // due to a special case for iOS in settings.NeedWebrtcAudioProcessing() |
| {true}; |
| #else |
| {false, true}; |
| #endif |
| for (bool use_apm : kUseApmValues) { |
| // No need to test stereo with APM if disabled. |
| if (use_apm && !use_multichannel_processing) { |
| continue; |
| } |
| SCOPED_TRACE(testing::Message() << "use_apm=" << use_apm); |
| |
| AudioProcessingSettings settings{.multi_channel_capture_processing = |
| use_multichannel_processing}; |
| if (!use_apm) { |
| // Turn off the audio processing. |
| DisableDefaultSettings(settings); |
| } |
| // Turn on the stereo channels mirroring. |
| settings.stereo_mirroring = true; |
| std::unique_ptr<AudioProcessor> audio_processor = AudioProcessor::Create( |
| mock_capture_callback_.Get(), LogCallbackForTesting(), settings, |
| params_, GetProcessorOutputParams(params_, settings)); |
| EXPECT_EQ(audio_processor->has_webrtc_audio_processing(), use_apm); |
| // There's no sense in continuing if this fails. |
| ASSERT_EQ(2, audio_processor->output_format().channels()); |
| |
| // Run the test consecutively to make sure the stereo channels are not |
| // flipped back and forth. |
| const base::TimeTicks pushed_capture_time = base::TimeTicks::Now(); |
| |
| for (int num_preferred_channels = 0; num_preferred_channels <= 5; |
| ++num_preferred_channels) { |
| SCOPED_TRACE(testing::Message() |
| << "num_preferred_channels=" << num_preferred_channels); |
| for (int i = 0; i < kNumberOfPacketsForTest; ++i) { |
| SCOPED_TRACE(testing::Message() << "packet index i=" << i); |
| EXPECT_CALL(mock_capture_callback_, Run(_, _, _)).Times(1); |
| // Pass audio for processing. |
| audio_processor->ProcessCapturedAudio( |
| *data_bus, pushed_capture_time, num_preferred_channels, 0.0, false); |
| } |
| // At this point, the audio processing algorithms have gotten past any |
| // initial buffer silence generated from resamplers, FFTs, and whatnot. |
| // Set up expectations via the mock callback: |
| EXPECT_CALL(mock_capture_callback_, Run(_, _, _)) |
| .WillRepeatedly([&](const media::AudioBus& processed_audio, |
| base::TimeTicks audio_capture_time, |
| std::optional<double> new_volume) { |
| EXPECT_EQ(audio_capture_time, pushed_capture_time); |
| if (!use_apm) { |
| EXPECT_FALSE(new_volume.has_value()); |
| } |
| float left_channel_energy = 0.0f; |
| float right_channel_energy = 0.0f; |
| for (int i = 0; i < processed_audio.frames(); ++i) { |
| left_channel_energy += |
| processed_audio.channel(0)[i] * processed_audio.channel(0)[i]; |
| right_channel_energy += |
| processed_audio.channel(1)[i] * processed_audio.channel(1)[i]; |
| } |
| if (use_apm && num_preferred_channels <= 1) { |
| // Mono output. Output channels are averaged. |
| EXPECT_NE(left_channel_energy, 0); |
| EXPECT_NE(right_channel_energy, 0); |
| } else { |
| // Stereo output. Output channels are independent. |
| // Note that after stereo mirroring, the _right_ channel is |
| // non-zero. |
| EXPECT_EQ(left_channel_energy, 0); |
| EXPECT_NE(right_channel_energy, 0); |
| } |
| }); |
| // Process one more frame of audio. |
| audio_processor->ProcessCapturedAudio(*data_bus, pushed_capture_time, |
| num_preferred_channels, 0.0, false); |
| } |
| } |
| } |
| |
| struct AudioProcessorDefaultOutputFormatTest |
| : public ::testing::Test, |
| public ::testing::WithParamInterface<std::tuple<bool, int>> { |
| static std::string PrintTestName( |
| const testing::TestParamInfo<ParamType>& info) { |
| auto [multichannel, sample_rate] = info.param; |
| return base::StringPrintf("MultichannelApm%sSampleRate%d", |
| multichannel ? "True" : "False", sample_rate); |
| } |
| }; |
| |
| INSTANTIATE_TEST_SUITE_P( |
| /*no prefix*/, |
| AudioProcessorDefaultOutputFormatTest, |
| ::testing::Combine(::testing::Bool(), |
| ::testing::ValuesIn(kSupportedSampleRates)), |
| &AudioProcessorDefaultOutputFormatTest::PrintTestName); |
| |
| TEST_P(AudioProcessorDefaultOutputFormatTest, GetDefaultOutputFormat) { |
| AudioProcessingSettings settings{.multi_channel_capture_processing = |
| std::get<0>(GetParam())}; |
| const int sample_rate = std::get<1>(GetParam()); |
| |
| media::AudioParameters input_params( |
| media::AudioParameters::AUDIO_PCM_LOW_LATENCY, |
| ChannelLayoutConfig::Stereo(), sample_rate, sample_rate / 100); |
| AudioParameters output_params = |
| AudioProcessor::GetDefaultOutputFormat(input_params, settings); |
| |
| // TODO(crbug.com/1336055): Investigate why chromecast devices need special |
| // logic here. See https://crrev.com/c/1572807 and |
| // https://crrev.com/c/3621456/comments/2e73cc96_0e9773cd for details. |
| const int expected_sample_rate = |
| #if BUILDFLAG(IS_CASTOS) || BUILDFLAG(IS_CAST_ANDROID) |
| std::min(sample_rate, media::kAudioProcessingSampleRateHz); |
| #else |
| media::kAudioProcessingSampleRateHz; |
| #endif |
| const int expected_output_channels = |
| settings.multi_channel_capture_processing ? input_params.channels() : 1; |
| |
| EXPECT_EQ(output_params.sample_rate(), expected_sample_rate); |
| EXPECT_EQ(output_params.channels(), expected_output_channels); |
| EXPECT_EQ(output_params.frames_per_buffer(), expected_sample_rate / 100); |
| } |
| |
| // Ensure that discrete channel layouts do not crash with audio processing |
| // enabled. |
| TEST_F(AudioProcessorTest, DiscreteChannelLayout) { |
| AudioProcessingSettings settings; |
| |
| // Test both 1 and 2 discrete channels. |
| for (int channels = 1; channels <= 2; ++channels) { |
| media::AudioParameters params(media::AudioParameters::AUDIO_PCM_LOW_LATENCY, |
| {media::CHANNEL_LAYOUT_DISCRETE, channels}, |
| 48000, 480); |
| std::unique_ptr<AudioProcessor> audio_processor = AudioProcessor::Create( |
| mock_capture_callback_.Get(), LogCallbackForTesting(), settings, params, |
| AudioProcessor::GetDefaultOutputFormat(params, settings)); |
| EXPECT_TRUE(audio_processor->has_webrtc_audio_processing()); |
| } |
| } |
| |
| class AudioProcessorPlayoutTest : public AudioProcessorTest { |
| protected: |
| AudioProcessorPlayoutTest() |
| : mock_webrtc_apm_( |
| rtc::make_ref_counted<webrtc::test::MockAudioProcessing>()), |
| audio_processor_(mock_capture_callback_.Get(), |
| LogCallbackForTesting(), |
| params_, |
| params_, |
| mock_webrtc_apm_, |
| /*stereo_mirroring=*/false, |
| /*needs_playout_reference=*/true) {} |
| |
| rtc::scoped_refptr<webrtc::test::MockAudioProcessing> mock_webrtc_apm_; |
| AudioProcessor audio_processor_; |
| }; |
| |
| TEST_F(AudioProcessorPlayoutTest, OnPlayoutData_ForwardsDataToWebrtcApm) { |
| std::unique_ptr<media::AudioBus> data_bus = |
| media::AudioBus::Create(/*channels=*/2, /*frames=*/480); |
| data_bus->Zero(); |
| |
| EXPECT_CALL(*mock_webrtc_apm_, AnalyzeReverseStream(_, _)).Times(1); |
| |
| audio_processor_.OnPlayoutData(*data_bus, 48000, base::Milliseconds(10)); |
| } |
| |
| TEST_F(AudioProcessorPlayoutTest, OnPlayoutData_BuffersPlayout) { |
| std::unique_ptr<media::AudioBus> data_bus = |
| media::AudioBus::Create(/*channels=*/2, /*frames=*/48000 * 4 / 1000); |
| data_bus->Zero(); |
| |
| // 5 buffers of 4 ms yields two 10 ms calls to APM. |
| EXPECT_CALL(*mock_webrtc_apm_, AnalyzeReverseStream(_, _)).Times(2); |
| |
| audio_processor_.OnPlayoutData(*data_bus, 48000, base::Milliseconds(10)); |
| audio_processor_.OnPlayoutData(*data_bus, 48000, base::Milliseconds(10)); |
| audio_processor_.OnPlayoutData(*data_bus, 48000, base::Milliseconds(10)); |
| audio_processor_.OnPlayoutData(*data_bus, 48000, base::Milliseconds(10)); |
| audio_processor_.OnPlayoutData(*data_bus, 48000, base::Milliseconds(10)); |
| } |
| |
| TEST_F(AudioProcessorPlayoutTest, OnPlayoutData_HandlesVariableInputSize) { |
| std::unique_ptr<media::AudioBus> long_data_bus = |
| media::AudioBus::Create(/*channels=*/2, /*frames=*/48000 * 25 / 1000); |
| long_data_bus->Zero(); |
| std::unique_ptr<media::AudioBus> short_data_bus = |
| media::AudioBus::Create(/*channels=*/2, /*frames=*/48000 * 5 / 1000); |
| short_data_bus->Zero(); |
| |
| // 25 ms + 5 ms yields three 10 ms calls to APM. |
| EXPECT_CALL(*mock_webrtc_apm_, AnalyzeReverseStream(_, _)).Times(3); |
| |
| audio_processor_.OnPlayoutData(*long_data_bus, 48000, base::Milliseconds(10)); |
| audio_processor_.OnPlayoutData(*short_data_bus, 48000, |
| base::Milliseconds(10)); |
| } |
| |
| TEST_F(AudioProcessorPlayoutTest, OnPlayoutData_HandlesSampleRateChange) { |
| std::unique_ptr<media::AudioBus> high_rate_data_bus = |
| media::AudioBus::Create(/*channels=*/2, /*frames=*/48000 * 12 / 1000); |
| high_rate_data_bus->Zero(); |
| std::unique_ptr<media::AudioBus> low_rate_data_bus = |
| media::AudioBus::Create(/*channels=*/2, /*frames=*/32000 * 18 / 1000); |
| low_rate_data_bus->Zero(); |
| |
| // 12 ms yields one 10 ms call to APM and leaves 2 ms in the buffer. |
| EXPECT_CALL(*mock_webrtc_apm_, AnalyzeReverseStream(_, _)).Times(1); |
| audio_processor_.OnPlayoutData(*high_rate_data_bus, 48000, |
| base::Milliseconds(10)); |
| |
| // 18 ms yields one 10 ms call to APM. Any previous buffer content should have |
| // been discarded, otherwise there would be more than one call to APM. |
| EXPECT_CALL(*mock_webrtc_apm_, AnalyzeReverseStream(_, _)).Times(1); |
| audio_processor_.OnPlayoutData(*low_rate_data_bus, 32000, |
| base::Milliseconds(10)); |
| } |
| |
| // When audio processing is performed, processed audio should be delivered as |
| // soon as 10 ms of audio has been received. |
| TEST(AudioProcessorCallbackTest, |
| ProcessedAudioIsDeliveredAsSoonAsPossibleWithShortBuffers) { |
| MockProcessedCaptureCallback mock_capture_callback; |
| AudioProcessingSettings settings; |
| // Set buffer size to 4 ms. |
| media::AudioParameters params(media::AudioParameters::AUDIO_PCM_LOW_LATENCY, |
| ChannelLayoutConfig::Stereo(), 48000, |
| 48000 * 4 / 1000); |
| std::unique_ptr<AudioProcessor> audio_processor = AudioProcessor::Create( |
| mock_capture_callback.Get(), LogCallbackForTesting(), settings, params, |
| AudioProcessor::GetDefaultOutputFormat(params, settings)); |
| ASSERT_TRUE(audio_processor->has_webrtc_audio_processing()); |
| int output_sample_rate = audio_processor->output_format().sample_rate(); |
| std::unique_ptr<media::AudioBus> data_bus = |
| media::AudioBus::Create(params.channels(), params.frames_per_buffer()); |
| data_bus->Zero(); |
| |
| auto check_audio_length = [&](const media::AudioBus& processed_audio, |
| base::TimeTicks, std::optional<double>) { |
| EXPECT_EQ(processed_audio.frames(), output_sample_rate * 10 / 1000); |
| }; |
| |
| // 4 ms of data: Not enough to process. |
| EXPECT_CALL(mock_capture_callback, Run(_, _, _)).Times(0); |
| audio_processor->ProcessCapturedAudio(*data_bus, base::TimeTicks::Now(), -1, |
| 1.0, false); |
| // 8 ms of data: Not enough to process. |
| EXPECT_CALL(mock_capture_callback, Run(_, _, _)).Times(0); |
| audio_processor->ProcessCapturedAudio(*data_bus, base::TimeTicks::Now(), -1, |
| 1.0, false); |
| // 12 ms of data: Should trigger callback, with 2 ms left in the processor. |
| EXPECT_CALL(mock_capture_callback, Run(_, _, _)) |
| .Times(1) |
| .WillOnce(check_audio_length); |
| audio_processor->ProcessCapturedAudio(*data_bus, base::TimeTicks::Now(), -1, |
| 1.0, false); |
| // 2 + 4 ms of data: Not enough to process. |
| EXPECT_CALL(mock_capture_callback, Run(_, _, _)).Times(0); |
| audio_processor->ProcessCapturedAudio(*data_bus, base::TimeTicks::Now(), -1, |
| 1.0, false); |
| // 10 ms of data: Should trigger callback. |
| EXPECT_CALL(mock_capture_callback, Run(_, _, _)) |
| .Times(1) |
| .WillOnce(check_audio_length); |
| audio_processor->ProcessCapturedAudio(*data_bus, base::TimeTicks::Now(), -1, |
| 1.0, false); |
| } |
| |
| // When audio processing is performed, input containing 10 ms several times over |
| // should trigger a comparable number of processing callbacks. |
| TEST(AudioProcessorCallbackTest, |
| ProcessedAudioIsDeliveredAsSoonAsPossibleWithLongBuffers) { |
| MockProcessedCaptureCallback mock_capture_callback; |
| AudioProcessingSettings settings; |
| // Set buffer size to 35 ms. |
| media::AudioParameters params(media::AudioParameters::AUDIO_PCM_LOW_LATENCY, |
| ChannelLayoutConfig::Stereo(), 48000, |
| 48000 * 35 / 1000); |
| std::unique_ptr<AudioProcessor> audio_processor = AudioProcessor::Create( |
| mock_capture_callback.Get(), LogCallbackForTesting(), settings, params, |
| AudioProcessor::GetDefaultOutputFormat(params, settings)); |
| ASSERT_TRUE(audio_processor->has_webrtc_audio_processing()); |
| int output_sample_rate = audio_processor->output_format().sample_rate(); |
| std::unique_ptr<media::AudioBus> data_bus = |
| media::AudioBus::Create(params.channels(), params.frames_per_buffer()); |
| data_bus->Zero(); |
| |
| auto check_audio_length = [&](const media::AudioBus& processed_audio, |
| base::TimeTicks, std::optional<double>) { |
| EXPECT_EQ(processed_audio.frames(), output_sample_rate * 10 / 1000); |
| }; |
| |
| // 35 ms of audio --> 3 chunks of 10 ms, and 5 ms left in the processor. |
| EXPECT_CALL(mock_capture_callback, Run(_, _, _)) |
| .Times(3) |
| .WillRepeatedly(check_audio_length); |
| audio_processor->ProcessCapturedAudio(*data_bus, base::TimeTicks::Now(), -1, |
| 1.0, false); |
| // 5 + 35 ms of audio --> 4 chunks of 10 ms. |
| EXPECT_CALL(mock_capture_callback, Run(_, _, _)) |
| .Times(4) |
| .WillRepeatedly(check_audio_length); |
| audio_processor->ProcessCapturedAudio(*data_bus, base::TimeTicks::Now(), -1, |
| 1.0, false); |
| } |
| |
| // When no audio processing is performed, audio is delivered immediately. Note |
| // that unlike the other cases, unprocessed audio input of less than 10 ms is |
| // forwarded directly instead of collecting chunks of 10 ms. |
| TEST(AudioProcessorCallbackTest, |
| UnprocessedAudioIsDeliveredImmediatelyWithShortBuffers) { |
| MockProcessedCaptureCallback mock_capture_callback; |
| AudioProcessingSettings settings; |
| DisableDefaultSettings(settings); |
| // Set buffer size to 4 ms. |
| media::AudioParameters params(media::AudioParameters::AUDIO_PCM_LOW_LATENCY, |
| ChannelLayoutConfig::Stereo(), 48000, |
| 48000 * 4 / 1000); |
| std::unique_ptr<AudioProcessor> audio_processor = AudioProcessor::Create( |
| mock_capture_callback.Get(), LogCallbackForTesting(), settings, params, |
| AudioProcessor::GetDefaultOutputFormat(params, settings)); |
| ASSERT_FALSE(audio_processor->has_webrtc_audio_processing()); |
| int output_sample_rate = audio_processor->output_format().sample_rate(); |
| std::unique_ptr<media::AudioBus> data_bus = |
| media::AudioBus::Create(params.channels(), params.frames_per_buffer()); |
| data_bus->Zero(); |
| |
| auto check_audio_length = [&](const media::AudioBus& processed_audio, |
| base::TimeTicks, std::optional<double>) { |
| EXPECT_EQ(processed_audio.frames(), output_sample_rate * 4 / 1000); |
| }; |
| |
| EXPECT_CALL(mock_capture_callback, Run(_, _, _)) |
| .Times(1) |
| .WillOnce(check_audio_length); |
| audio_processor->ProcessCapturedAudio(*data_bus, base::TimeTicks::Now(), -1, |
| 1.0, false); |
| EXPECT_CALL(mock_capture_callback, Run(_, _, _)) |
| .Times(1) |
| .WillOnce(check_audio_length); |
| audio_processor->ProcessCapturedAudio(*data_bus, base::TimeTicks::Now(), -1, |
| 1.0, false); |
| } |
| |
| // When no audio processing is performed, audio is delivered immediately. Chunks |
| // greater than 10 ms are delivered in chunks of 10 ms. |
| TEST(AudioProcessorCallbackTest, |
| UnprocessedAudioIsDeliveredImmediatelyWithLongBuffers) { |
| MockProcessedCaptureCallback mock_capture_callback; |
| AudioProcessingSettings settings; |
| DisableDefaultSettings(settings); |
| // Set buffer size to 35 ms. |
| media::AudioParameters params(media::AudioParameters::AUDIO_PCM_LOW_LATENCY, |
| ChannelLayoutConfig::Stereo(), 48000, |
| 48000 * 35 / 1000); |
| std::unique_ptr<AudioProcessor> audio_processor = AudioProcessor::Create( |
| mock_capture_callback.Get(), LogCallbackForTesting(), settings, params, |
| AudioProcessor::GetDefaultOutputFormat(params, settings)); |
| ASSERT_FALSE(audio_processor->has_webrtc_audio_processing()); |
| int output_sample_rate = audio_processor->output_format().sample_rate(); |
| std::unique_ptr<media::AudioBus> data_bus = |
| media::AudioBus::Create(params.channels(), params.frames_per_buffer()); |
| data_bus->Zero(); |
| |
| auto check_audio_length = [&](const media::AudioBus& processed_audio, |
| base::TimeTicks, std::optional<double>) { |
| EXPECT_EQ(processed_audio.frames(), output_sample_rate * 10 / 1000); |
| }; |
| |
| // 35 ms of audio --> 3 chunks of 10 ms, and 5 ms left in the processor. |
| EXPECT_CALL(mock_capture_callback, Run(_, _, _)) |
| .Times(3) |
| .WillRepeatedly(check_audio_length); |
| audio_processor->ProcessCapturedAudio(*data_bus, base::TimeTicks::Now(), -1, |
| 1.0, false); |
| // 5 + 35 ms of audio --> 4 chunks of 10 ms. |
| EXPECT_CALL(mock_capture_callback, Run(_, _, _)) |
| .Times(4) |
| .WillRepeatedly(check_audio_length); |
| audio_processor->ProcessCapturedAudio(*data_bus, base::TimeTicks::Now(), -1, |
| 1.0, false); |
| } |
| |
| class ApmTellsIfPlayoutReferenceIsNeededParametrizedTest |
| : public ::testing::TestWithParam<bool> { |
| public: |
| ApmTellsIfPlayoutReferenceIsNeededParametrizedTest() { |
| if (GetParam()) { |
| feature_list_.InitAndEnableFeature( |
| features::kWebRtcApmTellsIfPlayoutReferenceIsNeeded); |
| } else { |
| feature_list_.InitAndDisableFeature( |
| features::kWebRtcApmTellsIfPlayoutReferenceIsNeeded); |
| } |
| } |
| |
| private: |
| ::base::test::ScopedFeatureList feature_list_; |
| }; |
| |
| // Checks that, when all the audio processing settings are disabled, APM does |
| // not need the playout reference. |
| TEST_P(ApmTellsIfPlayoutReferenceIsNeededParametrizedTest, |
| DoesNotNeedPlayoutReference) { |
| AudioProcessingSettings settings; |
| DisableDefaultSettings(settings); |
| |
| MockProcessedCaptureCallback mock_capture_callback; |
| media::AudioParameters params(media::AudioParameters::AUDIO_PCM_LOW_LATENCY, |
| ChannelLayoutConfig::Stereo(), 48000, 480); |
| std::unique_ptr<AudioProcessor> audio_processor = AudioProcessor::Create( |
| mock_capture_callback.Get(), LogCallbackForTesting(), settings, params, |
| AudioProcessor::GetDefaultOutputFormat(params, settings)); |
| |
| EXPECT_FALSE(audio_processor->needs_playout_reference()); |
| } |
| |
| // Checks that, with echo cancellation, APM always needs the playout reference. |
| #if BUILDFLAG(IS_IOS) |
| #define MAYBE_NeedsPlayoutReference DISABLED_NeedsPlayoutReference |
| #else |
| #define MAYBE_NeedsPlayoutReference NeedsPlayoutReference |
| #endif |
| // TODO: This test is disabled for ios-blink platform as per the discussion on |
| // bug https://crbug.com/1417474 |
| TEST_P(ApmTellsIfPlayoutReferenceIsNeededParametrizedTest, |
| MAYBE_NeedsPlayoutReference) { |
| AudioProcessingSettings settings; |
| DisableDefaultSettings(settings); |
| settings.echo_cancellation = true; |
| |
| MockProcessedCaptureCallback mock_capture_callback; |
| media::AudioParameters params(media::AudioParameters::AUDIO_PCM_LOW_LATENCY, |
| ChannelLayoutConfig::Stereo(), 48000, 480); |
| std::unique_ptr<AudioProcessor> audio_processor = AudioProcessor::Create( |
| mock_capture_callback.Get(), LogCallbackForTesting(), settings, params, |
| AudioProcessor::GetDefaultOutputFormat(params, settings)); |
| |
| EXPECT_TRUE(audio_processor->needs_playout_reference()); |
| } |
| |
| INSTANTIATE_TEST_SUITE_P(AudioProcessor, |
| ApmTellsIfPlayoutReferenceIsNeededParametrizedTest, |
| ::testing::Bool()); |
| |
| } // namespace media |