Remove additional channel constraints when Beamforming is enabled in AudioProcessing

The general constraints on number of channels for AudioProcessing is:
num_in_channels == num_out_channels || num_out_channels == 1

When Beamforming is enabled and additional constraint was added forcing:
num_out_channels == 1

This artificial constraint was removed by adding upmixing support in CopyTo, since it was already supported for the AudioFrame interface using InterleaveTo.

Review URL: https://codereview.webrtc.org/1571013002

Cr-Commit-Position: refs/heads/master@{#11215}
This commit is contained in:
aluebs
2016-01-11 20:32:29 -08:00
committed by Commit bot
parent e93ad1b129
commit b2328d11dc
7 changed files with 35 additions and 23 deletions

View File

@ -78,6 +78,7 @@ class FakeAudioProcessing : public webrtc::AudioProcessing {
WEBRTC_STUB_CONST(proc_sample_rate_hz, ()); WEBRTC_STUB_CONST(proc_sample_rate_hz, ());
WEBRTC_STUB_CONST(proc_split_sample_rate_hz, ()); WEBRTC_STUB_CONST(proc_split_sample_rate_hz, ());
WEBRTC_STUB_CONST(num_input_channels, ()); WEBRTC_STUB_CONST(num_input_channels, ());
WEBRTC_STUB_CONST(num_proc_channels, ());
WEBRTC_STUB_CONST(num_output_channels, ()); WEBRTC_STUB_CONST(num_output_channels, ());
WEBRTC_STUB_CONST(num_reverse_channels, ()); WEBRTC_STUB_CONST(num_reverse_channels, ());
WEBRTC_VOID_STUB(set_output_will_be_muted, (bool muted)); WEBRTC_VOID_STUB(set_output_will_be_muted, (bool muted));

View File

@ -150,7 +150,7 @@ void AudioBuffer::CopyFrom(const float* const* data,
void AudioBuffer::CopyTo(const StreamConfig& stream_config, void AudioBuffer::CopyTo(const StreamConfig& stream_config,
float* const* data) { float* const* data) {
assert(stream_config.num_frames() == output_num_frames_); assert(stream_config.num_frames() == output_num_frames_);
assert(stream_config.num_channels() == num_channels_); assert(stream_config.num_channels() == num_channels_ || num_channels_ == 1);
// Convert to the float range. // Convert to the float range.
float* const* data_ptr = data; float* const* data_ptr = data;
@ -173,6 +173,11 @@ void AudioBuffer::CopyTo(const StreamConfig& stream_config,
output_num_frames_); output_num_frames_);
} }
} }
// Upmix.
for (int i = num_channels_; i < stream_config.num_channels(); ++i) {
memcpy(data[i], data[0], output_num_frames_ * sizeof(**data));
}
} }
void AudioBuffer::InitForNewData() { void AudioBuffer::InitForNewData() {

View File

@ -226,9 +226,9 @@ AudioProcessingImpl::AudioProcessingImpl(const Config& config,
#else #else
capture_(config.Get<ExperimentalNs>().enabled, capture_(config.Get<ExperimentalNs>().enabled,
#endif #endif
config.Get<Beamforming>().enabled,
config.Get<Beamforming>().array_geometry, config.Get<Beamforming>().array_geometry,
config.Get<Beamforming>().target_direction) config.Get<Beamforming>().target_direction),
capture_nonlocked_(config.Get<Beamforming>().enabled)
{ {
{ {
rtc::CritScope cs_render(&crit_render_); rtc::CritScope cs_render(&crit_render_);
@ -345,7 +345,7 @@ int AudioProcessingImpl::MaybeInitialize(
int AudioProcessingImpl::InitializeLocked() { int AudioProcessingImpl::InitializeLocked() {
const int fwd_audio_buffer_channels = const int fwd_audio_buffer_channels =
capture_.beamformer_enabled capture_nonlocked_.beamformer_enabled
? formats_.api_format.input_stream().num_channels() ? formats_.api_format.input_stream().num_channels()
: formats_.api_format.output_stream().num_channels(); : formats_.api_format.output_stream().num_channels();
const int rev_audio_buffer_out_num_frames = const int rev_audio_buffer_out_num_frames =
@ -428,9 +428,8 @@ int AudioProcessingImpl::InitializeLocked(const ProcessingConfig& config) {
return kBadNumberChannelsError; return kBadNumberChannelsError;
} }
if (capture_.beamformer_enabled && if (capture_nonlocked_.beamformer_enabled &&
(static_cast<size_t>(num_in_channels) != capture_.array_geometry.size() || static_cast<size_t>(num_in_channels) != capture_.array_geometry.size()) {
num_out_channels > 1)) {
return kBadNumberChannelsError; return kBadNumberChannelsError;
} }
@ -500,8 +499,9 @@ void AudioProcessingImpl::SetExtraOptions(const Config& config) {
} }
#ifdef WEBRTC_ANDROID_PLATFORM_BUILD #ifdef WEBRTC_ANDROID_PLATFORM_BUILD
if (capture_.beamformer_enabled != config.Get<Beamforming>().enabled) { if (capture_nonlocked_.beamformer_enabled !=
capture_.beamformer_enabled = config.Get<Beamforming>().enabled; config.Get<Beamforming>().enabled) {
capture_nonlocked_.beamformer_enabled = config.Get<Beamforming>().enabled;
if (config.Get<Beamforming>().array_geometry.size() > 1) { if (config.Get<Beamforming>().array_geometry.size() > 1) {
capture_.array_geometry = config.Get<Beamforming>().array_geometry; capture_.array_geometry = config.Get<Beamforming>().array_geometry;
} }
@ -537,6 +537,11 @@ int AudioProcessingImpl::num_input_channels() const {
return formats_.api_format.input_stream().num_channels(); return formats_.api_format.input_stream().num_channels();
} }
int AudioProcessingImpl::num_proc_channels() const {
// Used as callback from submodules, hence locking is not allowed.
return capture_nonlocked_.beamformer_enabled ? 1 : num_output_channels();
}
int AudioProcessingImpl::num_output_channels() const { int AudioProcessingImpl::num_output_channels() const {
// Used as callback from submodules, hence locking is not allowed. // Used as callback from submodules, hence locking is not allowed.
return formats_.api_format.output_stream().num_channels(); return formats_.api_format.output_stream().num_channels();
@ -771,7 +776,7 @@ int AudioProcessingImpl::ProcessStreamLocked() {
ca->num_channels()); ca->num_channels());
} }
if (capture_.beamformer_enabled) { if (capture_nonlocked_.beamformer_enabled) {
private_submodules_->beamformer->ProcessChunk(*ca->split_data_f(), private_submodules_->beamformer->ProcessChunk(*ca->split_data_f(),
ca->split_data_f()); ca->split_data_f());
ca->set_num_channels(1); ca->set_num_channels(1);
@ -793,7 +798,7 @@ int AudioProcessingImpl::ProcessStreamLocked() {
if (constants_.use_new_agc && if (constants_.use_new_agc &&
public_submodules_->gain_control->is_enabled() && public_submodules_->gain_control->is_enabled() &&
(!capture_.beamformer_enabled || (!capture_nonlocked_.beamformer_enabled ||
private_submodules_->beamformer->is_target_present())) { private_submodules_->beamformer->is_target_present())) {
private_submodules_->agc_manager->Process( private_submodules_->agc_manager->Process(
ca->split_bands_const(0)[kBand0To8kHz], ca->num_frames_per_band(), ca->split_bands_const(0)[kBand0To8kHz], ca->num_frames_per_band(),
@ -1183,7 +1188,7 @@ VoiceDetection* AudioProcessingImpl::voice_detection() const {
} }
bool AudioProcessingImpl::is_data_processed() const { bool AudioProcessingImpl::is_data_processed() const {
if (capture_.beamformer_enabled) { if (capture_nonlocked_.beamformer_enabled) {
return true; return true;
} }
@ -1293,12 +1298,12 @@ void AudioProcessingImpl::InitializeTransient() {
public_submodules_->transient_suppressor->Initialize( public_submodules_->transient_suppressor->Initialize(
capture_nonlocked_.fwd_proc_format.sample_rate_hz(), capture_nonlocked_.fwd_proc_format.sample_rate_hz(),
capture_nonlocked_.split_rate, capture_nonlocked_.split_rate,
formats_.api_format.output_stream().num_channels()); num_proc_channels());
} }
} }
void AudioProcessingImpl::InitializeBeamformer() { void AudioProcessingImpl::InitializeBeamformer() {
if (capture_.beamformer_enabled) { if (capture_nonlocked_.beamformer_enabled) {
if (!private_submodules_->beamformer) { if (!private_submodules_->beamformer) {
private_submodules_->beamformer.reset(new NonlinearBeamformer( private_submodules_->beamformer.reset(new NonlinearBeamformer(
capture_.array_geometry, capture_.target_direction)); capture_.array_geometry, capture_.target_direction));
@ -1320,12 +1325,12 @@ void AudioProcessingImpl::InitializeIntelligibility() {
} }
void AudioProcessingImpl::InitializeHighPassFilter() { void AudioProcessingImpl::InitializeHighPassFilter() {
public_submodules_->high_pass_filter->Initialize(num_output_channels(), public_submodules_->high_pass_filter->Initialize(num_proc_channels(),
proc_sample_rate_hz()); proc_sample_rate_hz());
} }
void AudioProcessingImpl::InitializeNoiseSuppression() { void AudioProcessingImpl::InitializeNoiseSuppression() {
public_submodules_->noise_suppression->Initialize(num_output_channels(), public_submodules_->noise_suppression->Initialize(num_proc_channels(),
proc_sample_rate_hz()); proc_sample_rate_hz());
} }

View File

@ -102,6 +102,7 @@ class AudioProcessingImpl : public AudioProcessing {
int proc_sample_rate_hz() const override; int proc_sample_rate_hz() const override;
int proc_split_sample_rate_hz() const override; int proc_split_sample_rate_hz() const override;
int num_input_channels() const override; int num_input_channels() const override;
int num_proc_channels() const override;
int num_output_channels() const override; int num_output_channels() const override;
int num_reverse_channels() const override; int num_reverse_channels() const override;
int stream_delay_ms() const override; int stream_delay_ms() const override;
@ -280,7 +281,6 @@ class AudioProcessingImpl : public AudioProcessing {
struct ApmCaptureState { struct ApmCaptureState {
ApmCaptureState(bool transient_suppressor_enabled, ApmCaptureState(bool transient_suppressor_enabled,
bool beamformer_enabled,
const std::vector<Point>& array_geometry, const std::vector<Point>& array_geometry,
SphericalPointf target_direction) SphericalPointf target_direction)
: aec_system_delay_jumps(-1), : aec_system_delay_jumps(-1),
@ -292,7 +292,6 @@ class AudioProcessingImpl : public AudioProcessing {
output_will_be_muted(false), output_will_be_muted(false),
key_pressed(false), key_pressed(false),
transient_suppressor_enabled(transient_suppressor_enabled), transient_suppressor_enabled(transient_suppressor_enabled),
beamformer_enabled(beamformer_enabled),
array_geometry(array_geometry), array_geometry(array_geometry),
target_direction(target_direction), target_direction(target_direction),
fwd_proc_format(kSampleRate16kHz), fwd_proc_format(kSampleRate16kHz),
@ -306,7 +305,6 @@ class AudioProcessingImpl : public AudioProcessing {
bool output_will_be_muted; bool output_will_be_muted;
bool key_pressed; bool key_pressed;
bool transient_suppressor_enabled; bool transient_suppressor_enabled;
bool beamformer_enabled;
std::vector<Point> array_geometry; std::vector<Point> array_geometry;
SphericalPointf target_direction; SphericalPointf target_direction;
rtc::scoped_ptr<AudioBuffer> capture_audio; rtc::scoped_ptr<AudioBuffer> capture_audio;
@ -318,16 +316,18 @@ class AudioProcessingImpl : public AudioProcessing {
} capture_ GUARDED_BY(crit_capture_); } capture_ GUARDED_BY(crit_capture_);
struct ApmCaptureNonLockedState { struct ApmCaptureNonLockedState {
ApmCaptureNonLockedState() ApmCaptureNonLockedState(bool beamformer_enabled)
: fwd_proc_format(kSampleRate16kHz), : fwd_proc_format(kSampleRate16kHz),
split_rate(kSampleRate16kHz), split_rate(kSampleRate16kHz),
stream_delay_ms(0) {} stream_delay_ms(0),
beamformer_enabled(beamformer_enabled) {}
// Only the rate and samples fields of fwd_proc_format_ are used because the // Only the rate and samples fields of fwd_proc_format_ are used because the
// forward processing number of channels is mutable and is tracked by the // forward processing number of channels is mutable and is tracked by the
// capture_audio_. // capture_audio_.
StreamConfig fwd_proc_format; StreamConfig fwd_proc_format;
int split_rate; int split_rate;
int stream_delay_ms; int stream_delay_ms;
bool beamformer_enabled;
} capture_nonlocked_; } capture_nonlocked_;
struct ApmRenderState { struct ApmRenderState {

View File

@ -174,7 +174,7 @@ int EchoCancellationImpl::ProcessCaptureAudio(AudioBuffer* audio) {
} }
assert(audio->num_frames_per_band() <= 160); assert(audio->num_frames_per_band() <= 160);
assert(audio->num_channels() == apm_->num_output_channels()); assert(audio->num_channels() == apm_->num_proc_channels());
int err = AudioProcessing::kNoError; int err = AudioProcessing::kNoError;

View File

@ -435,7 +435,7 @@ int GainControlImpl::ConfigureHandle(void* handle) const {
int GainControlImpl::num_handles_required() const { int GainControlImpl::num_handles_required() const {
// Not locked as it only relies on APM public API which is threadsafe. // Not locked as it only relies on APM public API which is threadsafe.
return apm_->num_output_channels(); return apm_->num_proc_channels();
} }
int GainControlImpl::GetHandleError(void* handle) const { int GainControlImpl::GetHandleError(void* handle) const {

View File

@ -288,6 +288,7 @@ class AudioProcessing {
virtual int proc_sample_rate_hz() const = 0; virtual int proc_sample_rate_hz() const = 0;
virtual int proc_split_sample_rate_hz() const = 0; virtual int proc_split_sample_rate_hz() const = 0;
virtual int num_input_channels() const = 0; virtual int num_input_channels() const = 0;
virtual int num_proc_channels() const = 0;
virtual int num_output_channels() const = 0; virtual int num_output_channels() const = 0;
virtual int num_reverse_channels() const = 0; virtual int num_reverse_channels() const = 0;