SciELO - Scientific Electronic Library Online

 
vol.24 issue3Ground Truth Spanish Automatic Extractive Text Summarization BoundsAutonomous Drone Racing with an Opponent: A First Approach author indexsubject indexsearch form
Home Pagealphabetic serial listing  

Services on Demand

Journal

Article

Indicators

Related links

  • Have no similar articlesSimilars in SciELO

Share


Computación y Sistemas

On-line version ISSN 2007-9737Print version ISSN 1405-5546

Comp. y Sist. vol.24 n.3 Ciudad de México Jul./Sep. 2020  Epub June 09, 2021

https://doi.org/10.13053/cys-24-3-3485 

Articles

Lightweight Online Separation of the Sound Source of Interest through BLSTM-Based Binary Masking

Alejandro Maldonado1 

Caleb Rascón2  * 

Ivette Vélez1 

11 Universidad Nacional Autónoma de México, Posgrado en Ciencia e Ingeniería de la Computación, Mexico

22 Universidad Nacional Autónoma de México, Instituto de Investigaciones en Matemáticas Avanzadas y en Sistemas, Mexico


Abstract:

Online audio source separation has been an important part of auditory scene analysis and robot audition. The main type of technique to carry this out, because of its online capabilities, has been spatial filtering (or beamforming), where it is assumed that the location (mainly, the direction of arrival; DOA) of the source of interest (SOI) is known. However, these techniques suffer from considerable interference leakage in the final result. In this paper, we propose a two step technique: 1) a phase-based beamformer that provides, in addition to the estimation of the SOI, an estimation of the cumulative environmental interference; and 2) a BLSTM-based TF binary masking stage that calculates a binary mask that aims to separate the SOI from the cumulative environmental interference. In our tests, this technique provides a signal-to-interference ratio (SIR) above 20 dB with simulated data. Because of the nature of the beamformer outputs, the label permutation problem is handled from the beginning. This makes the proposed solution a lightweight alternative that requires considerably less computational resources (almost an order of magnitude) compared to current deep-learning based techniques, while providing a comparable SIR performance.

Keywords: Beamforming; BLSTM; permutation problem; binary mask

1 Introduction

Sound source separation is an essential step in the processing chain of events in computational auditory scene analysis [32] (CASA) and robot audition [28, 19] (RA). Currently, many sound-related tasks such as automatic speech recognition, speaker identification, and mood classification, assume that their input bares only the audio data from the source to be analyzed.

Many of the techniques used for carrying these tasks are based on machine learning methods, which could be made robust against multiple-source scenarios by augmenting their corresponding training corpora. However, another alternative could be to have a source separation phase beforehand that provides the audio information from one source at a time. To this effect, current techniques could still be used by this alternative, without requiring impractical amounts of space and time for training.

In terms of sound source separation, it is of interest to carry it out in an online manner (meaning, “on the fly”), for scenarios in which the user is interacting with a CASA/RA system, such as a service robot, a virtual assistant, a security system, etc. This is opposed to an offline manner, which records the audio from the environment and returns the results of the audio analysis after the interaction is completed.

It is important to note that we are differentiating between carrying this analysis out in an online manner and carrying it out in real-time, since the latter involves discussion of specific response times thresholds [12]. What we define as online analysis is that the system’s response time is less than the length of the time window that is to be processed. Meaning that, even though results are given while the user is interacting, they may be given with a certain delay. In certain applications, like human-robot interaction (HRI), online results are essential for the interaction to be successful, while a reasonable delay ( 1 second) is acceptable [12].

One of the most popular type of techniques to carry out online source separation is by spatial filtering or beamforming. It is important to note that the terms “source separation” and “beamforming” are not usually used in the same context. While source separation aims to separate all sources present in the recorded environment, beamforming aims to separate one source of interest (the location of which is known a priori) from the rest of the environment. In this paper, we are equating these two terms, since beamforming is carrying out a type of source separation, and it is usually designed to be carried out in an online manner. And, while beamforming only separates one sound source from the environment, it is compatible with the aforementioned CASA/RA/HRI application scenarios in which one user is attended at a time, i.e. the source of interest (SOI).

Unfortunately, an important issue with beam-forming is that of interference leakage, in which sound sources different from the SOI are still present in the final result. Although this interference presence can be low if the beamformer is configured appropriately [11], it is still high enough to be perceivable (with signal-to-interference ratios less than 15 dB), which may have an impact in subsequent CASA/RA modules [22]. Thus, beamforming techniques tend to employ a high number of microphones to overcome this issue [25].

On the other hand, deep learning strategies have shown impressive results when carrying out source separation, even when only a single microphone is used [15, 14, 3]. A popular methodology is to classify which frequency bins belong to which sound source, i.e. frequency masking. To carry out this through time, many of these techniques track the frequential variation of each source, so that in each time window the appropriate time-frequency (TF) bins is assigned to the correct source. If this tracking is done incorrectly, one source may be assigned data from others in different time windows, corrupting the overall output. Solving this requires complex solutions that require an important amount of computational resources (which may be an important issue for some CASA/RA/HRI application scenarios). Or, in the worst case, the problem is bounded such that the proposed techniques are only tested with recordings with a few amount of interferences [15, 14, 3].

To overcome this problem, we proposed a novel lightweight source separation technique which carries out deep-learning-based frequency masking from a beamforming output. The proposed solution can be run online and is robust against variations of interferences and number of microphones. It is composed of two parts: 1) a phase-based time-frequency-masking beamforming that provides both the estimation of the SOI and the estimation of the cumulative environmental interference; and 2) a time-frequency binary masking stage based on a bidirectional long short-term memory (BLSTM) network, that aims to use these two estimations to separate the SOI from the environmental interference estimation.

Since the proposed beamformer is already providing a preliminary separation of the TF bins of the SOI from the TF bins of the interferences, the permutation problem [14] is solved from the beginning. This means that the complexity and size of the BLSTM network architecture is low enough to be run in an online manner, even with modest computer equipment. The full system is found herefn.

The work here presented has the following structure: Section 2 provides a brief summary of the related works and background relevant to the proposed technique; Section 3 details the proposed technique; Section 4 describes the evaluation methodology against a deep-clustering-based source separation approach and presents the results; Section 5 discusses the insights obtained from these results; and Section 6 provides our conclusions and future work.

2 Background and Related Work

As mentioned before, we are aiming to use a beamforming paradigm to carry out online sound source separation, which implies that the location of the source of interest (SOI) is known a priori. This approach is popular in Robot Audition (RA), as shown by HARK [20] and ManyEars [9], both of which employ a real-time variation of the geometric source separation technique [29].

This technique merges both the beamforming paradigm with a blind sound source separation approach. It is worth mentioning that HARK has modified this technique even further by introducing adaptability to the inner mechanisms of geometric source separation [21], and that ManyEars has pushed for being more lightweight, with its ODAS project [10]fn. In all these circumstances, the direction of arrival (DOA) of the sound source is assumed to be known a priori, or estimated by applying one of the many sound source localization techniques reported in literature [25]. However, evaluation of these beamforming techniques has been bounded by the use of a considerable amount of microphones, which reduces the presence of interferences in the resulting SOI estimation.

It would be of interest to use less microphones, while avoiding the aforementioned interference leakage issue, when carrying out online sound source separation. A possible alternative to this would be to apply recent developments in mono-aural sound source separation, the vast majority of which employ deep-learning techniques such as bidirectional long short-term memory (BLSTM) networks [7]. This type of techniques are a type of recurrent network which are ideal to be used with temporal data, and have been a good answer to issues specific to recurrent networks, such as the vanishing gradient problem [16] and localized classification [4]. To this effect, they have shown very good results for speech recognition [6, 5] and text recognition [27].

However when applied to sound source separation, an important issue has been found: the permutation problem [33, 14]. Many of these techniques aim to classify each time-frequency (TF) bin of the input signal as to belonging to one of various possible sources. When carrying out this process throughout several time windows, it is essential that the TF bins are appropriately assigned such that classifications of previous time windows are correctly followed; if not, one source may be assigned TF bins of other sources in subsequent time windows, with unwanted overall results. To avoid this, several methods have been used such as permutation invariant training [37], deep clustering [14] and deep attractors [3].

Specifically, deep clustering has been successful in recent years for sound source separation. The Chimera network [18] is a representative example of this. Even though it was originally proposed for voice separation in music, it has been recently modified for its use in mono-aural source separation [34, 36, 35]. The Chimera network uses a two-front approach to design its objective loss function: 1) a type of deep-clustering loss function that transforms the input signal to a domain in which it is able to keep track of which sound source is which (by means of clustering methodologies), and 2) a magnitude spectrum approximation objective that aims to infer the TF mask to apply to the input signal. By training with this loss function, the network is made to consider which source is being assigned to which TF bin, resulting in strong Signal-to-Distortion performances.

However, as it can be deduced, an important amount of complexity is encountered when carrying out this methodology. This results in a complex solution space that the training optimization algorithm is expected to solve. It would be of interest to avoid this issue altogether, which not only would simplify the solution space to solve, but also may reduce its memory requirements and its response time.

It is important to mention the work of [24], which carries out a similar technique to ours. However, the authors of this work feed the network with features extracted from the beamformer weights. Although this process solves the permutation problem from the beginning, it implicitly trains the network with the array geometry (since the beamformer weights are based on it). It would be of interest to solve the permutation problem while the system is robust against array geometry changes.

Additionally, a hybrid approach of a beamformer and deep learning techniques has been employed before [33]. However, this hybrid approach is usually carried out by either: 1) making the deep learning network emulate the task of the beam-former, or 2) feeding the beamformer estimation of the source of interest as the mono-aural input to the deep learning network. As far as we know, feeding the deep learning network a two-channel input (one of the preliminary estimation of the SOI, and the other of the preliminary estimation of the cumulative environmental interference) has not been proposed before.

3 Proposed System

An overall summary of the proposed system is shown in Figure 1. As it can be seen, there are two core modules. First, the audio data from the microphone array and the direction of arrival of the source of interest (SOI) is fed to a phase-based frequency-masking beamformer that provides a preliminary estimation of the SOI (ZSOI) as well as of the cumulative environmental interference (ZINT ). Second, a time-frequency binary masking stage, based in a bidirectional long short-term memory (BLSTM) network, provides a time-frequency (TF) binary mask (BSOI) that separates the SOI from the cumulative environmental interference estimation. This mask is then applied to the signal of the reference microphone for the final SOI estimation (YSOI). In this section, these two core modules are detailed.

Fig. 1 An overall diagram of the proposed system 

3.1 Phase-Based Frequency Masking Beamformer

The proposed beamformer is summarized in Figure 2.

Fig. 2 A diagram summarizing the phase-based frequency-masking beamformer stage. 

Let X be the input matrix of size M × N, where M is the number of microphones and N is the time-window length in samples, as well as the length of the resulting frequency masks.

The columns of X are the Fourier transformed time-windows of each microphone input. Additionally, let θ be the direction of arrival (DOA) of the SOI. The first stage of the beamformer carries out a time-alignment of the columns of X such that the information received by the microphone array in the planar direction θ is in phase. This is carried out as described in Equation 1:

Xa[m;f]=X[m;f]ei2πftm;θ, (1)

where Xa is the phase-aligned version of X towards θ, m is the microphone index, f is the frequency bin, and tm;θ is the delay in seconds applied to the input data of microphone m based on θ.

Using the positions of the microphones relative to the reference microphone, each respective delay can be calculated via different methodologies, such as the far-field model [25] presented in Equation 2:

tm;θ=rmccos(θmθ), (2)

where c is the speed of sound (~ 343 meters per second), and rm and θm are the polar coordinates of microphone m in relation to the reference microphone (m = 1).

The average phase difference is then calculated for each frequency bin f, as described in Equation 3:

|φ|f=2M(M1)i=1M1j=i+1M|φi;fφj;f|, (3)

where M is the number of microphones, |ϕ|f is the average phase difference at frequency bin f, and ϕm;f is the phase at frequency bin f of microphone m.

Consequently, two frequency masks are created via an angular threshold (ϕmax), as described in Equations 4 and 5:

PSOI[f]={1,0,if|φ|fφmaxotherwise, (4)

PINT[f]={0,1,if|φ|fφmaxotherwise, (5)

where PSOI and PINT are the 1 × N frequency masks for the SOI and for the cumulative environmental interference, respectively.

The 1 × N estimations of the SOI (ZSOI) and the cumulative environmental interference (ZINT) are calculated by applying the corresponding frequency mask to the reference microphone, as described in Equations 6 and 7:

ZSOI[f]=PSOI[f]X[1;f], (6)

ZINT[f]=PINT[f]X[1;f]. (7)

Variations of this beamformer have been proposed before. The authors of [1] use a similar method, but instead of creating a binary mask, they create a soft mask by assuming a frequency-dependent phase variance and empirically accounting for it. It is important to note, however, that this work does not provide an estimation of the cumulative environmental interference.

Another similar work is that of [13], where the authors employ an interference-leakage removal strategy that requires the estimation of the frequency covariance matrix. This is similar to the strategy employed by the well known minimum variance distortionless response (MVDR) beamformer [17, 2], which has been shown to be too complex to be run in an online manner using the whole frequency spectrum [30]. It is important to note that variations of MVDR have been developed to run online, but the strategy employed in [13] has not been shown to do so.

As it can be concluded, the phase-based frequency masking beamformer proposed here is much less complex than those presented in the aforementioned works. Additionally, and more importantly, it provides the estimation of the cumulative environmental interference. As it will be discussed later, this is essential to solve the permutation problem for the BLSTM-based TF binary-masking stage, resulting in having a relatively low complexity.

It is important to mention that, although X represents a time-window length of N samples of input data, the input length NB used for the subsequent binary masking stage is conformed of several of these N-length windows, using a Hann-window-based overlap-and-add strategy (to avoid discontinuities when applying the short-time Fourier transform). To this effect, NB can be considered independent of N, in only that NB is a multiple of N.

3.2 BLSTM-Based TF Binary Masking

In Figure 3, the BLSTM-based time-frequency binary masking stage is summarized. In an overall sense, the purpose of this stage is to calculate a time-frequency binary mask (BSOI) which, when applied to the input data of the reference microphone, the SOI is separated from the cumulative environmental interference.

Fig. 3 Architecture of the proposed BLSTM network. 

As it can be seen in Figure 3, the BLSTM-based TF binary masking stage expects two inputs, one with the SOI estimation and another with the estimation of the cumulative environmental interference. These two time-domain inputs of length NB are transformed to the time-frequency (TF) domain via the short-time Fourier transform, using a Hann window with a length of NH samples and a 50% overlap.

This results in two matrices of size T × F. The size of time dimension T depends on NB such that T = (NB 2) + 1, because of the 50% overlap (with zero-padded Hann windows at the edges). As for the size of the frequency dimension F, it depends on NH: to avoid redundant weight calculations in the subsequent BLSTM network, only the lower half (with the DC component) of the mirrored Fourier transform is used, thus F=NH2+1.

The energy at each input TF bin is converted into the decibel scale (dB), and then standardized to have zero-median and one-standard-deviation. These two steps are important to mold the solution space to a shape that is easier to converge when training the subsequent BLSTM network. The standardized inputs are then concatenated in the frequency dimension.

The proposed BLSTM network is made up of L amount of BLSTM stacked layers with H amount of hidden units, which are then fed into a fully-connected layer, and has a softmax output layer that estimates the probability of the TF bin belonging to the SOI. Thus, the BLSTM network carries out a binary classification, which results in two T × F binary masks: one for the SOI (BSOI) and one for the cumulative environmental interference (BINT ), although only BSOI is used in later stages.

Once trained, as shown in Figure 1, the resulting BSOI is applied to the input data of the reference microphone, which is transformed to the time-frequency domain in the same manner as the outputs of the beamformer. This process, as described by Equation 8, results in the final SOI estimation YSOI in the time-frequency domain:

YSOI[t;f]=BSOI[t;f]X[1;t;f]. (8)

If the application requires it, the final estimation of the cumulative environmental interference (YINT ) can be obtained by applying BINT to the reference microphone, as shown in Equation 9:

YINT[t;f]=BINT[t;f]X[1;t;f]. (9)

3.2.1 Training and Validation

For training, the LibriSpeech corpus [23] was used, which is composed of 500 hours of clean recordings of users reading text, sampled at 16 kHz. The users were chosen randomly from 80% of this corpus to act as sound sources which were artificially mixed to simulate the inputs of a 2-microphone array; the other 20% was used for validation purposes. For the second microphone, each source was delayed according to a randomly chosen DOA for each sound source, applying the far-field model shown in Equation 2. The DOA was chosen in the [90o, 90o] range, at 45o intervals in the horizontal plane.

Additionally, the ideal TF mask (Ok) of each source k was calculated from the clean corpus recordings, and used as part of a magnitude spectrum approximation (MSA) objective function (), described in Equation 10.

=k=12(OkBk)S22, (10)

where k is either 1 for SOI or 2 for INT; Bk is the predicted mask; and S indicates the magnitude of the TF bin of the mixture from the reference microphone. This is similar as to what was carried out in [18].

During training, before delivering BSOI to the loss function, a simple voice-activity detection (VAD) mechanism [14] is employed, described in Equation 11:

ψ[t;f]={1,0,ifX[1;t;f]>X[1]maxVotherwise,BSOI[t;f]=ψ[t;f]BSOI[t;f], (11)

where ψ[t; f] is the VAD mask, the operator ||·|| calculates the decibel energy of a TF bin, V is the VAD energy threshold, and X[1]max is the maximum decibel energy of the reference microphone X[1] in an input length.

It is important to mention that the VAD step is only necessary during training, and not during testing. This is because, given the design of the loss function, the BLSTM network implicitly learns to ignore the TF bins usually discarded by the VAD process.

During training, the RMSProp optimizer was used with a learning rate of 10e5 and a momentum of 0.9, as employed by [14].

3.2.2 Architecture Selection

To select the architecture for the proposed BLSTM network, we evaluated different architecture configurations, trained with up to 3 sources (including the source of interest, meaning, with up to 2 interferences). In Table 1, their performance is reported in terms of the signal-to-interference ratio (SIR) in the output. This was measured using the BSS_EVAL_SOURCES algorithm [31] using the clean recordings of LibriSpeech as the basis of comparison. This table also reports the memoryfn occupied by each model.

Table 1 Evaluation of different configurations of proposed BLSTM model with up to 3 sources. 

NB H L Memory (MB) SIR (dB)
8192 200 1 16 19.69
8192 200 3 38 22.44
8192 200 5 60 22.68
8192 300 1 26 20.87
8192 300 3 76 23.67
8192 300 5 125 22.06
8192 400 1 39 21.82
8192 400 3 127 22.94
8192 400 5 215 22.17
8192 500 4 259 26.02
16384 200 1 16 20.99
16384 200 3 38 24.66
16384 200 5 60 22.88
16384 300 1 26 21.36
16384 300 3 76 23.38
16384 300 5 125 21.93
16384 400 1 39 22.68
16384 400 3 127 23.82
16384 400 5 215 22.65
16384 500 4 259 27.75
Average 98.1 22.82

The configurations vary in terms of number of BLSTM stacked layers (L), number of hidden units (H) and input length (NB). The results when varying other parameters are not reported since they did not provide considerable differences in the evaluations. Meaning, in these evaluations, the Hann-window length NH was set at 512 samples and V is set at 40 dB. In [18] the authors employed 4 BLSTM stacked layers and 500 hidden units, and obtained robust performances in mismatched conditions. Since the aim of this work is to minimize memory usage, these were chosen as the combined upper bound for L and H.

For H < 500, we tested L values of 1, 3 and 5 to provide a balanced view of the performance fluctuation when varying L. We also set ϕmax to 60o.

It is of interest to select an architecture configuration that both maximizes its SIR performance while minimizing its memory usage. To this effect, we calculate the area under the curve as defined in Equation 12 for each of the architecture configurations in Table 1:

y(x)={0,ifx<;0,(σaμa)if0<;x<;μa,σa,otherwise, (12)

where σa and µa are (respectively) the SIR and memory usage for each architecture configuration a presented in Table 1. The architecture configuration shown in bold in Table 1 (L: 3, H: 200, NB: 16384) has the largest area under the curve and, thus, the one we recommend to use. However, consideration should be given to the configuration shown in italics (L: 3, H: 300, NB: 8192), since it not only provides the second largest area under the curve, but it also uses a smaller NB (which is close to 0.5 seconds when sampling at 16 kHz).

4 Evaluation and Results

To investigate the behavior of the proposed system, three evaluations were carried out, two of which use the Chimera model [18] as a point of comparison, since it is arguably a representative example of current deep-learning-based sound source separation techniques [34, 36, 35]. The evaluated Chimera network is a modified version to the one originally presented in [18], such that it was able to receive both outputs of the beamformer described in Section 3.1.

It is important to mention that the original version of the Chimera network was not built for generalized sound source separation. However, with slight modifications, such as the one proposed in this work, as well as more complex such as the ones shown in [34, 36, 35], its performance can be quite impressive.

To this effect, a similar evaluation to the one described in Section 3.2.2 (whose results are shown in Table 1) was carried out for the Chimera network, trained with up to 3 sources. Different configurations were evaluated, which varied in terms of NB and the embedding dimension used by one of the heads of the Chimera network (D). H and L were kept at 4 and 500, respectively, since these are the recommended values used in [18].

The results of these evaluations, carried out with up to 3 sources (including the source of interest; meaning, 2 interferences), are shown in Table 2.

Table 2 Evaluation of different configurations of the modified Chimera model with up to 3 sources 

Srcs. NB D Mem. (MB) SIR (dB)
3 8192 5 268 24.55
3 8192 10 283 25.59
3 8192 20 312 24.92
3 8192 40 371 25.29
3 16384 5 268 27.25
3 16384 10 283 27.12
3 16384 20 312 27.29
3 16384 40 371 27.28
Average 308.5 26.16

In this section some perspectives are provided that show the applicability of the proposed system. The results of three evaluations are reported:

  • — The relationship of the SIR performance against memory usage, for both Chimera and the proposed system.

  • — The relationship of the SIR performance against number of sources, for both Chimera and the proposed system.

  • — The robustness against changes in array geometry of the proposed system.

For these evaluations, 100 speakers were randomly chosen from the validation subset, and for each speaker 10 consecutive NB-length windows were selected for the 16384-input length models, and 20 of these were chosen for the 8192-input length models. Both of these types of segments are approximately 10 seconds. When varying the number of sources, these segments were mixed with the segments of other randomly selected speakers from the validation subset.

4.1 SIR vs Memory Usage

In Figure 4 each data point represents an architecture configuration shown in Tables 1 and 2; blue crosses belong to the proposed BLSTM-based models, and red circles to the Chimera-based models. The horizontal axis represents memory usage and the vertical axis its SIR. The blue dot-dashed lines represent the memory usage and SIR of the recommended configuration of the proposed BLSTM-based architecture; and the red dashed lines represent the memory usage and SIR of the similarly selected recommended configuration from the Chimera variations (shown in bold in Table 2).

Fig. 4 Memory Requirements vs SIR. The blue dot-dashed lines represent the respective SIR and memory usage of the recommended configuration BLSTM-based architecture, and red dashed lines the memory usage and SIR of the similarly selected recommended Chimera architecture. 

As it can be seen, although the difference between the SIR of Chimera and the proposed BLSTM-based architecture configuration is low ( 3 dB), the difference between their memory usage is substantial (> 200 MB).

4.2 SIR vs Number of Sources

It is also of interest to investigate the impact that the number of sources has on the performance. To this effect, we compare the performance of the recommended configuration of our proposed system (shown in bold in Table 1) as well as the best performing configuration of Chimera (underlined in Table 2), as the amount of sources is increased. The results are shown in Figure 5.

Fig. 5 Number of sources vs SIR of the output of the trained models 

It is important to note that both models were trained with up to 3 simultaneous sources, so these results reflect their ability to extrapolate the separation capabilities with more sources than they were trained with.

As it can be seen, both models have comparable SIR performance, and the obvious tendency is that as the number of sources increases, the SIR decreases (which is to be expected). An explanation for this is that the beamformer provides both an estimation of the source of interest, as well as an estimation of cumulative environmental interference from which the SOI should be separated.

This means that the permutation problem is solved from the beginning. Thus, the deep clustering part of Chimera that aims to solve this problem is rendered unnecessary for this test scenario.

4.3 SIR vs Number of Microphones

Since the models were trained using the output of the beamformer that was fed the simulated inputs of a two-microphone array, it is of interest to investigate the impact of the system if the number of microphones varies.

In Figure 6, the SIR performance is shown for both the recommended configuration of the BLSTM model as well as the best performing configuration of the Chimera when the number of microphones of the linear array is increased up to 10 microphones. No re-training was carried out and the same sources were used throughout the increase in number of microphones.

Fig. 6 Number of microphones vs SIR of the output of the trained models 

Additionally, to investigate the impact of changing the geometry of the simulated microphone array, the SIR performance as the number of sources increases when using a linear, triangle, square, pentagonal and hexagonal array is shown in Figure 7.

Fig. 7 Array geometry vs SIR of the output of the recommended BLSTM architecture configuration 

In both Figures 6 and 7, the same tendency observed in the previous section is still present: as the number of sources increases, the SIR decreases. More on topic, it can also be seen that, overall, as the number of microphones increases, so does the SIR. A possible explanation for this is that the quality of the beamformer output is affected by the number of microphones used, as shown in Figure 8.

Fig. 8 Number of microphones vs SIR of the beamformer output 

When comparing the SIR of the beamformer output (reported in Figure 8) and the SIR of the overall system output (reported in Figures 6 and 7), a substantial SIR increase can be observed in all the tested numbers of microphones, ranging from 10 to 20 dB difference in performance. This indicates that the BLSTM-based TF binary masking stage is essential in obtaining the reported performance.

More importantly, it is clear that the proposed system is quite robust against changes in the microphone geometry (being linear and the tested 2D geometries). In fact, in most cases, the SIR performance increases when more microphones are added, regardless of the employed geometry.

5 Results Discussion

It is important to point out that the recommended configuration of the proposed BLSTM model not only provides comparable SIR performance to the Chimera model, but in a considerable amount of cases, it actually outperformed it.

The reason this is important is that such a configuration only occupies nearly 10% of the amount of memory that the Chimera model occupies.

Moreover, in a considerable amount of cases both models provided a SIR close to or above the 20 dB mark, which can be considered as a high level of SIR for most auditory scene analysis [22].

Additionally, it can be seen in Table 1 that the proposed architectures configured with L = 3 obtain a higher SIR than their counterparts with L = 1 and L = 5, while keeping every other parameter the same. A possible explanation is that this number of BLSTM stacked layers may be a kind of “sweet spot” in the established solution space. However, this definitely merits further investigation.

It is also important to mention that the response time of all of the proposed architecture configurations is smaller than the length of the time window that it is fed. Meaning, all these architectures are able to carry out online sound source separation (although with up to a 1-second delay; 0.5-second delay, if using the other recommended configuration in italics in Table 1). Is is also worth considering that the computer used for these evaluations has an i7-4700MQ at 2.4 GHz (which is a moderate CPU by today’s standards), and no GPU was used to run the evaluated configurations.

This means that the proposed system provides a high separation performance (an average SIR higher than 20 dB), with moderate computational requirements.

6 Conclusion

There is a growing interest in online sound source separation in several areas of application. Deep learning techniques have reached an important level of performance, but require considerable computational resources. In this work, we propose a two step system that first carries out a preliminary estimation of both the source of interest and the cumulative environmental interference, via phase-based frequency masking. These two estimations are then fed to a BLSTM-based model that aims to estimate a time-frequency binary mask that, when applied to the signal of the reference microphone, provides a separation of the source of interest from the cumulative environmental interference.

The system was compared to a variation of the Chimera model, which applies deep clustering to solve the permutation problem encountered when carrying out sound source separation. It was shown that the proposed BLSTM-based system achieved comparable results and even in some cases even obtained slightly higher SIR results. And, it accomplished this only using nearly 10% of the memory occupied by the Chimera model in a moderately equipped computer. The reasoning behind this is that the first stage of the system (the phase-based beamformer) is solving the permutation problem from the beginning and, thus, the deep clustering parts of the Chimera model are not necessary to properly separate the source of interest.

The results shown here were all carried out with simulated data, with no noise and reverberation present. To this effect, for future work, we propose to investigate several methods of data augmentation that adds this effects to the data, to achieve acceptable SIR performance in real-life scenarios. We also propose to employ the AIRA corpus [26] to evaluate this next version of the proposed system.

And finally we will reduce the 1-second delay the system presents by a combination of low-grade GPUs (that still keep the computational requirements low) and shifting processing buffers.

Acknowledgments

This work was supported by CONACYT through the project 251319, and PAPIIT through the project IA100120. The authors would also like to thank David Kant, from UC Santa Cruz, for his insight during the initial development of the phase-based beamformer here described.

References

1.  1. Brutti, A., Tsiami, A., Katsamanis, A., & Maragos, P. (2016). A phase-based time-frequency masking for multi-channel speech enhancement in domestic environments. Interspeech 2016, pp. 2875–2879. [ Links ]

2.  2. Capon, J. (1969). High-resolution frequency-wavenumber spectrum analysis. Proceedings of the IEEE, Vol. 57, No. 8, pp. 1408–1418. [ Links ]

3.  3. Chen, Z., Luo, Y., & Mesgarani, N. (2017). Deep attractor network for single-microphone speaker separation. 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 246–250. [ Links ]

4.  4. Graves, A., Fernández, S., Gómez, F., & Schmidhuber, J. (2006). Connectionist temporal classification: Labelling unsegmented sequence data with recurrent neural networks. Proceedings of the 23rd International Conference on Machine Learning, ICML ’06, ACM, New York, NY, USA, pp. 369–376. [ Links ]

5.  5. Graves, A., Jaitly, N., & Mohamed, A. (2013). Hybrid speech recognition with deep bidirectional LSTM. 2013 IEEE Workshop on Automatic Speech Recognition and Understanding, pp. 273–278. [ Links ]

6.  6. Graves, A., Mohamed, A., & Hinton, G. (2013). Speech recognition with deep recurrent neural networks. 2013 IEEE International Conference on Acoustics, Speech and Signal Processing, pp. 6645–6649. [ Links ]

7.  7. Graves, A., & Schmidhuber, J. (2005). Framewise phoneme classification with bidirectional LSTM networks. Proceedings of IEEE International Joint Conference on Neural Networks, volume 4, pp. 2047–2052. [ Links ]

8.  8. Grondin, F. (2019). ODAS: Open embedded audition system. https://github.com/introlab/odas. Accessed: 2019-08-01. [ Links ]

9.  9. Grondin, F., Létourneau, D., Ferland, F., Rousseau, V., & Michaud, F. (2013). The ManyEars open framework. Autonomous Robots, Vol. 34, No. 3, pp. 217–232. [ Links ]

10.  10. Grondin, F., & Michaud, F. (2019). Lightweight and optimized sound source localization and tracking methods for open and closed microphone array configurations. Robotics and Autonomous Systems, Vol. 113, pp. 63–80. [ Links ]

11.  11. Hadad, E., Doclo, S., & Gannot, S. (2016). The binaural LCMV beamformer and its performance analysis. IEEE/ACM Transactions on Audio, Speech, and Language Processing, Vol. 24, No. 3, pp. 543–558. [ Links ]

12.  12. Harriott, C. E., & Adams, J. A. (2017). Towards reaction and response time metrics for real-world human-robot interaction. 2017 26th IEEE International Symposium on Robot and Human Interactive Communication (ROMAN), pp. 799–804. [ Links ]

13.  13. He, L., Zhou, Y., & Liu, H. (2019). Phase time-frequency masking based speech enhancement algorithm using circular microphone array. 2019 IEEE International Conference on Multimedia and Expo (ICME), pp. 808–813. [ Links ]

14.  14. Hershey, J. R., Chen, Z., Le Roux, J., & Watanabe, S. (2016). Deep clustering: Discriminative embeddings for segmentation and separation. 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 31–35. [ Links ]

15.  15. Huang, P., Kim, M., Hasegawa-Johnson, M., & Smaragdis, P. (2014). Deep learning for monaural speech separation. 2014 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 1562–1566. [ Links ]

16.  16. Kolen, J. F., & Kremer, S. C. (2001). Gradient Flow in Recurrent Nets: The Difficulty of Learning LongTerm Dependencies, chapter 14. IEEE, pp. 237–243. [ Links ]

17.  17. Levin, M. (1964). Maximum-likelihood array processing. Seismic Discrimination Semi-Annual Technical Summary Report. [ Links ]

18.  18. Luo, Y., Chen, Z., Hershey, J. R., Le Roux, J., & Mesgarani, N. (2017). Deep clustering and conventional networks for music separation: Stronger together. 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 61–65. [ Links ]

19.  19. Nakadai, K., Nakajima, H., Hasegawa, Y., & Tsujino, H. (2009). Sound source separation of moving speakers for robot audition. 2009 IEEE International Conference on Acoustics, Speech and Signal Processing, pp. 3685–3688. [ Links ]

20.  20. Nakadai, K., Okuno, H. G., & Mizumoto, T. (2017). Development, Deployment and Applications of Robot Audition Open Source Software HARK. Journal of Robotics and Mechatronics, Vol. 29, No. 1, pp. 16–25. [ Links ]

21.  21. Nakajima, H., Nakadai, K., Hasegawa, Y., & Tsujino, H. (2010). Blind source separation with parameter-free adaptive step-size method for robot audition. IEEE Transactions on Audio, Speech, and Language Processing, Vol. 18, No. 6, pp. 1476–1485. [ Links ]

22.  22. Okuno, H. G., & Nakadai, K. (2015). Robot audition: Its rise and perspectives. 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 5610–5614. [ Links ]

23.  23. Panayotov, V., Chen, G., Povey, D., & Khudanpur, S. (2015). Librispeech: An ASR corpus based on public domain audio books. 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 5206–5210. [ Links ]

24.  24. Pertilä, P., & Nikunen, J. (2015). Distant speech separation using predicted time–frequency masks from spatial features. Speech Communication, Vol. 68, pp. 97–106. [ Links ]

25.  25. Rascon, C., & Meza, I. (2017). Localization of sound sources in robotics: A review. Robotics and Autonomous Systems, Vol. 96, pp. 184–210. [ Links ]

26.  26. Rascon, C., Meza, I., Millan-González, A., Velez, I., Fuentes, G., Mendoza, D., & Ruiz-Espitia, O. (2018). Acoustic interactions for robot audition: A corpus of real auditory scenes. The Journal of the Acoustical Society of America, Vol. 144, No. 5. [ Links ]

27.  27. Ray, A., Rajeswar, S., & Chaudhury, S. (2015). Text recognition using deep BLSTM networks. 2015 Eighth International Conference on Advances in Pattern Recognition (ICAPR), pp. 1–6. [ Links ]

28.  28. Valin, J., Rouat, J., & Michaud, F. (2004). Enhanced robot audition based on microphone array source separation with post-filter. 2004 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) (IEEE Cat. No.04CH37566), volume 3, pp. 2123–2128 vol.3. [ Links ]

29.  29. Valin, J., Yamamoto, S., Rouat, J., Michaud, F., Nakadai, K., & Okuno, H. G. (2007). Robust recognition of simultaneous speech by a mobile robot. IEEE Transactions on Robotics, Vol. 23, No. 4, pp. 742–752. [ Links ]

30.  30. van de Sande, J. (2012). Real-time Beamforming and Sound Classification Parameter Generation in Public Environments. Master’s thesis, Delft University of Technology, Netherlands. [ Links ]

31.  31. Vincent, E., Gribonval, R., & Fevotte, C. (2006). Performance measurement in blind audio source separation. IEEE Transactions on Audio, Speech, and Language Processing, Vol. 14, No. 4, pp. 1462–1469. [ Links ]

32.  32. Wang, D. (2005). On ideal binary mask as the computational goal of auditory scene analysis. In Speech separation by humans and machines. Springer, pp. 181–197. [ Links ]

33.  33. Wang, D., & Chen, J. (2018). Supervised speech separation based on deep learning: An overview. IEEE/ACM Transactions on Audio, Speech, and Language Processing, Vol. 26, No. 10, pp. 1702– 1726. [ Links ]

34.  34. Wang, Z., Roux, J. L., & Hershey, J. R. (2018). Alternative objective functions for deep clustering. 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 686–690. [ Links ]

35.  35. Wang, Z., Tan, K., & Wang, D. (2019). Deep learning based phase reconstruction for speaker separation: A trigonometric perspective. ICASSP 2019 - 2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 71–75. [ Links ]

36.  36. Wang, Z., & Wang, D. (2019). Combining spectral and spatial features for deep learning based blind speaker separation. IEEE/ACM Transactions on Audio, Speech, and Language Processing, Vol. 27, No. 2, pp. 457–468. [ Links ]

37.  37. Yu, D., Kolbæk, M., Tan, Z., & Jensen, J. (2017). Permutation invariant training of deep models for speaker-independent multi-talker speech separation. 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 241–245. [ Links ]

2In this reference, ODAS does not report any source separation capabilities, but its authors have already added this functionality to its base code [8].

3We define “memory” as the amount of RAM (measured in MB) the model occupies when not carrying out any operations, as a representation of the computational resources it requires to run.

Received: June 14, 2020; Accepted: July 21, 2020

* Corresponding author: Caleb Rascón, e-mail: caleb.rascon@iimas.unam.mx

Creative Commons License This is an open-access article distributed under the terms of the Creative Commons Attribution License