Next Article in Journal
Bearing Fault Diagnosis of Induction Motors Using a Genetic Algorithm and Machine Learning Classifiers
Previous Article in Journal
Security Architecture for Defining and Enforcing Security Profiles in DLT/SDN-Based IoT Systems
Previous Article in Special Issue
End-to-End Automatic Pronunciation Error Detection Based on Improved Hybrid CTC/Attention Architecture
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Joint Optimization of Deep Neural Network-Based Dereverberation and Beamforming for Sound Event Detection in Multi-Channel Environments

Department of Electronics and Computer Engineering, Hanyang University, Seoul 04763, Korea
*
Author to whom correspondence should be addressed.
Sensors 2020, 20(7), 1883; https://doi.org/10.3390/s20071883
Submission received: 28 February 2020 / Revised: 21 March 2020 / Accepted: 25 March 2020 / Published: 28 March 2020

Abstract

:
In this paper, we propose joint optimization of deep neural network (DNN)-supported dereverberation and beamforming for the convolutional recurrent neural network (CRNN)-based sound event detection (SED) in multi-channel environments. First, the short-time Fourier transform (STFT) coefficients are calculated from multi-channel audio signals under the noisy and reverberant environments, which are then enhanced by the DNN-supported weighted prediction error (WPE) dereverberation with the estimated masks. Next, the STFT coefficients of the dereverberated multi-channel audio signals are conveyed to the DNN-supported minimum variance distortionless response (MVDR) beamformer in which DNN-supported MVDR beamforming is carried out with the source and noise masks estimated by the DNN. As a result, the single-channel enhanced STFT coefficients are shown at the output and tossed to the CRNN-based SED system, and then, the three modules are jointly trained by the single loss function designed for SED. Furthermore, to ease the difficulty of training a deep learning model for SED caused by the imbalance in the amount of data for each class, the focal loss is used as a loss function. Experimental results show that joint training of DNN-supported dereverberation and beamforming with the SED model under the supervision of focal loss significantly improves the performance under the noisy and reverberant environments.

1. Introduction

Sound event detection (SED) is desired as a task that detects the onset and offset times for each sound event in an audio segment. Various sounds always occur around us, and SED enables many services, including social care [1], audio surveillance [2,3], drone detection [4], and bird detection [5], by allowing machines to recognize sound events like the human auditory system. In recent years, the Detection and Classification of Acoustic Scenes and Events (DCASE) challenge and deep learning have significantly accelerated the research of SED. In the first DCASE challenge held in 2013, all proposed algorithms were based on shallow learning such as the hidden Markov model (HMM), support vector machine (SVM), and the Gaussian mixture model (GMM). In DCASE 2013, only a small number of teams participated, and the performances of the systems turned out not to be desirable [6]. Since the deep neural network (DNN)-based polyphonic SED algorithm was proposed in 2015 [7], deep learning-based SED studies have begun to pour out with the DCASE challenge (2016, 2017, 2018, 2019). In particular, deep learning structures based on the convolutional neural network (CNN) [8,9], recurrent neural network (RNN) [10,11,12], and convolutional recurrent neural network (CRNN) [13] showed the state-of-the-art performance, and data augmentation methods were proposed to maximize the learning ability of deep learning models [14]. Since the CNN can extract optimized features with trainable convolutional filters, it achieves a better performance than image-like inputs such as the log-scale Mel filter bank (LMFB), which is most used in the SED domain. Furthermore, the RNN, which can remember previous inputs through time, also performs well due to the time-series characteristics of the audio signal. Recently, the CNN and CRNN models with additional techniques, including the modified CNN [15] and pooling methods [16], were proposed. Furthermore, further studies combined with other tasks such as sound event detection and segmentation using the weakly labeled data [17] and joint sound event detection and localization [18] were proposed.
In contrast, in the speech recognition domain, the use of the LMFB as a feature vector and the CNN, RNN, and CRNN as the classifiers for the acoustic model are similar to the SED, but studies of integrating a preprocessor, such as acoustic beamforming or dereverberation with the acoustic model using multi-channel audio signals, have been actively conducted to improve recognition accuracy [19,20,21]. Furthermore, the joint optimization method on DNN-supported dereverberation and beamforming with an end-to-end speech recognition model was recently proposed [22]. Similarly, the joint optimization onto DNN-supported dereverberation and beamforming with the x-vector net was introduced in the speaker verification domain [23]. However, unlike speech recognition and speaker verification, there have not been many studies to enhance the audio signals for the SED because it is challenging to distinguish the evident audio from ambient noise due to the wide variety of target sounds. Sometimes, audio enhancement even degrades the SED performance, so some studies have been conducted on a limited basis with a weak level of noise reduction [24] or adaptive noise reduction [25] for the SED. Additionally, when using multi-channel audio signals, only studies using binaural features [26] or spatial features [27] and classification with the 3D CNN [28] have been reported, rather than combining them with preprocessing algorithms. Nevertheless, research on the combination of the preprocessor and the SED to take advantage of the multi-microphone has to be carried out for further performance enhancement.
In this paper, we propose a joint optimization method on DNN-supported dereverberation and beamforming for the SED under noisy and reverberant conditions. Because deep learning seamlessly optimizes beamforming and dereverberation through training, it is possible to combine beamforming and dereverberation with the SED to obtain optimized overall performance. One significant contribution compared to previous studies is the effectiveness of the cascade of DNN-supported weighted prediction error (WPE) dereverberation, the DNN-supported minimum variance distortionless response (MVDR) beamformer, and the SED network. Further, we present jointly training the final objective, the cost function, of the SED task. Furthermore, we employ the focal loss [29] within this task, since it is challenging to equalize the data amount of each sound class because the audio lengths of each class are all different in reality. Specifically, a mini-batch balancing method [30] in the training process is proposed to overcome this problem, but focal loss further helps to compensate this problem naturally in the training process. The evaluation was conducted based on the Tampere University (TAU) Spatial Sound Events 2019 dataset (http://dcase.community/challenge2019/task-sound-event-localization-and-detection), which showed significant improvement compared to conventional methods in the F-score and error rate.
Section 2 describes the proposed system, which is composed of the DNN-supported WPE dereverberation, DNN-supported MVDR beamformer, and SED. The dataset, evaluation metrics, experimental setup, and results are described in Section 3. Finally, conclusions are provided in Section 4.

2. Proposed System Overview

In this section, we fully describe our proposed system, which consists of three parts, as depicted in Figure 1. The first part of the system is designed for dereverberation by the DNN-supported WPE using multi-channel signals (in Section 2.1). In the second part, the DNN-supported MVDR beamforming is performed using the multi-channel output of the first part, and the single-channel beamformed signal is estimated as a result. Finally, the CRNN based SED assesses the presence or absence of sound events, including the onset and offset detection. Then, all the parts of the system are jointly optimized with the focal loss as a loss function. The details of each part of the proposed system are described in the following subsections.

2.1. DNN-Supported WPE Dereverberation

This subsection explains in detail the DNN-supported WPE dereverberation part including classical WPE dereverberation and DNN-supported WPE dereverberation reported in [20,21,22,23]. When we observe a signal using D microphones in a noisy and reverberant environment, the observed signal y t , f , d can be represented in the short-time Fourier transform (STFT) domain as follows:
y t , f , d = x t , f , d ( early ) + x t , f , d ( late ) + n t , f , d
where x t , f , d ( early ) , x t , f , d ( late ) , and n t , f , d denote the source signal convolved with the early part of the room impulse response (RIR) and with the late reflection and noise signal, respectively. Furthermore, t is the time frame index; f is the frequency bin index; and d is the microphone channel index, respectively. We assume that the first 50 ms after the main peak of the RIR contributes to the early reflection, and the remaining part becomes the late reflection. The purpose of dereverberation is to subtract late reflection components from the observed signal as follows:
x ^ t , f , d ( early ) = y t , f , d G f , d H y ˜ t Δ , f
where G f , d H , y ˜ t Δ , f , and Δ are the stacked representations of the linear prediction (LP) filter (WPE filter in Figure 1), the observation, and a delay for LP, respectively. To estimate the early reflection component, the classical WPE algorithm finds the LP filter based on the maximum likelihood (ML) for which the WPE assumes that the desired signal follows a zero-mean complex Gaussian distribution with a time-varying variance λ t , f . There is no closed-form solution of the ML optimization problem, but an iterative procedure alternates between estimating the filter coefficients G f , d H and the time-varying variance λ t , f to find G f , d H as follows:
λ t , f = 1 ( δ + 1 + δ ) D i = t δ t + δ d | x ^ t , f , d ( early ) | 2
R f = t y ˜ t Δ , f y ˜ t Δ , f H λ t , f C D K × D K
P f = t y ˜ t , f y t Δ , f H λ t , f C D K × D
G f = R f 1 P f C D K × D
where ( δ + 1 + δ ) means the number of context frames to improve the variance estimate, R f is the correlation matrix, P f is the correlation vector, and K is the order of the LP filter. DNN-supported WPE dereverberation replaces the iterative procedure to estimate power spectrum | x ^ t , f , d ( early ) | 2 in Equation (3). For this, we estimate the masks for calculating the desired power spectrum | x ^ t , f , d ( early ) | 2 from the given input power spectrum | y t , f , d | 2 . Specifically, the log-scale power spectra (LPS) y t , f , d are used as the input of the DNNs, which use ReLU with max clamp of one for the activation function of the output layer so as to limit the estimated masks within [ 0 , 1 ] . We can calculate | x ^ t , f , d ( early ) | 2 with | y t , f , d | 2 and estimated masks. Finally, x ^ t , f , d ( early ) is estimated by following sequence ( 3 ) ( 4 ) ( 5 ) ( 6 ) ( 2 ) , and it is tossed as the input of the DNN-supported MVDR beamformer part. Since the range of the masks is bounded within [ 0 , 1 ] , the DNN is easier to optimize than the direct prediction method of the desired power spectrum when jointly training the full networks [22].

2.2. DNN-Supported MVDR Beamformer

Originally, the MVDR beamformer used the steering vector, which depends on the angle of the desired signal from the source to minimize the residual noise while constraining the distortion of the signal. The steering vector can be obtained from an estimate of the direction of arrival (DoA) and the optimal signal is calculated by inducing the maximum beam gain in the steering vector direction and the minimum beam gain in the remaining direction. However, the MVDR beamformer also can be derived by speech and noise power spectral density (PSD) matrices without the steering vector. According to [31], the enhanced single-channel output x ^ t , f can be found by multiplying the gain H M V D R H (MVDR filter in Figure 1) by the observed multi-channel input signal y t , f as follows:
H M V D R H = Φ n n 1 Φ x x t r ( Φ n n 1 Φ x x ) u C D
x ^ t , f = H M V D R H y t , f
where Φ x x and Φ n n respectively denote the PSD matrices of the source and noise components and u is a one-hot vector for the reference microphone. In addition, t r means the trace of the matrix. In the DNN-supported MVDR beamformer, similar to the DNN-supported WPE dereverberation, two networks are separately trained for estimating masks in calculating the source and noise PSD matrices, where v denotes the signal attribute and θ f is a predefined decision threshold, respectively [19,20,22,23]. These masks are averaged over the microphone channel d. As a result, the PSD matrices of the source and noise are found as follows:
Φ v v = t M ^ t , f ( v ) y t , f y t , f H / t M ^ t , f ( v ) v { x , n }
where M ^ t , f ( v ) [ 0 , 1 ] denotes the estimated time-frequency mask calculated by the DNN, which uses the sigmoid as the activation function of the output layer. Finally, single-channel beamformed STFT coefficients x ^ t , f are estimated by following order ( 9 ) ( 7 ) ( 8 ) . Then, x ^ t , f is conveyed to the SED model for predicting sound events.

2.3. Sound Event Detection

For the SED, the LMFB is used as an input feature, which can be calculated by multiplying the magnitude spectrum with the Mel filter and then taking the logarithm. The input features are normalized using the global mean and variance statistics before being fed to the CRNN-based SED model, which is illustrated in Figure 2. Figure 2a–c show the CRNN-based SED model, the conventional convolutional block of the DCASE 2019 Task 3 baseline [18], and the proposed convolutional block, respectively. Unlike the conventional method using the three layers of the 3 × 3 convolution filter, the proposed convolutional block consists of two parallel parts inspired by VGGNet [32] and Inception V2 [33]. The first part conducts the convolution in the direction of the frequency axis only, and the second part performs the convolution in the direction of the time-frequency axis, then the two parts are concatenated. For the second part, the 3 × 3 convolution is divided into 1 × 3 convolution and 3 × 1 convolution. Finally, 1 × 1 convolution is used to reduce the computational cost. The output of the convolutional block is fed to the two layers of the bi-directional gated recurrent unit (GRU) RNN. Next, the output of the bi-directional GRU is connected to the fully connected layers and the output layer with the sigmoid function as an activation function, so that the value of the outputs is selected between zero and one for each class.

2.4. Joint Optimization

This section summarizes and explains how the DNN-supported dereverberation, beamforming, and the CRNN-based SED models are organized into a cascaded network. First, when the D-channel audio signal is input, the magnitudes of the STFT coefficients are calculated and then fed to the DNN. The DNN estimates the dereverberation mask, and then the magnitudes of the STFT coefficients of the dereverberated signal can be calculated using Equations (2), (3) and (6). Next, this output is fed into another DNN to estimate the source and noise masks for the neural MVDR beamformer. Using Equations (7)–(9), the magnitudes of the STFT coefficient of the single-channel enhanced signal are obtained. Then, multiplying these values with the Mel filters, the LMFB is calculated, which serves as an input for the CRNN-based SED model. The whole network is trained by the loss, which is calculated with the label and the SED output. At this time, the focal loss is considered as a loss function for further improving the performance. As for the SED, equalizing the data amount of each class is challenging because the audio lengths of each class are all different. The focal loss is useful for compensating for this problem naturally when training the deep learning model by giving a stronger loss to those that fail to estimate [29]. The focal loss is defined as follows:
FL ( p t ) = ( 1 p t ) γ log ( p t )
p t = p , if g t = 1 1 p , otherwise
where g t represents the ground truth, p [ 0 , 1 ] is the model’s estimated probability, and γ denotes the tunable focusing parameter.
All of the processes described above are differentiable, so the backpropagation with the chain rule is possible. Motivated by this, in the end, we perform joint training for the cascaded architecture of DNN-supported WPE dereverberation, the DNN-supported MVDR beamformer, and the SED network according to the focal loss, as depicted in Figure 2.
As for the joint optimization, we demand complex-valued operations including the complex-valued inverse in Equations (6) and (7). As in [19,23], the complex-valued operations using real-valued operations are implemented by separately computing real and imaginary parts. When C is a complex-valued matrix and A and B are real-valued matrices corresponding to real and imaginary parts, C can be expressed as C = A + i B . At this time, the complex-valued matrix inverse operations can be calculated as follows [34]:
( C 1 ) = ( A + BA 1 B ) 1
( C 1 ) = ( A + BA 1 B ) 1 BA 1

3. Experiments and Results

3.1. Dataset

The proposed algorithm was evaluated with the TAU Spatial Sound Events 2019 dataset. The dataset consists of the two datasets, Ambisonic and Microphone Array [35]. The TAU Spatial Sound Events 2019-Ambisonic dataset provides four-channel first-order ambisonic (FOA) recordings, while the TAU Spatial Sound Events 2019-Microphone Array dataset provides four-channel directional microphone recordings from a tetrahedral array configuration. Each dataset consists of 500 audio files, 400 for development and 100 for evaluation. The records are one minute long, the sampling frequency 48,000 Hz, and the signal-to-noise ratio (SNR) for sound events and ambient noise 30 dB. These recordings were synthesized using the spatial room impulse response (IRs) collected from five indoor locations at 504 unique combinations of azimuth-elevation-distance. The collected IRs were convolved with the DCASE 2016 Task 2 dataset. In the DCASE 2016 Task 2 dataset, there are 11 classes of sound events such as clearing throat, coughing, door knock, door slam, drawer, laughter, keyboard, keys (putting on table), page-turning, phone ringing, and speech, and each class consists of 20 audio files. Finally, each development dataset was divided into four cross-validation [36] splits of 100 recordings each. Additionally, to consider the noisy environment, we mixed the datasets with the ambient noise recorded at an indoor location inside the Hanyang University campus in Seoul, Korea, under 10 dB SNR. Two simple data augmentation methods (pitch shifting [14] and block mixing [10] using monophonic audio clips) were applied in the training process for the model generalization to reduce overfitting.

3.2. Evaluation Metrics

To evaluate the performance of the SED model, we measured the segment-based F-score and error rate (ER) in the same way as the DCASE 2019 Task 3. The F-score and ER were calculated in segments of one second with no overlap [37,38]. Therefore, the labels and the SED outputs were generated on average for segments of one second to calculate metrics. First, the F-score, which measures the effectiveness of retrieval, is calculated as follows:
F = 2 · k = 1 K T P 2 · k = 1 K T P + 2 · k = 1 K F P + 2 · k = 1 K F N
where K is the number of segments and T P ( k ) denotes the number of true positives, which is the total number of sound event classes that were active in both the reference and predictions for the segment. In addition, F P ( k ) denotes the number of false positives, which is the number of sound event classes that were active in the prediction, but were inactive in the reference. Similarly, F N ( k ) is the number of false negatives, which is the number of sound event classes inactive in the predictions, but active in the reference. Additionally, the ER, which measures the amount of errors, is given as follows:
E R = k = 1 K S ( k ) + k = 1 K D ( k ) + k = 1 K I ( k ) k = 1 K N ( k )
where N ( k ) is the total number of active sound event classes in the reference. In addition, S ( k ) , D ( k ) , and I ( k ) are called the substitution, deletion, and insertion, respectively, which are mathematically defined as:
S ( k ) = min ( F N ( k ) , F P ( k ) ) ,
D ( k ) = max ( 0 , F N ( k ) F P ( k ) ) ,
I ( k ) = max ( 0 , F P ( k ) F N ( k ) ) .
As for the ideal case, it is noted that the F-score and ER become one and zero, respectively.

3.3. Experimental Setup

The evaluation was performed with a window length of 40 ms, a hop length of 20 ms, and a fast Fourier transform (FFT) size of 2048 points. Therefore, we obtained 3000 frames in one file since the file was 60 seconds long, and the input sequence length T for training was 128. For dereverberation and beamforming, the multi-layer perceptron, which consisted of three hidden layers with 1024 nodes, was used. ReLU was chosen for the activation function at the hidden layers. For the DNN input, we used the log-scale power spectra (folded frequency bins were discarded) as features that were spliced with three left and three right context frames. Note that the parameters of the LP filter for the WPE were fixed to ( Δ , K ) = ( 3 , 10 ) . For sound event detection, first, the number of Mel filters for LMFB C was 240. Next, the number of CNN filters for each layer was [64, 64, 64], and the max pooling sizes along the frequency axis ( M P 1 , M P 2 , and M P 3 ) were 6, 5, and 4, respectively. Additionally, the size of two GRU layers and two fully connected (FC) layers was [128, 128] and [256, 256], and the drop-out rate for the FC layers was 0.5. We summarize the configurations of the neural networks in Table 1 and Table 2. The batch size was 16, and an early stopping method was applied. Batch normalization [39] was applied to all networks, and the networks were optimized by Adam [40]. The focus parameter γ of the focal loss was set to two.

3.4. Results

Table 3 and Table 4 show the results with the TAU Spatial Sound Events 2019-Ambisonic development dataset and TAU Spatial Sound Events 2019-Microphone Array development dataset, respectively. First, by replacing the convolutional block, the F-score increased by approximately 1.6% on average compared to the conventional method in both datasets, and the ER also improved to 0.05. This result exhibited that using the different types of blocks in the convolutional block to extract features and concatenate them also worked well for the SED. Next, the performance was improved in all cases where the WPE was combined with the SED, the MVDR was combined with the SED, and the WPE and MVDR were connected with the SED and then jointly trained, respectively. The one point of these results was that MVDR was much more useful than WPE. However, this may be because the reverberation of the dataset was not active. Finally, the focal loss also turned out to be helpful in gaining the performances for the unbalanced dataset. The performance of Split 2, which had a slightly lower performance than the other splits, was relatively increased. Subsequently, the average F-score increased by 13.1%, and the ER improved 0.23 compared to the conventional method. For the DCASE 2019 Task 3 challenge results, two systems showed better performance than our proposed system with this dataset, and they achieved the F-score of 98.2%, while Xue_JDAI_task3_1 [41] achieved the F-score of 93.4%. However, MazzonYasuda_NTT_task3_3 [42] used 134M parameters for a vast ensemble model because the DCASE 2019 challenge did not require limited complexity. In contrast, the number of parameters in our system was 21M only. Table 5 and Table 6 show the results at 10 dB SNR for the Ambisonic and Microphone Array development datasets, respectively. Similar to the original 30 dB datasets, the performance in the noisy environment was also improved in all cases where the WPE was combined with the SED, the MVDR was combined with the SED, and the WPE and MVDR were attached to the SED and then jointly trained, respectively. Table 7 shows the F-score and ER results of the evaluation dataset. Compared to the DCASE 2019 Task 3 algorithms, the proposed algorithm showed 4% better performance under the 10 dB SNR environment.

4. Conclusions

The CRNN-based SED model, which combines the DNN-supported WPE dereverberation and the DNN-supported MVDR beamformer, was jointly trained using a single loss function. Since the DNN-supported WPE dereverberation and MVDR beamformer were all differentiable, the gradients derived from the SED part could be backpropagated to update all the parameters of the DNN-supported dereverberation and beamforming. As for the loss function, we used the focal loss to compensate for the imbalance in the amount of data between classes. Experimental results showed that the joint training and focal loss improved the F-score and error rate of the SED, especially noisy environments.

Author Contributions

Conceptualization, K.N. and J.-H.C.; methodology, K.N. and J.-H.C.; software, K.N.; validation, K.N.; writing, original draft preparation, K.N.; writing, review and editing, J.-H.C.; supervision, J.-H.C.; project administration, J.-H.C.; funding acquisition, J.-H.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by Projects for Research and Development of Police science and Technology under Center for Research and Development of Police science and Technology and Korean National Police Agency funded by the Ministry of Science and ICT (PA-J000001-2017-101).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Goetze, S.; Schroder, J.; Gerlach, S.; Hollosi, D.; Appell, J.E.; Wallhoff, F. Acoustic monitoring and localization for social care. J. Comput. Sci. Eng. 2012, 6, 40–50. [Google Scholar] [CrossRef] [Green Version]
  2. Crocco, M.; Cristani, M.; Trucco, A.; Murino, V. Audio surveillance: A systematic review. ACM Comput. Surv. 2016, 48, 52. [Google Scholar] [CrossRef]
  3. Alsina-Pagès, R.; Navarro, J.; Alías, F.; Hervás, M. homesound: Real-time audio event detection based on high performance computing for behaviour and surveillance remote monitoring. Sensors 2017, 17, 854. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Anwar, M.Z.; Kaleem, Z.; Jamalipour, A. Machine learning inspired sound-based amateur drone detection for public safety applications. IEEE Trans. Veh. Technol. 2019, 68, 2526–2534. [Google Scholar] [CrossRef]
  5. Stowell, D.; Wood, M.; Stylianou, Y.; Glotin, H. Bird detection in audio: A survey and a challenge. In Proceedings of the IEEE International Workshop on Machine Learning for Signal Processing (MLSP), Pittsburgh, PA, USA, 13–16 October 2016; pp. 1–6. [Google Scholar]
  6. Giannoulis, D.; Benetos, E.; Stowell, D.; Rossignol, M.; Lagrange, M.; Plumbley, M.D. Detection and classification of acoustic scenes and events: An IEEE AASP challenge. In Proceedings of the IEEE Workshop on Applications of Signal Processing to Audio and Acoustics (WASPAA), New Paltz, NY, USA, 20–23 October 2013; pp. 1–4. [Google Scholar]
  7. Cakir, E.; Heittola, T.; Huttunen, H.; Virtanen, T. Polyphonic sound event detection using multi label deep neural networks. In Proceedings of the International Joint Conference on Neural Networks (IJCNN), Budapest, Hungary, 14–19 July 2015; pp. 1–7. [Google Scholar]
  8. Zhang, H.; McLoughlin, I.; Song, Y. Robust sound event recognition using convolutional neural networks. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brisbane, Australia, 19–24 April 2015; pp. 559–563. [Google Scholar]
  9. Phan, H.; Hertel, L.; Maass, M.; Mertins, A. Robust audio event recognition with 1-max pooling convolutional neural networks. In Proceedings of the Interspeech, San Francisco, CA, USA, 8–12 September 2016; pp. 3653–3657. [Google Scholar]
  10. Parascandolo, G.; Huttunen, H.; Virtanen, T. Recurrent neural networks for polyphonic sound event detection in real life recordings. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Shanghai, China, 20–25 March 2016; pp. 6440–6444. [Google Scholar]
  11. Lu, R.; Duan, Z. Bidirectional GRU for Sound Event Detection. Available online: http://www.cs.tut.fi/sgn/arg/dcase2017/documents/challenge_technical_reports/DCASE2017_Lu_137.pdf (accessed on 28 March 2020).
  12. Hayashi, T.; Watanabe, S.; Toda, T.; Hori, T.; Le Roux, J.; Takeda, K. Duration-controlled LSTM for polyphonic sound event detection. IEEE/ACM Trans. Audio Speech Lang. Process. 2017, 25, 2059–2070. [Google Scholar] [CrossRef]
  13. Cakır, E.; Parascandolo, G.; Heittola, T.; Huttunen, H.; Virtanen, T. Convolutional recurrent neural networks for polyphonic sound event detection. IEEE/ACM Trans. Audio Speech Lang. Process. 2017, 25, 1291–1303. [Google Scholar] [CrossRef] [Green Version]
  14. Salamon, J.; Bello, J.P. Deep convolutional neural networks and data augmentation for environmental sound classification. IEEE Signal Process Lett. 2017, 24, 279–283. [Google Scholar] [CrossRef]
  15. Xu, Y.; Kong, Q.; Wang, W.; Plumbley, M.D. Large-scale weakly supervised audio classification using gated convolutional neural network. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Calgary, AB, Canada, 15–20 April 2018; pp. 6440–6444. [Google Scholar]
  16. He, K.X.; Shen, Y.H.; Zhang, W.Q. Hierarchical Pooling Structure for Weakly Labeled Sound Event Detection. Available online: https://arxiv.org/pdf/1903.11791.pdf (accessed on 28 March 2020).
  17. Kong, Q.; Xu, Y.; Sobieraj, I.; Wang, W.; Plumbley, M.D. Sound event detection and time–frequency segmentation from weakly labelled data. IEEE/ACM Trans. Audio Speech Lang. Process. 2019, 27, 777–787. [Google Scholar] [CrossRef]
  18. Adavanne, S.; Politis, A.; Nikunen, J.; Virtanen, T. Sound event localization and detection of overlapping sources using convolutional recurrent neural networks. IEEE J. Sel. Top. Sign. Proces. 2019, 13, 34–48. [Google Scholar] [CrossRef] [Green Version]
  19. Ochiai, T.; Watanabe, S.; Hori, T.; Hershey, J.R.; Xiao, X. Unified architecture for multichannel end-to-end speech recognition with neural beamforming. IEEE J. Sel. Top. Sign. Process. 2017, 11, 1274–1288. [Google Scholar] [CrossRef]
  20. Drude, L.; Boeddeker, C.; Heymann, J.; Haeb-Umbach, R.; Kinoshita, K.; Delcroix, M.; Nakatani, T. Integrating neural network based beamforming and weighted prediction error dereverberation. In Proceedings of the Interspeech, Hyderabad, India, 2–6 September 2018; pp. 3043–3047. [Google Scholar]
  21. Heymann, J.; Drude, L.; Haeb-Umbach, R.; Kinoshita, K.; Nakatani, T. Joint optimization of neural network-based WPE dereverberation and acoustic model for robust online ASR. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brighton, UK, 12–17 May 2019; pp. 6655–6659. [Google Scholar]
  22. Subramanian, A.S.; Wang, X.; Watanabe, S.; Taniguchi, T.; Tran, D.; Fujita, Y. An Investigation of End-to-End Multichannel Speech Recognition for Reverberant and Mismatch Conditions. Available online: https://arxiv.org/pdf/1904.09049.pdf (accessed on 28 March 2020).
  23. Yang, J.Y.; Chang, J.H. Joint optimization of neural acoustic beamforming and dereverberation with x-vectors for robust speaker verification. In Proceedings of the Interspeech, Graz, Austria, 15–19 September 2019; pp. 4075–4079. [Google Scholar]
  24. Choi, I.; Kwon, K.; Bae, S.H.; Kim, N.S. DNN-based sound event detection with exemplar-based approach for noise reduction. In Proceedings of the Workshop on Detection and Classification of Acoustic Scenes and Events (DCASE2016 Workshop), Budapest, Hungary, 3 September 2016; pp. 16–19. [Google Scholar]
  25. Zhou, Q.; Feng, Z.; Benetos, E. Adaptive noise reduction for sound event detection using subband-weighted NMF. Sensors 2019, 19, 3206. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  26. Adavanne, S.; Virtanen, T. A Report on Sound Event Detection with Different Binaural Features. Available online: http://www.cs.tut.fi/sgn/arg/dcase2017/documents/challenge_technical_reports/DCASE2017_Adavanne_130.pdf (accessed on 28 March 2020).
  27. Adavanne, S.; Pertilä, P.; Virtanen, T. Sound event detection using spatial features and convolutional recurrent neural network. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), New Orleans, LA, USA, 5–9 March 2017; pp. 771–775. [Google Scholar]
  28. Adavanne, S.; Politis, A.; Virtanen, T. Multichannel sound event detection using 3D convolutional neural networks for learning inter-channel features. In Proceedings of the International Joint Conference on Neural Networks (IJCNN), Rio, Brazil, 8–13 July 2018; pp. 1–7. [Google Scholar]
  29. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal loss for dense object detection. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
  30. Kong, Q.; Xu, Y.; Wang, W.; Plumbley, M.D. Audio set classification with attention model: A probabilistic perspective. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Calgary, AB, Canada, 15–20 April 2018; pp. 316–320. [Google Scholar]
  31. Souden, M.; Benesty, J.; Affes, S. On optimal frequency-domain multichannel linear filtering for noise reduction. IEEE Trans. Audio Speech Lang. Process. 2009, 18, 260–276. [Google Scholar] [CrossRef]
  32. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. Available online: https://arxiv.org/pdf/1409.1556.pdf (accessed on 28 March 2020).
  33. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the inception architecture for computer vision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 26–30 June 2016; pp. 2818–2826. [Google Scholar]
  34. Petersen, K.B.; Pedersen, M.S. The Matrix Cookbook; Techical University of Denmark: Kongens Lyngby, Denmark, 2008; Volume 7, p. 15. [Google Scholar]
  35. Adavanne, S.; Politis, A.; Virtanen, T. A Multi-Room Reverberant Dataset for Sound Event Localization and Detection. Available online: https://arxiv.org/pdf/1905.08546.pdf (accessed on 28 March 2020).
  36. Forman, G.; Scholz, M. Apples-to-apples in cross-validation studies: Pitfalls in classifier performance measurement. ACM SIGKDD Explorations Newsletter 2010, 12, 49–57. [Google Scholar] [CrossRef]
  37. Mesaros, A.; Heittola, T.; Virtanen, T. Metrics for polyphonic sound event detection. Appl. Sci. 2016, 6, 162. [Google Scholar] [CrossRef]
  38. Mesaros, A.; Heittola, T.; Ellis, D. Datasets and evaluation. In Computational Analysis of Sound Scenes and Events; Virtanen, T., Plumbley, M.D., Ellis, D., Eds.; Springer: Cham, Switzerland, 2018; pp. 147–179. [Google Scholar]
  39. Ioffe, S.; Szegedy, C. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift. Available online: https://arxiv.org/pdf/1502.03167.pdf (accessed on 28 March 2020).
  40. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. Available online: https://arxiv.org/pdf/1412.6980.pdf (accessed on 28 March 2020).
  41. Xue, W.; Tong, Y.; Zhang, C.; Ding, G. Multi-Beam and Multi-Task Learning for Joint Sound Event Detection and Localization. Available online: http://dcase.community/documents/challenge2019/technical_reports/DCASE2019_Xue_91.pdf (accessed on 28 March 2020).
  42. Mazzon, L.; Yasuda, M.; Koizumi, Y.; Harada, N. Sound Event Localization and Detection Using FOA Domain Spatial Augmentation. Available online: http://dcase.community/documents/challenge2019/technical_reports/DCASE2019_MazzonYasuda_93.pdf (accessed on 28 March 2020).
  43. Kapka, S.; Lewandowski, M. Sound Source Detection, Localization and Classification Using Consecutive Ensemble of CRNN Models. Available online: http://dcase.community/documents/challenge2019/technical_reports/DCASE2019_Kapka_26.pdf (accessed on 28 March 2020).
  44. Cao, Y.; Iqbal, T.; Kong, Q.; Galindo, M.; Wang, W.; Plumbley, M. Two-Stage Sound Event Localization and Detection Using Intensity Vector and Generalized Cross-Correlation. Available online: http://dcase.community/documents/challenge2019/technical_reports/DCASE2019_Cao_74.pdf (accessed on 28 March 2020).
Figure 1. Block diagram of the proposed system. WPE, weighted prediction error; MVDR, minimum variance distortionless response; SED, sound event detection.
Figure 1. Block diagram of the proposed system. WPE, weighted prediction error; MVDR, minimum variance distortionless response; SED, sound event detection.
Sensors 20 01883 g001
Figure 2. Overview of the SED model: (a) convolutional block of the Detection and Classification of Acoustic Scenes and Events (DCASE) 2019 Task 3 baseline [18]; (b) CRNN-based SED model; (c) proposed convolutional block. LMFB, log-scale Mel filter bank.
Figure 2. Overview of the SED model: (a) convolutional block of the Detection and Classification of Acoustic Scenes and Events (DCASE) 2019 Task 3 baseline [18]; (b) CRNN-based SED model; (c) proposed convolutional block. LMFB, log-scale Mel filter bank.
Sensors 20 01883 g002
Table 1. Configuration of DNNs for dereverberation and the beamformer.
Table 1. Configuration of DNNs for dereverberation and the beamformer.
LayersOutput size
(Channels × Frames × Frequency Bins)
Input
(Log power spectra)
4 × 128 × (1025 × 7)
(Including left and right context frames)
Hidden Layer 14 × 128 × 1024
Hidden Layer 24 × 128 × 1024
Hidden Layer 34 × 128 × 1024
Output
(Mask)
4 × 128 × 1025
Table 2. Configuration of CRNN.
Table 2. Configuration of CRNN.
LayersOutput Size
Input
(Log Mel filter bank)
1 × 128 × 240
(Feature maps × Frames × Mel bins)
Convolutional Layer 164 × 128 × 40/64 × 128 × 40
Convolutional Layer 264 × 128 × 8/64 × 128 × 8
Convolutional Layer 364 × 128 × 2/64 × 128 × 2
Concatenate128 × 128 × 2
1 × 1 Convolution64 × 128 × 2
GRU Layer 1128 × 128
GRU Layer 2128 × 128
Fully Connected Layer 1128 × 256
Fully Connected Layer 2128 × 256
Output128 × 11
(frames × classes)
Table 3. F-score and error rate (ER) results on the TAU Spatial Sound Events 2019-Ambisonic development dataset.
Table 3. F-score and error rate (ER) results on the TAU Spatial Sound Events 2019-Ambisonic development dataset.
DCASE 2019
Task 3
Baseline [18]
Proposed
SED
Proposed
WPE
+SED
Proposed
MVDR
+SED
Proposed
WPE+MVDR
+SED
Proposed
WPE+MVDR
+SED
+FL
F-score
(%)
Split 181.282.883.089.691.792.5
Split 278.080.181.288.190.892.0
Split 380.581.282.589.591.593.0
Split 479.880.882.089.491.293.5
Overall79.981.282.289.291.392.8
Error
rate
Split 10.310.280.270.150.140.12
Split 20.370.320.310.170.160.14
Split 30.330.300.280.150.150.13
Split 40.340.310.280.160.150.12
Overall0.340.300.290.160.150.13
Table 4. F-score and ER results on the TAU Spatial Sound Events 2019-Microphone Array development dataset.
Table 4. F-score and ER results on the TAU Spatial Sound Events 2019-Microphone Array development dataset.
DCASE 2019
Task 3
Baseline [18]
Proposed
SED
Proposed
WPE
+SED
Proposed
MVDR
+SED
Proposed
WPE+MVDR
+SED
Proposed
WPE+MVDR
+SED
+FL
F-score
(%)
Split 181.582.983.890.592.393.6
Split 279.181.582.589.991.593.5
Split 380.582.183.090.392.193.7
Split 479.881.583.390.191.993.1
Overall80.282.083.290.292.093.5
Error
rate
Split 10.310.280.250.150.140.08
Split 20.370.300.270.170.150.10
Split 30.350.290.270.160.140.09
Split 40.330.300.280.180.130.09
Overall0.340.290.270.170.140.09
Table 5. F-score and ER results at 10 dB SNR: Ambisonic development dataset.
Table 5. F-score and ER results at 10 dB SNR: Ambisonic development dataset.
DCASE 2019
Task 3
Baseline [18]
Proposed
SED
Proposed
WPE
+SED
Proposed
MVDR
+SED
Proposed
WPE+MVDR
+SED
Proposed
WPE+MVDR
+SED
+FL
F-score
(%)
Split 169.470.372.079.482.082.4
Split 268.271.572.479.081.281.1
Split 370.272.674.480.483.383.2
Split 469.673.573.180.282.483.0
Overall69.472.073.079.882.282.4
Error
rate
Split 10.510.480.460.350.280.27
Split 20.530.480.450.330.290.29
Split 30.500.470.460.340.280.26
Split 40.500.460.460.350.270.25
Overall0.510.470.460.340.280.27
Table 6. F-score and ER results at 10 dB SNR: Microphone Array development dataset.
Table 6. F-score and ER results at 10 dB SNR: Microphone Array development dataset.
DCASE 2019
Task 3
Baseline [18]
Proposed
SED
Proposed
WPE
+SED
Proposed
MVDR
+SED
Proposed
WPE+MVDR
+SED
Proposed
WPE+MVDR
+SED
+FL
F-score
(%)
Split 170.171.073.178.581.582.9
Split 270.671.173.479.080.280.8
Split 370.471.474.280.682.482.6
Split 470.572.474.879.683.483.2
Overall70.471.573.979.481.982.4
Error
rate
Split 10.480.480.450.340.290.26
Split 20.510.490.470.370.330.30
Split 30.470.450.460.330.290.27
Split 40.460.450.430.320.260.24
Overall0.480.470.450.340.290.26
Table 7. F-score and ER results: evaluation dataset.
Table 7. F-score and ER results: evaluation dataset.
30 dB SNR10 dB SNR
AlgorithmsF-Score (%)ERF-Score (%)ER
Kapka_SRPOL_task3_2 [43]94.70.0881.00.31
Cao_Surrey_task3_4 [44]95.50.0879.40.32
DCASE 2019 Task 3 baseline [18]85.40.2873.20.47
Proposed WPE+MVDR+SED+FL93.30.1185.30.25

Share and Cite

MDPI and ACS Style

Noh, K.; Chang, J.-H. Joint Optimization of Deep Neural Network-Based Dereverberation and Beamforming for Sound Event Detection in Multi-Channel Environments. Sensors 2020, 20, 1883. https://doi.org/10.3390/s20071883

AMA Style

Noh K, Chang J-H. Joint Optimization of Deep Neural Network-Based Dereverberation and Beamforming for Sound Event Detection in Multi-Channel Environments. Sensors. 2020; 20(7):1883. https://doi.org/10.3390/s20071883

Chicago/Turabian Style

Noh, Kyoungjin, and Joon-Hyuk Chang. 2020. "Joint Optimization of Deep Neural Network-Based Dereverberation and Beamforming for Sound Event Detection in Multi-Channel Environments" Sensors 20, no. 7: 1883. https://doi.org/10.3390/s20071883

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop