Next Article in Journal
Imitative Reinforcement Learning Fusing Mask R-CNN Perception Algorithms
Next Article in Special Issue
A Full Loading-Based MVDR Beamforming Method by Backward Correction of the Steering Vector and Reconstruction of the Covariance Matrix
Previous Article in Journal
CORSIA Lower Carbon Aviation Fuels: An Assessment of the Greenhouse Gas Emission Reduction Potential
Previous Article in Special Issue
Packet Loss Concealment Based on Phase Correction and Deep Neural Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Effective Dereverberation with a Lower Complexity at Presence of the Noise

Speech and Audio Signal Processing Laboratory, Faculty of Information Technology, Beijing University of Technology, Beijing 100124, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2022, 12(22), 11819; https://doi.org/10.3390/app122211819
Submission received: 4 November 2022 / Revised: 17 November 2022 / Accepted: 18 November 2022 / Published: 21 November 2022
(This article belongs to the Special Issue Advances in Speech and Language Processing)

Abstract

:
Adaptive beamforming and deconvolution techniques have shown effectiveness for reducing noise and reverberation. The minimum variance distortionless response (MVDR) beamformer is the most widely used for adaptive beamforming, whereas multichannel linear prediction (MCLP) is an excellent approach for the deconvolution. How to solve the problem where the noise and reverberation occur together is a challenging task. In this paper, the MVDR beamformer and MCLP are effectively combined for noise reduction and dereverberation. Especially, the MCLP coefficients are estimated by the Kalman filter and the MVDR filter based on the complex Gaussian mixture model (CGMM) is used to enhance the speech corrupted by the reverberation with the noise and to estimate the power spectral density (PSD) of the target speech required by the Kalman filter, respectively. The final enhanced speech is obtained by the Kalman filter. Furthermore, a complexity reduction method with respect to the Kalman filter is also proposed based on the Kronecker product. Compared to two advanced algorithms, the integrated sidelobe cancellation and linear prediction (ISCLP) method and the weighted prediction error (WPE) method, which are very effective for removing reverberation, the proposed algorithm shows better performance and lower complexity.

1. Introduction

In real applications, reverberation and noise often cause the performance of speech communication or human–computer interaction to be degraded. Especially, due to the presence of reverberation, the speech signal [1] is often distorted, which seriously affects the localization of the sound sources. With the rapid development of artificial intelligence, human–computer interaction becomes increasingly more important [2,3]. If the reverberation and noise are not well cancelled in front of the intelligent devices, then speech communication will be greatly impaired.
In recent years, microphone array-based methods using spatial information have received extensive attention. The microphone array-based methods, such as the MVDR-based beamformer [4], the linear-constrained minimum variance (LCMV) beamformer [5,6], and the generalized sidelobe canceller (GSC) [7], are often used for speech enhancement. The microphone array-based beamforming methods can be used to receive signals from the direction of the target speaker and suppress the interference, such as reverberation and noise [8,9,10]. In [8], the GSC and a post-filter implemented with the Wiener filter were introduced to eliminate the reverberation and noise. In [9], the LCMV was used for the dereverberation in the case of multiple speech sources, in which the obtained sparse matrix was used to obtain a preliminary estimation of the desired signal and the LCMV was used to obtain the final estimation of the desired signal. In [10], a beamforming estimator based on minimum mean square error (MMSE) was proposed to obtain the early components of the reverberation signal. In MVDR beamforming, a relatively early transfer function (RETF) was used. The beamforming and post-filter were used to manage the late reverberation and noise. The methods that only use beamforming to reduce reverberation have limited performance and cannot suppress the reverberation very well.
Currently, the most effective dereverberation method is based on multichannel linear prediction (MCLP) deconvolution [11,12,13,14,15,16], which is also known as the direct inverse filtering algorithm. The MCLP-based methods model the reverberation components as a delay of the speech signal in the time domain, namely the linear prediction components. In these methods, the estimation of the room impulse response (RIR) is not required, and the reverberation is subtracted from the speech signal itself. Thus, this method can be applied to the situation where the RIR cannot be reliably estimated. In [13], a statistical model-based dereverberation method was proposed. It assumes that the speech source follows the time-varying Gaussian model. The reverberation is modeled as the delayed components after linear prediction, where linear prediction coefficients are approximated by estimating model parameters to obtain the late reverberation components. This modeling method is defined as the weighted prediction error (WPE). In [11], by combining WPE and MVDR, the noise and reverberation were well suppressed simultaneously. The methods of using Kalman filter to estimate linear prediction coefficients for obtaining reverberation components were also proposed [12,14,15]. In [14], the MVDR and the MCLP were used for simultaneous noise reduction and dereverberation, and the Kalman filter was used to estimate the beamforming coefficients and linear prediction coefficients simultaneously. This kind of dereverberation framework that combines beamforming and the MCLP has attracted increasingly more attention. In [15], two interacting Kalman filters were used to alternately estimate the MCLP coefficients and reverberation signal. In [16], a two-stage algorithm was proposed. In the first stage, the WPE was used to reduce reverberation, and in the second stage, the MVDR was utilized to remove the residual reverberation. This method treats the reverberation as a diffuse noise. The Kalman filter-based dereverberation algorithm that combines the MCLP and the GSC, known as the integrated sidelobe cancellation and linear prediction (ISCLP), has also shown good performance [12]. This method not only reduces the complexity of simply cascading the MCLP and beamforming, but also improves the performance. However, in the ISCLP, estimating linear prediction coefficients with the Kalman filter needs to consider the initialization of the Kalman filter parameters [17,18,19]. These parameters, such as the power spectral density (PSD) of the target signal, are critical to the performance of the algorithm. Some related dereverberation methods [20,21,22,23] also heavily rely on the PSD estimation of the speech signal.
This paper proposes a dereverberation algorithm by combining the complex Gaussian mixture model (CGMM)-based MVDR (CGMM-MVDR) beamforming [24] and the linear prediction Kalman filter. In this algorithm, the spectral model based on the CGMM is used to estimate the time–frequency masks of the covariance matrix of the target speech and the steering vector. The covariance matrices of the target speech and noise are estimated through the time–frequency masks while considering the noise only, where the eigenvector corresponding to the maximum eigenvalue of the covariance matrix of the target speech is an estimation of the steering vector of the MVDR beamforming. At the same time, the main diagonal elements of the covariance matrix of the target speech are considered as the estimated PSD of the target speech, which are also required by the Kalman filter. Since this method can obtain a more reasonable estimation for the parameters, the performance of dereverberation is better than the ISCLP. In addition, the low-complexity version of the proposed method is proposed as well, in which the multichannel linear prediction coefficients are expressed in terms of the spatial filter and the interframe filter by using the Kronecker product. By determining the spatial filter, the Kalman filter is used to estimate the interframe filter. Since the dimension of the interframe filter is lower than that of the original multichannel linear prediction filter, the size of matrices in the Kalman filter also decreases so that the amount of calculation is greatly reduced. Experiments show that this method can effectively reduce computational complexity and obtain relatively good results.
In the following parts, we arrange the paper as follows. In the Section 2, we briefly introduce the signal model of the CGMM-based MVDR beamformer and the related linear prediction model. In the Section 3, we present the proposed method by combining the CGMM-based MVDR beamformer and Kalman filter. In the Section 4, we show and analyze the experimental results. Finally, we conclude our method.

2. Signal Model

In this paper, one sound source is considered, and M microphones are used to collect the observation signal containing the reverberant and noisy speech signal. By performing a short-time Fourier transform (STFT), the received signal of the mth microphone can be expressed as ym (k,l) in the frequency domain, where k and l are the indices of the frequency bins and frames, respectively. The entire stacked microphone signals vector is denoted as follows:
y ( k , l ) = [ y 1 ( k , l ) , , y m ( k , l ) , , y M ( k , l ) ] T
where superscript “T” indicates the transpose of the matrix. Thus, the multimicrophone signals corrupted by reverberation and noise can be written as follows:
y ( k , l ) = x ( k , l ) + v ( k , l )
where the vectors x(k,l) and v(k,l) are the STFT results of the speech signal containing reverberation and the noise signal, respectively. Here, the speech and noise are assumed to be mutually independent. Since the algorithm is finished independently in each frequency bin, the index k is omitted for the simple expression in the subsequent parts.

2.1. Complex Gaussian Mixture Model

Reviewing MVDR beamforming based on the CGMM [24], the enhanced speech signal by the beamforming can be expressed as follows:
x b ( l ) = w b H y ( l )
where superscript “H” denotes the conjugate transpose, wb is the vector of the filter coefficients for minimizing the output power of the beamforming under the constraint wbHg = 1. Here, g is the steering vector of the desired speech signal and the subscript “b” indicates the beamforming. Considering the sparsity of the speech signal in the time–frequency domain [25,26], the observed signal can be classified into three cases, i.e., the mixture of speech, noise, and reverberation; the mixture of speech and reverberation; and noise-only. Thus, the observation signals of the microphones can be written as follows [24]:
y ( l ) = g ( d ) x b ( d ) ( l )
where the superscript “d” denotes the scenarios of the interference such as reverberation with noise (in this case, d is denoted as x + r + n), reverberation-only (in this case, d is denoted as x + r), and noise-only (in this case, d is denoted as n).
The complex Gaussian distribution of the random variable x with the mean μ and variance σ2 is generally expressed as follows [24]:
N ( x ; μ , σ 2 ) = 1 π σ 2 e | x μ | 2 σ 2
Assuming that x b ( d ) ( l ) follows a complex Gaussian distribution x b ( d ) ( l ) ~ N ( 0 , ϕ ( l ) ( d ) ) , namely, its mean and variance are zero and ϕ(l)(d), respectively. Here, ϕ(l)(d) also indicates the power spectral density of x b ( d ) ( l ) . Thus, at each time–frequency bin, the multichannel microphone signals obey the following complex Gaussian mixture distribution [24]:
y ( l ) | d ~ N ( 0 , ϕ ( l ) ( d ) R ( d ) )
where R(d) = g(d) (g(d))H. Thus, the multichannel microphone signals can be described by the CGMM in the case of reverberation with noise or noise-only, in which the maximum likelihood estimation of the CGMM can be obtained by the expectation maximization (EM) algorithm [24].
The Q function in the EM algorithm is defined as follows [24]:
Q ( Θ ) = l d λ ( l ) ( d ) log N ( y ( l ) ; 0 , ϕ ( l ) ( d ) R ( d ) )
where λ(l)(d) denotes the mask or probability of the different scenarios, which can be obtained as follows:
λ ( l ) ( d ) = N ( y ( l ) ; 0 , ϕ ( l ) ( d ) R ( d ) ) / d N ( y ( l ) ; 0 , ϕ ( l ) ( d ) R ( d ) )
In order to achieve the online MVDR beamforming, the speech signal is divided into a sequence of batches. Each batch contains a limited number of the speech frames. The number of the frames can be changed. Let c ∈ {1, 2,…, C} denote the batch index and let lc denote the time index within the cth batch, where C is the number of the batches. For each batch, the parameters of the CGMM are updated by the following formulation [24]:
ϕ ( l ) ( d ) = 1 M tr { y ( l ) y ( l ) H [ R c ( d ) ] 1 }
R c ( d ) = A c 1 ( d ) A c 1 ( d ) + l l c λ ( l ) ( d ) R c 1 ( d ) + 1 A c 1 ( d ) + l l c λ ( l ) ( d ) l l c λ ( l ) ( d ) 1 ϕ ( l ) ( d ) y ( l ) y ( l ) H
where A c 1 ( d ) denote the sum of the masks over all the previous frames, and it can be updated as follows:
A c ( d ) = A c 1 ( d ) + l l c λ ( l ) ( d )
It is worth noting that R c ( d ) is usually irreversible. In practical calculation, the inverse of R c ( d ) is usually replaced by a pseudo inverse.
For every batch, the EM algorithm can be used to obtain the most optimal mask value for the current time–frequency bin only containing noise, and the covariance matrix in the case of the reverberation with noise or noise-only at each frequency bin can be estimated as follows [24]:
M c ( d ) = A c 1 ( d ) A c 1 ( d ) + l l c λ ( l ) ( d ) M c 1 ( d ) + 1 A c 1 ( d ) + l l c λ ( l ) ( d ) l l c λ ( l ) ( d ) y ( l ) y ( l ) H
So, the covariance matrix in the case of reverberation can be given as follows:
M c ( x + r ) = M c ( x + r + n ) M c ( n )
Performing the following eigenvalue decomposition
M c ( x + r ) = P Λ P 1
where P is the matrix composed of the eigenvectors of the matrix Mc(x + r), and Λ is the diagonal matrix containing the eigenvalues, the eigenvector associated with the largest eigenvalue is extracted as the steering vector required for the MVDR beamforming. The coefficient vector wb of the MVDR filter can be expressed as follows [24]:
w b = [ M c ( n ) ] 1 g / { g H [ M c ( n ) ] 1 g }
In fact, the CGMM-MVDR is a beamforming-based method with blind estimation of the steering vector. The late reverberation leads to an inaccurate estimation of the steering vector. Since the blind estimation of the accurate steering vector from the reverberated speech with noise is a very difficult task, what we obtain here is the approximation of the steering vector. This does not mean that such approximations lead to false results, because the reverberation has been mostly removed by Kalman filtering rather than beamforming.

2.2. Multichannel Linear Prediction Reverberation Model

By modeling reverberation components using the MCLP and predicting reverberation components of the current frame using the signals coming from the previous (L − D) frames, where D and L are the frame numbers prior to the current frame, respectively, the speech signal with reverberation can be expressed as follows [12]:
x ( l ) = n = D L B r , n ( l ) x ( l n ) r ( l ) + x e ( l )
where Br,n (l) denotes the multichannel linear prediction coefficients, and L > D ≥ 1. The reverberation is divided into two parts: the early reverberation and the late reverberation. The result r (l) represents the late reverberation components. Here, the desired signal xe (l), which is the direct signal with early reverberation components, is also called the prediction error in the linear prediction model. The prediction error is expressed as x e ( l ) = [ x e 1 ( l ) , x e 2 ( l ) , , x e M ( l ) ] T , which contains the desired signal of each microphone. Note that τ = (LD) is the number of the delayed frames that are used to predict the late reverberation components. The selection of τ should make the correlation between the late reverberation and the expected signal approach zero.

3. Proposed Dereverberation Method

The block diagram of the proposed dereverberation method is shown in Figure 1. In the upper path, the CGMM-based MVDR beamformer is used to enhance the input signal y (l) collected by the microphones so that we can obtain the enhanced signal xb (l). The PSD of the desired signal used for the Kalman filter is approximately equal to the power spectral density ϕ x b ( l ) of the xb (l). In the lower path, according to the delayed signals t (l) of the input signal y (l), the MCLP is performed based on the Kalman filter to estimate the reverberation signal z (l) by linear prediction. The input to the Kalman filter-based linear predictor is t (l), and the corresponding output is the estimated reverberation signal z (l). By subtracting z (l) from xb (l), we can finally obtain the enhanced signal s (l). Different from ISCLP [12], this paper uses the MVDR beamforming to replace the GSC, because the GSC has the shortage of speech distortion, that is, if the steering vector of the desired speech used in the GSC is not accurate, the desired speech will be inevitably destroyed by its block matrix. In particular, in the ISCLP, the PSD of the target signal is estimated by a complex algorithm, which contains some prior information [12], whereas the proposed method uses a simpler approach without any prior information to estimate the PSD of the target signal.

3.1. Algorithm Architecture

In the upper path of Figure 1, by using Equations (2), (3) and (15), the output xb (l) of beamforming is obtained as follows:
x b ( l ) = w b H y ( l ) = x e ( l ) + w b H [ r ( l ) + v ( l ) ]
Although most of the noise can be suppressed by the MVDR beamformer, the effect of dereverberation is very limited [27]. The scalar xe (l) denotes the inner product wbHxe (l).
In the lower path of Figure 1, the past (L − D) frames are used as the input signal of the linear predictor. By combining the vectors from y (lD) to y (l − L + 1), the input matrix of the linear predictor at the lth frame is given as follows:
t ( l ) = [ y T ( l D ) , , y T ( l L + 1 ) ] T
and the prediction signal z(l) can be expressed as follows:
z ( l ) = w r H ( l ) t ( l )
where wr(l) is the linear prediction filter. An appropriate D should be selected according to the frame shift in the STFT operation in order to make the correlation between the prediction error xe(l) and the prediction signal z(l) zero, i.e.,
z , x e = l z ( l ) x e ( l ) = 0
The choice of D determines the amount of early reverberation preserved in the desired signal, which affects the performance of the algorithm [15].
According to the architecture of the proposed method, the final enhanced speech s(l) is obtained as follows:
s ( l ) = x b ( l ) w r H ( l ) t ( l )
where the output s (l) is regarded as an estimation of xe (l). If wr (l) is available, we can estimate the final enhanced signal s (l) from Equation (21). In the ISCLP method [12], the Kalman filter was used to jointly estimate the parameters of the sidelobe canceller (SC) and the linear predictor. Since the linear predictor needs information from the past frames, the canceller and predictor cannot work on the same frame simultaneously. It does not mean they are mutually independent due to the common inclusion of the late reverberation components. This implies that changes in the canceller parameters cause changes in the predictor parameters, and vice versa. To avoid this problem, in this paper, the coefficients of the beamformer and the coefficients of the linear predictor are estimated separately. Thus, the result of the beamforming will improve the result of the linear prediction as well.

3.2. Application of Kalman Filter

The linear prediction coefficients can be estimated effectively by the Kalman filter [28,29,30]. In the ISCLP [12], the Kalman filter used for estimating the linear prediction coefficients has shown good results for dereverberation. Here, we define the observation equation of the Kalman filter [12] as follows:
x b * ( l ) = t H ( l ) w r ( l ) + x e * ( l )
where the superscript “*” indicates the conjugation operation. The vector wr (l) is the state vector with zero mean, and its related covariance matrix is expressed as follows:
Φ w ( l ) = E [ w r ( l ) w r H ( l ) ]
and xe* (l) in Equation (22) is the desired signal; it is also known as the measurement error. Similarly, the measurement error can also be regarded as a Gaussian process with zero mean, as defined by Equation (6), and its power spectral density is ϕ x e ( l ) = E [ x e ( l ) x e * ( l ) ] .
The change in wr (l) means the change in the RIR or the change in the microphone positions, so wr (l) needs to be adjusted by the state equation. Considering the first-order Markov process, the state equation can be defined as follows [14]:
w r ( l ) = A H ( l ) w r ( l 1 ) + v w ( l )
where A (l) is the state transition matrix; it indicates the prediction result of the state vectors from the previous frames to the current frame. The zero-mean Gaussian model with the following covariance matrix can be used to model the process noise vw (l):
Φ v ( l ) = E [ v w ( l ) v w H ( l ) ]
According to Equations (22) and (24), the purpose of the Kalman filter is to find the optimal solution for the linear prediction coefficients based on the minimum mean square error (MMSE). Usually, A (l) and vw (l) will be adjusted at the beginning of the algorithm [31].
In the Kalman filtering process, the essence of the algorithm is to determine the required parameters step by step through two stages. In the first stage, the state vector wr (l) of the next moment is predicted through the state equation. This prediction is called a priori time update. In the second stage, the prediction result of the state vector in the first stage is weighted with the so-called Kalman gain that is calculated by the observation equation combined with the state vector predicted in the first stage. The expected optimal state vector wr (l) can be obtained by the iterations of these two stages.
Let ŵr (l) be the estimated result given by the state Equation (24) and ŵr+ (l) be the modified result from the observation Equation (22); their corresponding errors are obtained as follows:
e ( l ) = w r ( l ) w ^ r ( l )
e + ( l ) = w r ( l ) w ^ r + ( l )
the covariance matrices of e(l) and e+(l) are given as follows:
Φ e ( l ) = E [ e ( l ) e H ( l ) ]
Φ e + ( l ) = E [ e + ( l ) e + H ( l ) ]
It is worth noting that these two covariance matrices of the errors can be regarded as the evolution in the same frame, that is, Φ e + ( l ) is a further estimation of Φ e ( l ) . According to the following cost function:
J { w r ( l ) , w ^ r + ( l ) } = tr { E [ e + ( l ) e + ( l ) H ] }
the update process and the estimation of gain k(l) of the Kalman filter are expressed as follows [12]:
w ^ r ( l ) = A ( l ) w ^ r + ( l 1 )
Φ e ( l ) = A H ( l ) Φ e + ( l 1 ) A ( l ) + Φ v ( l )
s * ( l ) = x b * ( l ) t H ( l ) w ^ r ( l )
ϕ s ( l ) = t H ( l ) Φ e ( l ) t ( l ) + ϕ x e ( l )
k ( l ) = Φ e ( l ) t ( l ) ϕ s 1 ( l )
w ^ r + ( l ) = w ^ r ( l ) + k ( l ) s * ( l )
Φ e + ( l ) = Φ e ( l ) k ( l ) t H ( l ) Φ e ( l )
where Equations (31) and (32) are used to update the predictive coefficients and the covariance matrix of the errors in the current frame based on the information from the previous frame. In the following, the complex conjugate error signal s* (l) (i.e., the final enhanced signal) and its power spectral density ϕs (l) are calculated using Equations (33) and (34) by combining the observation Equation (22) for calculating the Kalman gain k (l) in Equation (35). The final predictive coefficients and covariance matrix of the errors in the current frame are finally obtained by Equations (36) and (37) with the Kalman gain k (l). The final results are also used to calculate the information needed for the next frame. In the calculation of the next frame, however, the error signal s (l) and the Kalman gain are reset, which means that they only represent the current frame and do not iterate through the frames. The complex conjugate error signal s* (l) is the signal we need, which is given by the Kalman filter. The initialization of the Kalman filter is introduced in the following content.

3.3. Low-Complexity Algorithm Based on the Kronecker Product

The Kalman filter has the disadvantage of high computational complexity. By observing update Equations (31)~(37) of the Kalman filter, it can be found that the calculation of the Kalman filter involves the multiplication and summation of many matrices with high dimension data, so the dimension reduction in the matrix is the key for reducing the calculation complexity. In the observation Equation (22), the stacked microphone signal t (l) is a vector with M (LD) elements, wr (l) is a vector including M (LD) linear prediction coefficients. According to the method proposed in [32], wr (l) can be expressed as the form of the Kronecker product of two subfilters, where two low-dimension subfilters can be estimated separately. Considering a signal model such as Equation (21), we have
x e ( l ) = w 2 H ( l ) y ( l ) [ w 1 ( l ) w 2 ( l ) ] H t ( l )
where xe (l) is the desired signal, symbol “⊗” represents the Kronecker product, x ^ ( l ) = w 2 H ( l ) y ( l ) , w r ( l ) = w 1 ( l ) w 2 ( l ) , w1 (l), and w2 (l) can be regarded as a (L − D)-order interframe filter and a M-order spatial filter, respectively. In fact, w1 (l) represents the filter for processing the consecutive frames, and the main purpose of w1 (l) is to predict the reverberation components from the past L − D frames. w2 (l) represents the filter that processes signal from all channels. Thus, w2 (l) can be regarded as the coefficients of the spatial filter. By using the Kronecker product, the original optimization problem becomes two optimal problems of the subfilters.
According to the characteristic of the Kronecker product [32], we can obtain
w 1 ( l ) w 2 ( l ) = [ I w 2 ( l ) ] w 1 ( l ) = W 2 ( l ) w 1 ( l )
where I is a (LD) × (LD) identity matrix and W2 (l) = Iw2 (l) is a M (LD) × (LD) matrix. Substituting (39) into (38), we have
x e ( l ) = w 2 H ( l ) y ( l ) w 1 H ( l ) W 2 H ( l ) t ( l ) = x ^ ( l ) w 1 H ( l ) t W 2 ( l )
where t W 2 ( l ) = W 2 H ( l ) t ( l ) .
Since w2(l) is a spatial filter, w2(l) in (40) can be replaced by the CGMM-MVDR filter, i.e.,
x e ( l ) = w b H y ( l ) w 1 H ( l ) W H t ( l ) = x b ( l ) w 1 H ( l ) t W ( l )
where wb = w2 (l), W = (Iwb), tw (l) = WH (l) t (l). It is worth noting that wb is different at each frequency bin of all the frames. tw (l) can be regarded as the speech signal that was processed by the spatial filter in advance. After such replacement, two optimal problems of the subfilters become a single optimal problem of the interframe filter that can be estimated by the Kalman filter. Moreover, the state equation and the observation equation of the improved low-complexity Kalman filter are given as follows:
w 1 ( l ) = A H ( l ) w 1 ( l 1 ) + w Δ ( l )
and
x b * ( l ) = t w H ( l ) w 1 ( l ) + x e * ( l )
where the process noise w (l) can be modeled by the zero-mean Gaussian process with the covariance matrix Φ w Δ ( l ) = E [ w Δ ( l ) w Δ H ( l ) ] .
Let ŵ1 (l) be the estimated result by the state Equation (42) and w ^ 1 + ( l ) be the modified result with the observation Equation (43); their corresponding errors are obtained as follows:
e 1 ( l ) = w 1 ( l ) w ^ 1 ( l )
e 1 + ( l ) = w 1 ( l ) w ^ 1 + ( l )
the covariance matrices of e1(l) and e 1 + ( l ) are given as follows:
Φ e 1 ( l ) = E [ e 1 ( l ) e 1 H ( l ) ]
Φ e 1 + ( l ) = E [ e 1 + ( l ) e 1 + ( l ) H ]
According to the state Equation (42), the observation Equation (43), and following cost function:
J { w 1 ( l ) , w ^ 1 + ( l ) } = tr { E [ e 1 + ( l ) e 1 + ( l ) H ] }
we can obtain update equations for Kalman filtering with low complexity similar to Equations (31)~(37) as follows:
w ^ 1 ( l ) = A ( l ) w ^ 1 + ( l 1 )
Φ e 1 ( l ) = A H ( l ) Φ e 1 + ( l 1 ) A ( l ) + Φ w Δ ( l )
s 1 * ( l ) = x b * ( l ) t W H ( l ) w ^ 1 ( l )
ϕ s 1 ( l ) = t W H ( l ) Φ e 1 ( l ) t W ( l ) + ϕ x e ( l )
k 1 ( l ) = Φ e 1 ( l ) t W ( l ) ϕ s 1 1 ( l )
w ^ 1 + ( l ) = w ^ 1 ( l ) + k 1 ( l ) s 1 * ( l )
Φ e 1 + ( l ) = Φ e 1 ( l ) k 1 ( l ) t W H ( l ) Φ e 1 ( l )
where s1* (l) is the final enhanced signal by the low-complexity algorithm with power spectral density ϕ s 1 ( l ) and k1 is the Kalman filter gain of the low-complexity algorithm. Thus, the coefficients of the interframe filter w1 (l) can be estimated. Through Equations from (38) to (55), the dimension of the filter coefficients wr (l) is significantly reduced so that the computational complexity is reduced effectively. The effectiveness of the low-complexity algorithm is discussed in the experiments.

3.4. Initialization of Kalman Filtering

Some parameters need to be initialized at the start of the Kalman filtering. In update Equations (31)~(37) and (49)~(55), these parameters include the state transition matrix A (l), covariance matrix Φv (l) of the process noise vw (l), covariance matrix Φ w Δ ( l ) of the process noise w (l), power spectral density ϕ x e ( l ) of the desired signal, and the covariance matrices Φe (l) and Φ e 1 ( l ) of the errors e (l) and e1 (l).
The state transition matrix A(l) is regarded as a time-invariant function similar to the forgetting factor in the update equation [31]. The covariance matrices Φe (l) and Φ e 1 ( l ) gradually become convergent during the Kalman filtering. The covariance matrix Φe (l) can be initialized as a diagonal matrix, i.e., Diag [Φe(l)] = φe, where vector φe includes M (L − D) elements that rise exponentially with every M elements [12], i.e.,
φ e = [ φ e 1 1 T , φ e 2 1 T , , φ e ( L D ) 1 T ] M ( L D ) × 1 T
where 1 = [ 1 , 1 , , 1 ] M × 1 T , φ e 1 , φ e 2 , …, φ e ( L D ) are L − D scalars that rise exponentially. Similarly, the covariance matrix Φe1 (l) is also initialized as a diagonal matrix, i.e., Diag [Φe1 (l)] = φe1, where vector φe1 only includes L − D elements that drop exponentially, i.e.,
φ e 1 = [ φ e 1 1 , φ e 1 2 , , φ e 1 ( L D ) ] T
The process noises vw (l) and w (l) are modeled by the zero-mean Gaussian model, and their covariance matrixes Φv (l) and Φw∆ (l) are usually assumed to be time-invariant and initialized as follows [12]:
Φ v ( l ) = ( 1 β ) Φ e ( l )
Φ w Δ ( l ) = ( 1 β ) Φ e 1 ( l )
where β ∈ (0,1) is the forgetting factor.
The weighted prediction error (WPE) algorithm also needs to initialize the PSD of the desired signal. In the WPE, the PSD of unprocessed signal is regarded as one of the desired signals to estimate the linear prediction coefficients used for obtaining the enhanced signal. Then, the PSD of the enhanced signal is used to estimate the new linear prediction coefficients. The above steps are iterated until the algorithm converges. When the WPE is used for a short-time speech signal, the frame number of the speech signal has to be reduced. In this case, the estimated PSD of the desired signal greatly degrades the performance of the algorithm because of using few frames. Therefore, a method of directly estimating the PSD of the desired signal from the reverberated speech by using the DNN was proposed [33]. In [14,15], the PSD of the desired signal is determined by a decision-oriented method, that is, the current PSD of the desired signal is composed of the PSD of the previous frame and the PSD of the current frame; the later one is obtained by the linear prediction coefficients of the previous frame. These two PSDs are weighted by a weight factor. This method relies on the setting of the weight factor, and how to choose an appropriate weight factor is crucial.
In this paper, the speech signal processed by the beamformer is utilized to initialize the PSD of the desired signal. We can redefine the covariance matrices presented in (11) and (12) at each time–frequency bin as follows:
C ( x + r + n ) ( l ) = y ( l ) y ( l ) H
C ( n ) ( l ) = λ ( l ) ( n ) y ( l ) y ( l ) H
where λ (l)(n) denotes the mask or probability of the scenarios that only contain noise. Thus, the covariance matrix of the noiseless speech at each time–frequency bin can be given as follows:
C ( x + r ) ( l ) = C ( x + r + n ) ( l ) C ( n ) ( l )
Obviously, the diagonal elements of the covariance matrix C(x+r)(l) represent the PSD of the speech signal. We use vector θ(l) to express its diagonal elements as follows:
θ ( l ) = Diag [ C ( x + r ) ( l ) ] = [ ϕ x b 1 ( l ) , ϕ x b 2 ( l ) , , ϕ x b m ( l ) , , ϕ x b M ( l ) ] T
where ϕ x b m ( l ) represents the power spectral density of the mth microphone signal that is processed by the beamformer. The diagonal elements are summed and averaged as the approximation of the PSD of the desired signal required in the Kalman filtering process, i.e.,
ϕ x b ( l ) = 1 M m = 1 M ϕ x b m ( l )
where ϕ x b ( l ) is an approximation of the power spectral density ϕ x e ( l ) of the desired signal, and the low-complexity version also utilize ϕ x b ( l ) as the initialization of ϕ x e ( l ) . Under ideal circumstances, the closer the initialized PSD is to the PSD of the desired speech, the better the algorithm will achieve, but this is obviously impractical. Blind estimation of the PSD of the desired signal is very difficult only from the signal received by the microphones. Compared with the PSD estimation method in ISCLP [12], the proposed method does not need the incident angle of the desired speech and other prior information. Applying such crude approximation does not mean a wrong result, because the linear prediction coefficients are used instead of the desired signal itself in the Kalman filtering, the effect of the approximation error can be mitigated during the Kalman filtering. Meanwhile, the computational complexity of this kind of PSD estimation is lower than that used in ISCLP. Therefore, it is feasible to adopt this approximation of the PSD, which is also confirmed by the following experiments.

4. Experiments and Evaluation

In this section, we give the experimental results. The proposed dereverberation method is compared with the reference methods including ISCLP and WPE, which have excellent performance at present. Firstly, we introduce the experimental settings used for the evaluation. Then, the reference methods are introduced briefly. Finally, the experimental results are evaluated and discussed.

4.1. Acoustic Scenario and Experimental Setup

The observation signals are generated the same as described in [34], and the TIMIT corpus [35] is used for the simulation. As shown in Figure 2, the length, width, and height of the room are 4, 3, and 3 m, respectively. A uniform linear microphone array with eight microphones is placed on a table for receiving speech signals from all angles in the room, the microphone spacing is 0.04 m, and the linear microphone array has a height of 0.75 m with the table. Meanwhile, there are 20 speech source positions in the room. Eight source positions, namely Tar.1, Tar.2, Tar.3, Tar.6, Tar.7, Tar.8, Tar.10, and Tar. 20, are chosen for the test (Figure 2), and 100 utterances are selected for each position. The selection of the source positions is based on their symmetry in Figure 2, that is, Tar.1 and Tar.2 are similar to Tar.5 and Tar.4, Tar.15 and Tar.16, Tar.19 and Tar.18, respectively. Tar.3 is similar to Tar.17. Tar.6 is similar to Tar.11. Tar.7 is similar to Tar.9, Tar.12 and Tar.14. Tar.8 is similar to Tar.13. These eight source positions are represented in Figure 2. Babble noise is used to obtain the reverberate speech with noise, the input signal-to-noise ratio (SNR) is set to −5, 0, 5, and 10 dB, and the reverberation time T60 is set to 100, 400, 500, 600, 700, and 800 ms. When investigating the dependence of the algorithm on the order L of the linear prediction filter and the effect of the different initializations of the PSD on the proposed method, we selected the speech signal from the first position as the evaluation object. When comparing with the reference methods, all 800 utterances of the eight aforementioned locations in Figure 2 were utilized.
The speech signal is sampled at 16 kHz, a Hanning window with a length of 512 samples has 50% overlap for analyzing the speech signal, and a 512-point STFT is used for the windowed speech signal. The frame number for each batch in the MVDR beamforming is set to 70. The delay D that preserves the early reflections is set to 1, and the order L of the linear prediction filter is set to 10. The state transition matrix A (l) = αI, where α = 0.998. Similar to the default setting in [12], we let 10 log10(φe) = −4 dB and 10 log10(φe1) = −4 dB in (56) and (57), whereas 10 log10 (1 − β) = −25 dB in (58) and (59).
To evaluate the proposed algorithm, we choose the perceptual evaluation of speech quality (PESQ) [36], the short-time objective intelligibility (STOI) [37], the cepstral distance (CD) [38], the frequency-weighted segmental SNR (fwSegSNR) [39], and the signal-to-interference ratio (SIR) [39] as the measures, where the reverberation and noise are considered as the interference. These measures have a reasonable evaluation of the perception of reverberation and the overall quality of speech [40]. The CD focuses on reflecting the distortion of the speech spectrum: the lower the CD, the less overall speech distortion. The PESQ, STOI, fwSegSNR, and SIR can better demonstrate the reduction in reverberation and noise.

4.2. Reference Methods

To demonstrate the effectiveness of the proposed method, the ISCLP and WPE methods that have excellent performance at present are chosen for the comparison.
ISCLP realized in [41] is a type of method for noise reduction and dereverberation by combining the MCLP and GSC. In the ISCLP, the filter coefficients of the MCLP and GSC are simultaneously estimated by the Kalman filter. In the case of an interfering speaker, the ISCLP can effectively eliminate the interfering speaker, but at the same time it also introduces distortion to the target speaker.
WPE realized in [42] is a type of dereverberation method based on linear prediction and is the most widely used dereverberation algorithm as well. It assumes that the speech signal follows the complex Gaussian distribution with time variance. Its model parameters and linear prediction filter coefficients can be obtained iteratively through maximum likelihood estimation. This algorithm has good suppression effect on reverberation, but it is easily interfered by the noise.

4.3. Analysis and Comparison of the Test Results

4.3.1. Effect of Filter Order

The effect of the filter order L on the algorithm performance is investigated in an exemplary simulation condition, namely, SNR = 10 dB and T60 = 800 ms. Here, the proposed method and ISCLP are evaluated based on the variation in L. Figure 3 shows the incremental results of the PESQ, STOI, CD, and SIR for different filter orders. The ISCLP shows better performance with respect to the PESQ, STOI, and CD in the case of shorter filter order, while for SIR, the ISCLP shows the dependence on the filter order. The proposed method shows the dependence on the filter order in the PESQ, STOI, and SIR results, and its performance is the best at eight options of the filter order and tends to be convergent at some fixed order L for different tests. In the evaluation results of the PESQ, STOI, and CD, the proposed method is better than ISCLP at all filter orders. In the evaluation result of the SIR, the proposed method is slightly worse than ISCLP. It can be seen from Figure 3 that the CD results of the proposed method and ISCLP have positive increments, which indicates that the speech distortion is increased. This may be caused by the inaccurate estimation of some key parameters in the Kalman filtering, but the distortion caused by the ISCLP is significantly larger than the proposed method.

4.3.2. Effect of the PSD Initialization

We measured the effect of different initialization methods of the PSD on the proposed method when the input SNR = 10 dB and T60 = {400, 500, 600, 700, 800} ms at the same source positions as given in Section 4.3.1. Four initialization methods of the PSD of the desired signal, i.e., the method given in ISCLP, the proposed initialization method, the method given in [33], and the method given in [14,15], are employed in the proposed dereverberation algorithm. The effects of these methods on the performance of the algorithms are compared. In addition, the initialization of the PSD using real PSD of the desired signal is also compared. Figure 4 shows the result comparison.
As expected, the result of using real PSD of the desired signal greatly improved the performance, which is the best among four evaluation indicators. The effect of the initialization of the ISCLP and the proposed method are far better than the initialization given in [33] and [14,15]. In terms of SIR, the effects of these four initialization methods (excluding the real PSD) are very similar.
The effect of the initialization of the ISCLP is slightly better than the proposed method in the results of the PESQ, STOI, and SIR, but the ISCLP needs prior information such as the incident angle of the desired signal and the position of the microphones in advance. In the ISCLP, the inaccurate incident angle of the desired signal makes the block matrix in the GSC filter miss useful signal, which causes the distortion of the desired signal in the subsequent processing. The proposed method only uses the signals received by the microphones to obtain the estimation of the PSD of the desired signal, which is an adaptive approach. At the same time, the computational complexity of the initialization of the PSD in the proposed method is far lower than that of the ISCLP. From the experimental results, we can find that obtaining the PSD close to the real PSD of the desired signal is an important way for effectively improving the performance of the algorithm. How to obtain the accurate PSD of the desired signal is a key issue for the dereverberation.

4.3.3. Performance Comparison with Reference Methods

Two proposed dereverberation algorithms and two reference methods are evaluated for different SNR and T60. Six kinds of reverberation time are considered at four SNR conditions. Table 1, Table 2, Table 3 and Table 4 show the results of performance evaluation for the unprocessed speech. It is worth noting that the PESQ value of the unprocessed signal increases with the increase in the reverberation time. This is because in the case of lower input SNR, compared with low reverberation, the high reverberation can resist more noise, so that the speech is not submerged into the annoying noise. Under the condition of high SNR, the performance of the proposed method is not very good compared with the reference methods, and especially worse than the WPE. The reason is that under the condition of high SNR, the speech distortion caused by the noise is far less than that caused by the reverberation. Meanwhile, WPE can remove reverberation well and improve the speech quality greatly. Under the condition of low reverberation, the ISCLP shows better results and the WPE shows worse results, because WPE has no denoising ability. Under the condition of low SNR, it can be found that the results of the PESQ, STOI, and fwSegSNR of the proposed method are higher than these two reference methods in most cases, and the result of STOI is significantly improved especially in the case of −5 dB, whereas the low-complexity version of the proposed method (called proposed-kron) is similar to the proposed method. As shown in Table 1 and Table 2, the fwSegSNR is also significantly improved in all cases. In the results for CD, the WPE shows the best performance. As the distortion measurement of the speech, the results of CD show that both the proposed method and ISCLP introduce the distortion, and the ISCLP causes more distortion. The distortion is due to the PSD error of the desired signal in the Kalman filtering. The GSC-based beamforming used in the ISCLP causes speech distortion because of the inaccurate RETF. For the SIR, the ISCLP shows the best effect. The ISCLP reduces much of the interference at the cost of introducing more speech distortion. The proposed method balances the speech distortion and interference reduction. Within a certain distortion range, the proposed method eliminates more interference signal than the WPE. The proposed-kron method also shows good results with low complexity. From the discussion above, the proposed method shows better performance under the condition of low SNR and high reverberation. In future work, the problem of speech distortion in the proposed method needs to be improved.

4.3.4. Comparison of the Spectrogram

Figure 5 shows the spectrogram comparison for the original speech, the enhanced speech obtained by the ISCLP with filter order L = 6, the enhanced speech obtained by the WPE, and the enhanced speech obtained by the proposed algorithm. The input SNR is set to 10 dB and T60 = 800 ms. Compared with the original speech, these three algorithms effectively reduce the reverberation and exhibit different characteristics at the same time. As shown in Figure 5b, much of the reverberation and noise are suppressed by the ISCLP, but the distortion of the desired signal is also serious. Figure 5c shows that in the WPE, most of the reverberation is suppressed and the desired speech is retained well, whereas the noise is hardly suppressed. Figure 5d shows that the proposed algorithm can reduce the noise and reverberation effectively, and can retain the high frequency components of the desired speech. Compared with the WPE algorithm, the proposed algorithm has better ability for suppressing the noise.

4.3.5. Comparison of Computational Complexity

In order to compare the computational complexity of the “proposed” method and the “proposed-kron” method, Table 5 lists the operations required for the multiplication, addition, and division of two proposed algorithms. The update Equations (31), (32), (49), and (50) of the Kalman filtering are used for the calculation of the first step, whereas update Equations (33)–(35) and (51)–(53) are utilized for the calculation of the second step. Furthermore, the update Equations (36), (37), (54), and (55) are employed for the calculation of the third step. The complexity reduction factor is defined as the ratio between the operations required for the “proposed-kron” and “proposed” methods in each step. We can see that the “proposed-kron” algorithm has greatly improved computational efficiency compared to the “proposed” algorithm. Especially, the multiplication operations in the first and third steps are reduced much more. The “proposed-kron” method not only effectively reduces the computational complexity but also guarantees the speech quality compared to the “proposed” method.

5. Conclusions

In this paper, a noise reduction and dereverberation algorithm was presented by combining the CGMM-based MVDR beamforming and the multichannel linear prediction with Kalman filtering. The noise and a part of reverberation were processed by the MVDR beamforming, meanwhile the covariance matrix of the target speech estimated by beamforming was utilized to estimate the PSD of the target signal, which is an important parameter for Kalman filtering. In addition, a low-complexity version of the proposed algorithm was given. The effectiveness of the proposed method in the experiment, the influence of different PSD initialization methods on the algorithm, and the importance of accurate PSD estimation to the algorithm were carefully analyzed. The evaluation results proved that the proposed algorithm is superior to the reference methods and the low-complexity version keeps similar effectiveness with much less complexity.

Author Contributions

Conceptualization, C.B., F.T. and J.Z.; methodology, C.B. and F.T.; software, F.T. and J.Z.; validation, C.B., F.T. and J.Z.; formal analysis, C.B. and F.T.; investigation, F.T.; resources, F.T.; data curation, F.T. and J.Z.; writing—original draft preparation, C.B., F.T. and J.Z.; writing—review and editing, C.B., F.T. and J.Z.; visualization, F.T.; supervision, C.B. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded by the National Natural Science Foundation of China (Grant No. 61831019).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Exclude this statement.

Acknowledgments

The authors are grateful to all the reviewers and editors.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Beutelmann, R.; Brand, T. Prediction of speech intelligibility in spatial noise and reverberation for normal-hearing and hearing-impaired listeners. J. Acoust. Soc. Amer. 2006, 120, 331–342. [Google Scholar] [CrossRef] [PubMed]
  2. Nakatani, T.; Kellermann, W.; Naylor, P.; Miyoshi, M.; Juang, B.H. Introduction to the special issue on processing reverberant speech: Methodologies and applications. IEEE/ACM Trans. Audio Speech Lang. Process. 2010, 18, 1673–1675. [Google Scholar] [CrossRef] [Green Version]
  3. Naylor, P.A.; Gaubitch, N.D. Speech Dereverberation; Springer: London, UK, 2010. [Google Scholar]
  4. van Veen, B.D.; Buckley, K.M. Beamforming: A versatile approach to spatial filtering. IEEE ASSP Mag. 1988, 5, 4–24. [Google Scholar] [CrossRef]
  5. Gannot, S.; Burshtein, D.; Weinstein, E. Signal enhancement using beamforming and nonstationarity with applications to speech. IEEE Trans. Signal Process. 2001, 49, 1614–1626. [Google Scholar] [CrossRef] [Green Version]
  6. Markovich, S.; Gannot, S.; Cohen, I. Multichannel eigenspace beamforming in a reverberant noisy environment with multiple interfering speech signals. IEEE Trans. Signal Process. 2009, 17, 1071–1086. [Google Scholar] [CrossRef]
  7. Schwartz, O.; Gannot, S.; Habets, E.A.P. Nested generalized sidelobe canceller for joint dereverberation and noise reduction. In Proceedings of the IEEE International Conference on Acoustic, Speech and Signal Processing (ICASSP), South Brisbane, QL, Australia, 19–24 April 2015; pp. 106–110. [Google Scholar]
  8. Arote, S.; Deshpande, M. Multichannel speech dereverberation using generalized sidelobe canceller and post filter. In Proceedings of the IEEE 5th International Conference on Computing, Communication, Control and Automation, Pune, India, 19–21 September 2019; pp. 1–4. [Google Scholar]
  9. Nathwani, K.; Hegde, R.M. Speech dereverberation in multisource environment using LCMV filter. In Proceedings of the IEEE International Symposium on Signal Processing & Information Technology, Noida, India, 15–17 December 2014; pp. 404–409. [Google Scholar]
  10. Schwartz, O.; Gannot, S.; Habets, E.A.P. Multi-microphone speech dereverberation and noise reduction using relative early transfer functions. IEEE/ACM Trans. Audio Speech Lang. Process. 2015, 23, 240–251. [Google Scholar] [CrossRef]
  11. Nakatani, T.; Kinoshita, K. A unified convolutional beamformer for simultaneous denoising and dereverberation. IEEE Signal Process. Lett. 2019, 26, 903–907. [Google Scholar] [CrossRef] [Green Version]
  12. Dietzen, T.; Doclo, S.; Moonen, M.; van Waterschoot, T. Integrated sidelobe cancellation and linear prediction Kalman filter for joint multi-microphone speech dereverberation, interfering speech cancellation, and noise reduction. IEEE/ACM Trans. Audio Speech Lang. Process. 2020, 28, 740–754. [Google Scholar] [CrossRef] [Green Version]
  13. Nakatani, T.; Yoshioka, T.; Kinoshita, K.; Miyoshi, M.; Juang, B.-H. Speech dereverberation based on variance-normalized delayed linear prediction. IEEE/ACM Trans. Audio Speech Lang. Process. 2010, 18, 1717–1731. [Google Scholar] [CrossRef]
  14. Hashemgeloogerdi, S.; Braun, S. Joint beamforming and reverberation cancellation using a constrained Kalman filter with multichannel linear prediction. In Proceedings of the IEEE International Conference on Acoustic, Speech and Signal Processing (ICASSP), Barcelona, Spain, 4–8 May 2020; pp. 481–485. [Google Scholar]
  15. Braun, S.; Habets, E.A.P. Linear prediction-based online dereverberation and noise reduction using alternating Kalman filters. IEEE/ACM Trans. Audio Speech Lang. Process. 2018, 26, 1119–1129. [Google Scholar] [CrossRef]
  16. Cohen, A.; Stemmer, G.; Ingalsuo, S.; Markovich-Golan, S. Combined weighted prediction error and minimum variance distortionless response for dereverberation. In Proceedings of the IEEE International Conference on Acoustic, Speech and Signal Processing (ICASSP), New Orleans, LA, USA, 5–9 March 2017; pp. 446–450. [Google Scholar]
  17. Dietzen, T.; Spriet, A.; Tirry, W.; Doclo, S.; Moonen, M.; van Waterschoot, T. Partitioned block frequency domain Kalman filter for multi-channel linear prediction based blind speech dereverberation. In Proceedings of the IEEE International Workshop on Acoustic Signal Enhancement (IWAENC), Xi’an, China, 13–16 September 2016; pp. 1–5. [Google Scholar]
  18. Braun, S.; Habets, E.A.P. Online dereverberation for dynamic scenarios using a Kalman filter with an autoregressive model. IEEE Signal Process. Lett. 2016, 23, 1741–1745. [Google Scholar] [CrossRef]
  19. Dietzen, T.; Doclo, S.; Spriet, A.; Tirry, W.; Moonen, M.; van Waterschoot, T. Low-Complexity Kalman filter for multi-channel linear-prediction-based blind speech dereverberation. In Proceedings of the IEEE Workshop on Applications of Signal Processing to Audio and Acoustics (WASPAA), New Paltz, NY, USA, 15–18 October 2017; pp. 284–288. [Google Scholar]
  20. Kodrasi, I.; Doclo, S. Analysis of eigenvalue decomposition-based late reverberation power spectral density estimation. IEEE/ACM Trans. Audio Speech Lang. Process. 2018, 26, 1106–1118. [Google Scholar] [CrossRef]
  21. Schwartz, O.; Gannot, S.; Habets, E.A.P. Joint estimation of late reverberant and speech power spectral densities in noisy environments using frobenius norm. In Proceedings of the 24th European Signal Processing Conference (EUSIPCO), Budapest, Hungary, 29 August–2 September 2016; pp. 1123–1127. [Google Scholar]
  22. Schwartz, O.; Gannot, S.; Habets, E.A.P. Joint maximum likelihood estimation of late reverberant and speech power spectral density in noisy environments. In Proceedings of the IEEE International Conference on Acoustic, Speech and Signal Processing (ICASSP), Shanghai, China, 20–25 March 2016; pp. 151–155. [Google Scholar]
  23. Kuklasiński, A.; Doclo, S.; Jensen, S.H.; Jensen, J. Maximum likelihood PSD estimation for speech enhancement in reverberation and noise. IEEE/ACM Trans. Audio Speech Lang. Process. 2016, 24, 1599–1612. [Google Scholar] [CrossRef] [Green Version]
  24. Higuchi, T.; Ito, N.; Yoshioka, T.; Nakatani, T. Robust MVDR beamforming using time-frequency masks for online/offline ASR in noise. In Proceedings of the IEEE International Conference on Acoustic, Speech and Signal Processing (ICASSP), Shanghai, China, 20–25 March 2016; pp. 5210–5214. [Google Scholar]
  25. Ito, N.; Araki, S.; Yoshioka, T.; Nakatani, T. Relaxed disjointness based clustering for joint blind source separation and dereverberation. In Proceedings of the IEEE International Workshop on Acoustic Signal Enhancement (IWAENC), Juan-les-Pins, France, 8–11 September 2014; pp. 268–272. [Google Scholar]
  26. Mandel, M.; Ellis, D.; Jebara, T. An EM algorithm for localizing multiple sound sources in reverberant environments. In Advances in Neural Information Processing Systems 19: Proceedings of the 2006 Conference; MIT Press: Cambridge, MA, USA, 2007; pp. 953–960. [Google Scholar]
  27. Pan, C.; Chen, J.; Benesty, J. Performance study of the MVDR beamformer as a function of the source incidence angle. IEEE/ACM Trans. Audio Speech Lang. Process. 2014, 22, 67–79. [Google Scholar] [CrossRef]
  28. Simon, D. Optimal State Estimation: Kalman, H infinity, and Nonlinear Approaches; Wiley: Hoboken, NJ, USA, 2006. [Google Scholar]
  29. Kay, S.M. Fundamentals of Statistical Signal Processing: Estimation Theory, 1st ed.; Prentice-Hall: Englewood Cliffs, NJ, USA, 1993. [Google Scholar]
  30. Haykin, S. Adaptive Filter Theory, 4th ed.; Prentice-Hall: Englewood Cliffs, NJ, USA, 2002. [Google Scholar]
  31. Enzner, G.; Vary, P. Frequency-domain adaptive Kalman filter for coustic echo control in hands-free telephones. Signal Process. 2006, 86, 1140–1156. [Google Scholar] [CrossRef]
  32. Yang, W.; Huang, G.; Chen, J.; Benesty, J.; Cohen, I.; Kellermann, W. Robust dereverberation with kronecker product based multichannel linear prediction. IEEE Signal Process. Lett. 2021, 28, 101–105. [Google Scholar] [CrossRef]
  33. Kinoshita, K.; Delcroix, M.; Kwon, H.; Mori, T.; Nakatani, T. Neural network-based spectrum estimation for online WPE dereverberation. In Proceedings of the Interspeech, Stockholm, Sweden, 20–24 August 2017; pp. 384–388. [Google Scholar]
  34. Cheng, R.; Bao, C.; Cui, Z. Mass: Microphone array speech simulator in room acoustic environment for multi-channel speech coding and enhancement. Appl. Sci. 2020, 10, 1484. [Google Scholar] [CrossRef] [Green Version]
  35. Garofolo, J.S.; Lamel, L.F.; Fisher, W.M.; Fiscus, J.G.; Pallett, D.S. DARP: A TIMIT Acoustic-Phonetic Continous Speech Corpus CD-ROM. NIST Speech Disc 1-1.1; NASA STI/Recon Technical Report N 93, 27403; NASA: Washington, DC, USA, 1993. [Google Scholar]
  36. ITU-T. Perceptual Evaluation of Speech Quality (PESQ): An Objective Method for End-to-End Speech Quality Assessment of Narrowband Telephone Networks and Speech Codecs; ITU-T Recommendation P.862; International Telecommunication Union: Geneva, Switzerland, 2001. [Google Scholar]
  37. Taal, C.H.; Hendriks, R.C.; Heusdens, R.; Jensen, J. An algorithm for intelligibility prediction of time-frequency weighted noisy speech. IEEE/ACM Trans. Audio Speech Lang. Process. 2011, 19, 2125–2136. [Google Scholar] [CrossRef]
  38. Hu, Y.; Loizou, P.C. Evaluation of objective quality measures for speech enhancement. IEEE/ACM Trans. Audio Speech Lang. Process. 2008, 16, 229–238. [Google Scholar] [CrossRef]
  39. Loizou, P.C. Speech Enhancement: Theory and Practice; CRC Press: Boca Raton, FL, USA, 2007. [Google Scholar]
  40. Goetze, S.; Warzybok, A.; Kodrasi, I.; Jungmann, J.O.; Cauchi, B.; Rennies, J.; Habets, E.A.P.; Mertins, A.; Gerkmann, T.; Doclo, S.; et al. A study on speech quality and speech intelligibility measures for quality assessment of single-channel dereverberation algorithms. In Proceedings of the 14th International Workshop on Acoustic Signal Enhancement (IWAENC), Juan-les-Pins, France, 8–11 September 2014; pp. 233–237. [Google Scholar]
  41. Dietzen, T. GitHub Repository: Integrated Sidelobe Cancellation and Linear Prediction Kalman Filter for Joint Multi-microphone Speech Dereverberation, Interfering Speech Cancellation, and Noise Reduction. July 2019. Available online: https://github.com/tdietzen/ISCLP-KF (accessed on 14 January 2021).
  42. Drude, L.; Heymann, J.; Boeddeker, C.; Haeb-Umbach, R. NARA-WPE: A Python package for weighted prediction error dereverberation in Numpy and Tensorflow for online and offline processing. In Proceedings of the 13th ITG-Symposium, Speech communication, Oldenburg, Germany, 10–12 October 2018; pp. 216–220. [Google Scholar]
Figure 1. Block diagram of the proposed dereverberation method.
Figure 1. Block diagram of the proposed dereverberation method.
Applsci 12 11819 g001
Figure 2. Simulation scene of a room considering 20 source locations and 8 microphones [34].
Figure 2. Simulation scene of a room considering 20 source locations and 8 microphones [34].
Applsci 12 11819 g002
Figure 3. Test results related to the filter order L between the proposed method (red line) and ISCLP (blue line).
Figure 3. Test results related to the filter order L between the proposed method (red line) and ISCLP (blue line).
Applsci 12 11819 g003aApplsci 12 11819 g003b
Figure 4. Test results related to T60 on four initialization methods of the PSD of the desired signal with real PSD (green line), [14,15] (black line), proposed (red line), ISCLP (blue line), and [34] (cyan line).
Figure 4. Test results related to T60 on four initialization methods of the PSD of the desired signal with real PSD (green line), [14,15] (black line), proposed (red line), ISCLP (blue line), and [34] (cyan line).
Applsci 12 11819 g004
Figure 5. Spectrogram comparison in the case of SNR = 10 dB and T60 = 800 ms.
Figure 5. Spectrogram comparison in the case of SNR = 10 dB and T60 = 800 ms.
Applsci 12 11819 g005
Table 1. Comparison of the test results at −5 dB for babble noise. (The bold in the table indicates the best result).
Table 1. Comparison of the test results at −5 dB for babble noise. (The bold in the table indicates the best result).
T60 (ms)100400500600700800
PESQ
unprocessed1.121.511.561.581.611.62
ISCLP1.501.581.631.661.691.71
WPE1.121.471.521.531.541.56
proposed1.201.661.721.771.781.79
proposed-kron1.191.681.751.771.781.79
STOI
unprocessed0.510.580.580.580.560.56
ISCLP0.590.570.570.570.570.57
WPE0.500.570.570.580.590.59
proposed0.540.610.620.620.620.62
proposed-kron0.540.610.620.620.620.61
CD (dB)
unprocessed7.045.985.925.895.895.90
ISCLP7.937.917.877.877.877.88
WPE7.005.985.905.865.825.80
proposed7.607.137.067.016.966.93
proposed-kron7.056.326.236.176.136.41
SIR (dB)
unprocessed11.70−4.37−5.52−6.11−6.89−7.03
ISCLP7.02−1.17−1.62−1.13−1.31−0.98
WPE6.34−1.42−2.28−2.97−3.43−3.98
proposed−2.00−2.17−2.73−2.74−2.31−1.90
proposed-kron−3.14−3.12−3.33−3.42−3.11−2.70
fwSegSNR (dB)
unprocessed0.651.932.112.222.302.35
ISCLP1.411.671.631.571.511.46
WPE0.832.022.172.282.352.41
proposed3.164.034.164.244.304.34
proposed-kron3.434.624.794.894.965.01
Table 2. Comparison of the test results at 0 dB for babble noise.
Table 2. Comparison of the test results at 0 dB for babble noise.
T60 (ms)100400500600700800
PESQ
unprocessed1.461.841.871.881.881.87
ISCLP2.021.841.881.901.901.90
WPE1.491.791.831.861.881.91
proposed1.512.012.062.102.112.11
proposed-kron1.452.062.092.102.092.08
STOI
unprocessed0.620.660.650.640.640.63
ISCLP0.710.640.640.640.630.63
WPE0.610.670.670.670.670.67
proposed0.630.690.700.700.690.69
proposed-kron0.640.690.690.680.670.67
CD (dB)
unprocessed6.425.275.275.305.355.40
ISCLP7.167.467.427.437.557.68
WPE6.305.165.105.065.045.13
Proposed6.976.286.326.346.356.37
proposed-kron6.405.385.475.565.665.75
SIR (dB)
unprocessed13.41−3.80−4.80−5.59−6.11−6.51
ISCLP8.81−0.11−0.15−0.38−0.41−0.58
WPE6.69−0.16−1.51−2.34−2.94−3.51
proposed−2.21−1.99−1.15−1.34−1.11−1.18
proposed-kron−3.34−2.08−2.12−2.26−2.12−2.25
fwSegSNR (dB)
unprocessed1.893.333.433.473.473.46
ISCLP3.562.802.652.502.402.30
WPE2.153.563.703.793.843.88
proposed4.565.235.315.335.375.37
proposed-kron4.915.875.966.006.026.01
Table 3. Comparison of the test results at 5 dB for babble noise. (The bold in the table indicates the best result).
Table 3. Comparison of the test results at 5 dB for babble noise. (The bold in the table indicates the best result).
T60 (ms)100400500600700800
PESQ
unprocessed1.852.082.062.032.001.97
ISCLP2.412.412.382.342.302.26
WPE1.902.172.202.232.242.25
proposed2.152.382.392.372.342.31
proposed-kron2.142.372.352.312.282.23
STOI
unprocessed0.730.720.710.690.680.66
ISCLP0.790.760.750.750.740.73
WPE0.720.750.750.750.750.75
proposed0.750.760.750.750.740.74
proposed-kron0.750.750.740.730.720.71
CD (dB)
unprocessed5.575.225.335.455.575.68
ISCLP6.427.077.157.217.287.34
WPE5.394.874.824.804.794.78
proposed6.396.746.686.646.606.59
proposed-kron5.596.016.006.006.006.02
SIR (dB)
unprocessed14.13−5.48−6.53−7.26−7.80−8.21
ISCLP9.38−1.02−1.62−2.03−2.29−2.49
WPE6.76−1.85−2.92−3.75−4.44−5.02
proposed−2.19−6.82−6.83−6.69−6.57−6.41
proposed-kron−3.32−8.03−8.06−7.94−7.87−7.71
fwSegSNR (dB)
unprocessed3.834.714.684.594.494.39
ISCLP5.633.853.343.163.022.88
WPE4.135.145.245.295.315.32
proposed4.124.254.244.234.204.14
proposed-kron4.634.944.944.904.844.76
Table 4. Comparison of the test results at 10 dB for babble noise. (The bold in the table indicates the best result).
Table 4. Comparison of the test results at 10 dB for babble noise. (The bold in the table indicates the best result).
T60 (ms)100400500600700800
PESQ
unprocessed2.242.322.242.182.121.97
ISCLP2.732.592.532.472.422.36
WPE2.302.512.542.552.562.55
proposed2.542.612.582.532.482.45
proposed-kron2.512.582.522.452.392.34
STOI
unprocessed0.820.760.740.720.700.69
ISCLP0.850.790.780.770.760.75
WPE0.810.800.800.800.800.80
proposed0.810.790.790.780.770.77
proposed-kron0.810.780.770.760.740.73
CD (dB)
unprocessed4.644.785.005.215.385.53
ISCLP5.806.736.846.947.027.10
WPE4.414.204.194.204.214.21
proposed5.646.296.276.266.256.24
proposed-kron5.325.715.755.795.825.85
SIR (dB)
unprocessed14.19−5.53−6.67−7.32−7.81−8.62
ISCLP9.55−0.98−1.57−1.96−2.21−2.42
WPE6.78−1.90−2.99−3.84−4.54−5.13
proposed−1.91−6.79−6.80−6.67−6.49−6.38
proposed-kron−3.03−8.05−8.09−7.92−7.82−7.73
fwSegSNR (dB)
unprocessed6.395.865.655.435.225.04
ISCLP7.334.143.843.633.453.28
WPE6.556.536.566.556.546.51
proposed5.865.245.135.044.944.86
proposed-kron6.275.765.645.535.405.27
Table 5. Complexity comparison.
Table 5. Complexity comparison.
Proposed-Kron
Step(×)(+)(÷)
1(2L − 1) (L − 1)22(L − 1)3 − (L − 1)——
2M (L − 1)2 (M + 1) + 2 (L − 1)2 + 2 (L − 1)(M2 + 2) (L − 1)2 − M (L − 1)L − 1
3(L − 1) + (L − 1)2 + (L − 1)3(L − 1)3 + (L − 1)——
Proposed
Step(×)(+)(÷)
12 M3 (L − 1)3 + M2 (L − 1)22 M3(L − 1)3 − M(L − 1)——
22 M2 (L − 1)2 + 2 M (L − 1)2 M2(L − 1)2M (L − 1)
3M (L − 1) + M2 (L − 1)2 + M3 (L − 1)3M (L − 1)3 + (L − 1)——
Complexity Reduction Factor
Stepλ (×)λ (+)λ (÷)
11/M31/M3——
21/2 + 1/2 M1/2 + 1/M21/M
31/M31/M——
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Tan, F.; Bao, C.; Zhou, J. Effective Dereverberation with a Lower Complexity at Presence of the Noise. Appl. Sci. 2022, 12, 11819. https://doi.org/10.3390/app122211819

AMA Style

Tan F, Bao C, Zhou J. Effective Dereverberation with a Lower Complexity at Presence of the Noise. Applied Sciences. 2022; 12(22):11819. https://doi.org/10.3390/app122211819

Chicago/Turabian Style

Tan, Fengqi, Changchun Bao, and Jing Zhou. 2022. "Effective Dereverberation with a Lower Complexity at Presence of the Noise" Applied Sciences 12, no. 22: 11819. https://doi.org/10.3390/app122211819

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop