Next Article in Journal
A Novel Photovoltaic Power Prediction Method Based on a Long Short-Term Memory Network Optimized by an Improved Sparrow Search Algorithm
Previous Article in Journal
Steel Surface Defect Detection Algorithm Based on YOLOv8
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Convex Regularized Recursive Minimum Error Entropy Algorithm

School of Physics and Electronic Information, Yantai University, Yantai 264005, China
*
Author to whom correspondence should be addressed.
Electronics 2024, 13(5), 992; https://doi.org/10.3390/electronics13050992
Submission received: 26 January 2024 / Revised: 29 February 2024 / Accepted: 3 March 2024 / Published: 6 March 2024

Abstract

:
It is well known that the recursive least squares (RLS) algorithm is renowned for its rapid convergence and excellent tracking capability. However, its performance is significantly compromised when the system is sparse or when the input signals are contaminated by impulse noise. Therefore, in this paper, the minimum error entropy (MEE) criterion is introduced into the cost function of the RLS algorithm in this paper, with the aim of counteracting the interference from impulse noise. To address the sparse characteristics of the system, we employ a universally applicable convex function to regularize the cost function. The resulting new algorithm is named the convex regularization recursive minimum error entropy (CR-RMEE) algorithm. Simulation results indicate that the performance of the CR-RMEE algorithm surpasses that of other similar algorithms, and the new algorithm excels not only in scenarios with sparse systems but also demonstrates strong robustness against pulse noise.

1. Introduction

Sparse adaptive filters have widespread applications in the field of signal processing, commonly employed in tasks such as acoustic echo cancellation [1,2] and communication channel identification [3]. Traditional adaptive filtering algorithms used for identifying sparse systems (referring to systems that contain many near-zero or a few large coefficients that need to be identified) often rely on prior knowledge of sparsity. In particular, many relevant algorithms have been proposed in this context: for the LMS algorithm, scholars have successively improved its cost function using the l0-norm and l1-norm regularization techniques [4]. This result significantly enhances the algorithm’s convergence speed in sparse systems, reduces steady-state errors, and simultaneously improves its robustness. Subsequently, other researchers utilized a more general convex function to regularize the LMS algorithm [5], making the proposed algorithm more versatile compared to the former. In [6], the author applied convex regularization techniques to enhance the RLS algorithm. By incorporating a general regularization term into the RLS cost function, they derived a new type of sparse adaptive filtering algorithm, referred to as the convex regularized recursive least squares (CR-RLS) algorithm. This approach allows the use of any convex function for regularization. In recent years, the concept of sparsity has even been applied to kernel adaptive filtering, as mentioned in [7,8,9,10].
The aforementioned adaptive algorithms based on the mean-square error criterion perform well in Gaussian noise environments. However, in real-life scenarios, signals may be contaminated by impulsive noise, and the mean-square error criterion may not yield satisfactory results in the presence of impulsive noise [11]. With the development of information theoretic learning (ITL), the maximum correntropy criterion (MCC) [12,13,14,15] and the minimum error entropy criterion (MEE) [16,17,18] have gradually been introduced. Due to their calculations involving higher-order statistical properties of the error signal, they exhibit less sensitivity to significant outliers, thus demonstrating a strong capability to effectively handle impulsive noise. Especially in non-Gaussian environments, information-theoretic criteria are notably superior to the most popular mean square error criterion. Furthermore, the literature indicates that under heavy-tailed noise distributions or when signals are contaminated by impulse noise or outliers, the MCC criterion and MEE criterion are significantly superior to the mean square error criterion.
In the present day, the maximum correlation entropy criterion and the minimum error entropy criterion have been successfully applied in adaptive filtering algorithms. For instance, a classic example is found in [19], where the correntropy-induced metric (CIM) penalized MCC criterion is utilized instead of the traditional minimum mean square error criterion for robust channel estimation against impulse noise. MCC is employed to combat impulse noise, while CIM is utilized to effectively exploit the sparsity of the channel. Hence, the proposed CIM-MCC algorithm demonstrates excellent performance. Furthermore, researchers have introduced the MCC criterion in kernel adaptive filtering [13], leading to the derivation of the kernel recursive maximum correntropy (KRMC) algorithm. Subsequent simulation results in alpha-stable noise environments for short-term chaotic time series prediction confirm the superior performance of KRMC. Similarly, the recursive minimum error entropy (RMEE) algorithm is proposed [20]. These algorithms exhibit strong suppression capabilities against pulse noise. Considering scenarios involving sparse systems, scholars have also proposed the convex regularized recursive maximum correntropy (CR-RMC) algorithm as mentioned in [21]. Inspired by the previous discussions, this paper introduces sparsity enhancements to the RMEE algorithm by incorporating a more general convex function as a sparse penalty term in the cost function. As a result, we obtain a sparse adaptive algorithm based on the MEE criterion, which is called the convex regularization recursive minimum error entropy (CR-RMEE) algorithm. Subsequent results reveal that this algorithm not only exhibits strong robustness in the presence of impulsive noise but also demonstrates good capabilities in identifying sparse systems.
The remaining part of the paper is organized as follows: In Section 2, relevant background knowledge is introduced, including basic concepts of adaptive filtering and minimum error entropy. We provide a detailed explanation of the proposed new algorithm in Section 3, including the derivation process and the selection of regularization factors. In Section 4, we conduct simulation experiments on the algorithm, considering both stationary and non-stationary systems and various noise environments, to test the algorithm’s convergence, tracking performance, and robustness. An actual scenario of echo cancellation is also considered to thoroughly evaluate the algorithm’s performance. Finally, the conclusion is presented in Section 5.

2. Basic Principles

When performing linear adaptive filtering, there is an input vector u, unknown tap parameter w, and desired response d. The desired response d(n) at each regression point can be obtained with the following expression:
d n = w o T u n + v n , n = 1 , 2 ,
where v is a zero-mean background noise with a variance of σ v 2 . In this context, the error signal can be represented as follows:
e n = d n w T n 1 u n
where w(n − 1) is the estimate of wo at point n − 1.
Additionally, in this study, we have made the following assumptions:
(a)
The input signal is a white sequence with a mean of zero and is wide-sense stationary. The additive noise is zero-mean.
(b)
The input sequences at different time instants and the noise are uncorrelated.
(c)
The inputs at different time instants and the additive noise are uncorrelated.
E u m u T n = σ u 2 I , m = n , 0 , m n ,
E u T m u n = M σ u 2 , m = n , 0 , m n ,
E v m v n = σ v 2 , m = n , 0 , m n .
E u H m v n = 0 , m n .
As a well-known learning criterion in Information Theoretic Learning (ITL), the Minimum Error Entropy was proposed by Alfred Renyi, building upon the foundation of Shannon entropy. Its expression can be derived from the Shannon entropy formula:
H α X = 1 1 α log i = 1 N p i α = 1 1 α log V α X
where α is a the order parameter, and you can change its value to alter the form of entropy expression. p i represents the probability of occurrence of the i t h data set; V α X is referred to as the Information Potential (IP), and the second-order Renyi entropy with α = 2 is more commonly used as follows:
V 2 X = E k σ 2 X
where κ σ X is the Gaussian kernel with bandwidth σ . Kernel functions primarily describe the inner product of random variables in the feature space. The Gaussian kernel function is commonly used, and its expression is as follows:
κ σ x x o = 1 2 π σ exp x x o 2 σ 2
Due to the difficulty in obtaining the Probability Density Function (PDF) of a signal accurately in practical situations, it is common to estimate the PDF. The most common method for PDF estimation is based on the Parzen method. Assuming the input vector data are u 1 , u 2 , , u N , the probability density after Parzen estimation can be represented as follows:
p ^ u , σ u = 1 N i = 1 N κ σ u u i
The estimate of its entropy can be derived by incorporating the kernel function into the second-order Renyi entropy formula:
H ^ 2 U = log 1 N 2 i = 1 N j = 1 N κ 2 σ u i u j
where u i and u i represent input signals at different time instants i and j . When replacing the input in the kernel function with the error e , the error entropy H ^ 2 e of the system can be defined.
From the derived error entropy formula above, based on the definition of error entropy, it can be concluded that when the error entropy is minimized, the system error signal becomes more stable. This implies that the system’s filtering algorithm has achieved convergence. Therefore, the cost function for algorithms based on the minimum error entropy criterion can be expressed as follows:
J M E E w = min H ^ 2 e   = min log 1 N 2 i = 1 N j = 1 N κ 2 σ e i e j
Because this function is a logarithmic function and monotonically increasing, the cost formula can be simplified to
J M E E w = max V ^ 2 , σ e   = max 1 N 2 i = 1 N j = 1 N κ 2 σ e i e j
where w represents the tap parameters of the adaptive filter, and V ^ 2 , σ e represents the estimate of the second-order information potential.

3. Proposed Algorithm

In this chapter, we first proposed a new cost function and derived the novel CR-RMEE algorithm based on it. We also introduced the updated formula for the regularization factor. Finally, towards the end of this chapter, we conducted a mathematical analysis of the convergence of the new algorithm, demonstrating its theoretical convergence.

3.1. CR-RMEE Algorithm

The traditional MEE cost function, as shown in (11), leads to the cost function for the exponential forgetting factor 0 < λ < 1 based on the MEE criterion as follows:
J w = 1 L 2 i = 0 L 1 j = 0 L 1 λ i + j G σ e i e j
where L is the length of the sliding window inside the filter, G σ · is the Gaussian kernel function, and σ is the kernel width.
We incorporate a convex function containing the weight vector into it for regularization, which allows the regularized cost function to leverage some prior knowledge about the unknown system, such as sparsity. In this paper, the regularized MEE cost function is as follows:
J w = 1 L 2 i = 0 L 1 j = 0 L 1 λ i + j G σ e i e j γ n f w
where f · is a convex function, γ n is a variable constant known as the regularization factor. The role of γ n is to maintain a balance between the MEE cost function and the regularization term. When γ n increases or decreases, the impact of the regularization term on the adaptive algorithm also changes accordingly.
There are subgradients for the regularization convex function in the cost function. The collection of subgradients is called subdifferential, which is denoted by f w . We denote a subgradient vector of f · at n with s f n f n .
Therefore, the gradient of the cost function (13) can be represented as follows:
J w = 1 L 2 σ 2 i = 0 L 1 j = 0 L 1 λ i + j u i u j G σ e i e j e i e j γ n s f w
To simplify the computation, at each time point n (the following derivations of equations are all under time point n) we have
u i = u n i , d i = d n i , e i = e n i , i = 0 , 1 , , L 1 .
Then, we can obtain
U L = u 0 , u 1 , , u L 1 = u 0 , U L 1 , D L = d 0 , d 1 , , d L 1 T = d 0 , D L 1 T , ε L = e 0 , e 1 , , e L 1 T = e 0 , ε L 1 T = D L U L T w .
So, the gradient of the cost function (14) can be rewritten as follows:
J w = 2 L 2 σ 2 U L P L Q L ε L γ n s f w = 2 L 2 σ 2 U L Φ L ε L γ n s f w
where
Q i + 1 j + 1 = λ i + j G σ e i e j , i , j = 0 , 1 , , L 1 , P i + 1 j + 1 = k = 0 L 1 λ i + j G σ e i e k , i = j , 0 , i j , Φ L = P L Q L = ϕ 0 φ T φ λ 2 Φ L 1 , φ i = λ i G σ e i e 0 , j = 0 , 1 , , L 1 ϕ 0 = k = 1 L 1 λ k G σ e 0 e k .
Setting the gradient equal to zero, we can solve for
w = U L Φ L U L T 1 U L Φ L D L 1 2 L 2 σ 2 γ n s f w
Let
R L = U L Φ L U L T r L = U L Φ L D L
Then, (19) can be rewritten as follows:
w = R L 1 r L 1 2 L 2 σ 2 γ L s f w
where R L is the weighted autocorrelation matrix of the input signal and can be rewritten as
R L = λ 2 U L 1 Φ L 1 U L 1 T + u 0 ϕ 0 u 0 T + U L 1 φ u 0 T + u 0 φ U L 1 T
r L is the weighted cross-correlation vector between the input signal and the desired response and can be rewritten as
r L = λ 2 U L 1 Φ L 1 D L 1 + u 0 ϕ 0 d 0 + U L 1 φ d 0 + u 0 φ D L 1 T
Under the conditions of Assumption (b) and Assumption (c), the following expectations have the following properties:
E U L 1 φ u 0 T + u 0 φ U L 1 T = 0 E U L 1 φ d 0 + u 0 φ D L 1 T = 0
So, (22) and (23) can be rewritten as follows:
R L = λ 2 U L 1 Φ L 1 U L 1 T + u 0 ϕ 0 u 0 T   = λ 2 R L 1 + u 0 ϕ 0 u 0 T
r L = λ 2 U L 1 Φ L 1 D L 1 + u 0 ϕ 0 d 0   = λ 2 r L 1 + u 0 ϕ 0 d 0
Define a new vector θ L :
θ L = r L 1 2 L 2 σ 2 γ L s f w
By substituting the update (26) into (27), assuming that γ L and s f w change only slightly at each time point [6], the vector θ L can also be recursively updated as follows:
θ L = λ 2 r L 1 + u 0 ϕ 0 d 0 1 2 L 2 σ 2 γ L s f w   = λ 2 r L 1 1 2 L 2 σ 2 γ L 1 s f w + u 0 ϕ 0 d 0   1 2 L 2 σ 2 γ L s f w + 1 2 λ 2 L 2 σ 2 γ L 1 s f w   λ 2 θ L 1 + u 0 ϕ 0 d 0 1 2 1 λ 2 L 2 σ 2 γ L 1 s f w
Then, (21) can be rewritten as
w = P L θ L
where P L = R L 1 . By using the matrix inversion lemma [20], we can obtain the updated formula
P L = λ 2 P L 1 k L u o T P L 1
The gain matrix k L is given by
k L = P L 1 u o λ 2 ϕ 0 + u o T P L 1 u o 1
Rewriting (29) recursively, we can obtain the following updated formula:
w L = w L 1 + k L e 0 1 2 L 2 σ 2 1 λ 2 γ P L 1 s f w

3.2. Selection of the Regularization Factor

The use of regularization functions in the field of adaptive filtering is because they can provide some prior knowledge about the unknown system. Here, we assume that the prior knowledge is an upper bound on the regularization function, i.e., f w ρ , where ρ is a constant. Let w ^ n be the solution obtained from the cost function without regularization, and ε = w w o and ε ^ = w ^ w o represent the differences between w and w ^ with respect to the true weights. Substituting the difference expressions into (19), we obtain
ε n = ε ^ n 1 2 L 2 σ 2 P n γ n s f w n
The squared deviation of ε n from D n can be computed as follows:
D n = ε T n ε n = ε n 2 2 = D ^ n L 2 σ 2 P n γ n s f w n ε ^ n   + 1 4 L 4 σ 4 γ 2 n P n s f w n 2 2
where D ^ n is the squared deviation from ε ^ n . One can obtain the following theorem from the (34) [6].
Theorem 1.
D n D ^ n  if  γ n 0 , max γ ^ n , 0 , where
γ ^ n = 4 s f w n P n ε ^ n L 2 σ 2 P n s f w n 2 2
Proof. 
From (34), when 1 4 L 4 σ 4 γ 2 n P n γ n s f w n 2 2 L 2 σ 2 P n γ n s f w n ε ^ n 0 , then we obtain D n D ^ n . Consequently, the inequality can be solved as follows:
0 γ n 4 s f w n P n ε ^ n L 2 σ 2 P n s f w n 2 2 , s f w n P n ε ^ n 0 γ n = 0 , s f w n P n ε ^ n < 0
From Theorem 1, it can be seen that if the method described in (36) is used to adaptively update the regularization factor, the performance of the CR-RMEE algorithm will be close to or even surpass the standard RMEE algorithm. However, since in practical use, we cannot obtain the true weight vector, γ ^ n is also not easy to calculate directly. Inspired by the work in [6], and using the subgradient definition of a convex function, γ ^ n can be approximated by the following expression when the input is a white sequence and n is sufficiently large:
γ n γ n = 4 t r P n N f w n ρ + s f w n T P n ε n L 2 σ 2 P n s f w n 2 2
where ε n = w ^ n w n , meaning the difference between the unregularized coefficients and the regularized coefficients, and t r · represents a tracking operator. According to (38), γ n can be automatically updated to γ n = max γ n , 0 in each iteration, and the computational complexity for each automatic update is only O N 2 , which effectively saves time in the search for the optimal value. □

3.3. Convergence Analysis

Whether an algorithm converges can be determined by observing the behavior when it reaches a steady state characterized by E w ˜ n , where w ˜ n = w o w n . So, from Equation (32), it can be concluded that
w ˜ n = w ˜ n 1 k n e n + 1 2 L 2 σ 2 1 λ 2 γ n 1 P n 1 s f w
Substituting Equation (1) into it, we obtain
w ˜ n = w ˜ n 1 k n u n w ˜ n 1 k n v n   + 1 2 L 2 σ 2 1 λ 2 γ n 1 P n 1 s f w
Substituting (30) and (31) into the previous expression yields
w ˜ n = λ 2 P n R n 1 w ˜ n 1 P n u n ϕ n v n   + 1 2 L 2 σ 2 1 λ 2 γ n 1 P n 1 s f w
From (25), the steady-state average value of A can be obtained as follows:
E R n = λ 2 E R n 1 + E ϕ n E u n u T n
Assuming the input signal u n is a wide-sense stationary process, combined with the assumptions mentioned earlier (a) and (41), we can obtain
E R n = λ 2 E R n 1 + E ϕ n σ u 2 I
where I is the unit matrix of the same dimension as R n . Summarizing the above expressions, we obtain
E R n = 1 λ 2 1 E ϕ n σ u 2 I
Although it is known from the above that P n = R n 1 , calculating the steady-state average of P n is still very challenging. Therefore, it is commonly approximated using the following expression [19].
E P n E 1 P 1 n = 1 λ 2 E 1 ϕ n σ u 2 I
From (37), it can be easily deduced that lim n γ n = 0 , and under the conditions of assumption (a), E v n = 0 can also be obtained. Therefore, the steady-state value of w ˜ n can be expressed as
E w ˜ n = λ 2 E w ˜ n 1
For 0 < λ < 1 , it follows that
lim n E w ˜ n = 0
Therefore, we can demonstrate that the CR-RMEE algorithm proposed by us possesses convergence.

4. Simulation Experiments

In order to compare the advantages of the proposed algorithm, our simulations were conducted separately in steady and non-steady environments, and all simulation results were obtained through 1000 independent Monte Carlo experiments. The simulations use normalized mean squared deviation (NMSD) as the performance metric to evaluate the algorithms. A smaller value of NMSD indicates better algorithm performance, and the stability of the resulting waveform is also an important indicator of algorithm stability. NMSD can be obtained using the following formula:
N M S D n = 10 log 10 w n w o 2 w o 2
In the simulations, the unknown weight vector w o is a 20 × 1 vector with S non-zero tap weights, and the sliding window width is 8. Unless otherwise stated, the input u is a white Gaussian random sequence, and the additive noise is impulsive with a mixed Gaussian distribution, v(n) ~ 0.95N(0,0.01) + 0.05N(0,25).
For the selection of the function in the regularization term, we will introduce a sparsity-inducing convex penalty function. Since the l0 norm is a non-convex function, we will use its approximate function instead as follows:
w 0 f β w = k = 0 N 1 1 e β w k
where β is a constant, and the approximate gradient of the above expression can be represented using the following formula:
s f β w k β sgn w k β 2 w k , w k 1 β 0 , e l s e w h e r e
Firstly, we present the performance of the CR-RMEE algorithm in a stationary sparse environment. For comparison, we also present the performance of the LMS algorithm, RLS algorithm, RMC algorithm, and the original RMEE algorithm. We set the unknown system as a sparse system with S = 1. For the RMC algorithm, RMEE algorithm, CR-RMC algorithm, and our proposed CR-RMEE algorithm, the same σ = 1 is chosen. Moreover, the step size for the LMS algorithm is 0.01. As for the choice of λ , RLS, RMC, RMEE, CR-RMC, and CR-RMEE algorithms are, respectively, chosen as λ = 0.992 , λ = 0.942 , λ 2 = 0.996 , λ 2 = 0.996 , and λ 2 = 0.996 . As we can see from Figure 1, it is evident that, with the initial convergence rates of various algorithms set to be similar, the NMSD of the LMS, RLS, and RMC algorithms is relatively large. Additionally, both the LMS and RMC algorithms exhibit significant fluctuations after convergence. Although both the RMEE and CR-RMC algorithms have very small NMSD values, the RMEE algorithm exhibits faster convergence, while the CR-RMC algorithm ultimately achieves a lower steady-state error. Compared to the LMS and RMC algorithms, our proposed CR-RMEE algorithm shows significantly reduced fluctuations within an acceptable range. Combining the advantages of the fast convergence of the RMEE algorithm and the low steady-state error of the CR-RMC algorithm, the CR-RMEE algorithm even achieves the lowest NMSD. This simulation experiment demonstrates that our proposed new algorithm exhibits excellent performance in a steady sparse environment.
Secondly, to demonstrate the convergence and tracking performance of the CR-RMEE algorithm in a non-stationary sparse environment, we identify another three-stage sparse system. On the other hand, to compare the performance of each algorithm in different sparse environments, we assume the parameter vector of the unknown system as follows:
w o = 0 ,   0 ,   0 ,   0 ,   0 ,   0 ,   0 ,   0 ,   0 ,   1 ,   0 ,   0 ,   0 ,   0 ,   0 ,   0 ,   0 ,   0 ,   0 ,   0 , n 2000 1 ,   0 ,   1 ,   0 ,   1 ,   0 ,   1 ,   0 ,   1 ,   0 ,   1 ,   0 ,   1 ,   0 ,   1 , 0 ,   1 ,   0 ,   1 ,   0 , 2000 < n 4000 1 , 1 ,   1 , 1 ,   1 , 1 ,   1 , 1 ,   1 , 1 ,   1 , 1 ,   1 , 1 ,   1 , 1 ,   1 , 1 ,   1 , 1 , 4000 < n
From the above equation, it can be seen that as S increases from “1” to “10” and finally to “20”, the sparsity of the system decreases over time. It is worth noting that due to the fact that the system is re-identified each time the algorithm detects a sudden change in the system, there will be a process of convergence again. And by changing the system parameters twice during the simulation, the subsequent convergence curves will also produce a three-step process. For LMS, RLS, RMS, and RMEE algorithms, the parameters are kept the same as before, and for the CR-RMC and CR-RMEE algorithms, γ is set to 0.1. From Figure 2, it can be seen that the performance of each algorithm in each sparse phase follows the same trend as in Figure 1. Due to not considering the impact of sparse systems, the LMS algorithm, RLS algorithm, RMC algorithm, and RMEE algorithm are less affected by the sparsity of the environment. Similarly, both CR-RMEE and CR-RMC are influenced by sparsity. When the sparsity level is high, the CR-RMEE algorithm has a significant advantage compared to other algorithms. Although the advantage decreases slightly as the sparsity decreases, regardless of how sparse the system is, the normalized mean squared deviation (NMSD) of the CR-RMEE algorithm remains the lowest. The system undergo two parameter changes in total, and our proposed algorithm responded quickly and converged rapidly to a steady state, which indicates that the new algorithm also possesses excellent tracking performance. In summary, compared to the other algorithms, our algorithm exhibits the lowest steady-state error and excellent performance.
Thirdly, we vary the parameter (σ, λ) for the CR-RMEE algorithm to test its impact on algorithm performance. All simulation results are the average of 200 independent runs, so we can only obtain an approximate value. The experimental results are shown in Table 1. To more intuitively illustrate the effect of different parameters on algorithm performance, we have selected a set of obvious results in graphical form, as shown in Figure 3. The parameters for CR-RMEE1, CR-RMEE2, CR-RMEE3, CR-RMEE4, and CR-RMEE5 are [σ = 3, λ2 = 0.942], [σ = 2, λ2 = 0.970], [σ = 1, λ2 = 0.980], [σ = 0.5, λ2 = 0.992], and [σ = 0.2, λ2 = 0.994]. Combining Table 1 and Figure 3, it can be observed that the NMSD of the algorithm is negatively correlated with the σ value and positively correlated with the λ2 value. As the steady-state error decreases, the convergence speed of the algorithm gradually slows down, and the level of post-convergence oscillation also gradually increases. This indicates that achieving a lower steady-state error often requires sacrificing the algorithm’s convergence speed and stability. Therefore, in practical applications, it is necessary to adjust the parameters judiciously to achieve the best performance. Additionally, from Table 1, it can be seen that the steady-state error of the CR-RMEE algorithm in Figure 1 is approximately −49.6 dB when [σ = 1, λ2 = 0.996].
Then, although the algorithms were tested separately in both steady and non-steady environments in the above experiments, demonstrating good tracking and convergence performance under certain conditions, the testing for algorithm robustness remains incomplete. Therefore, we continue to test the algorithm and comparative algorithms in four different noise environments. The four types of noise are Gaussian white noise with a distribution of v(n) ~ N(0,0.25), pulse noise with a distribution of v(n) ~ 0.95N(0,0.01) + 0.05N(0,1), pulse noise with a distribution of v(n) ~ 0.95N(0,0.01) + 0.05N(0,9), and pulse noise with a distribution of v(n) ~ 0.95N(0,0.01) + 0.05N(0,25), with an increasing intensity of pulse noise. These correspond, respectively, to Figure 4a–d. The selection of parameters for each algorithm aims to maintain optimal performance as much as possible. We set the unknown system as a sparse system, with the same choice of parameters for RMC algorithm, RMEE algorithm, CR-RMC algorithm, and the proposed CR-RMEE algorithm. The step size for the LMS algorithm is set to 0.01. The choices for λ are, respectively, λ = 0.992 , λ = 0.942 , λ 2 = 0.996 , λ 2 = 0.996 , and λ 2 = 0.996 .
The results of testing under the four noise conditions can be seen from Figure 4. Firstly, as the intensity of noise pulses increases, the gap between different algorithms also becomes larger (convergence plots initially cluster together but then become dispersed). This indicates that different algorithms have varying degrees of sensitivity to noise, with RMC algorithm being the most affected. While classic algorithms such as LMS and RLS exhibit larger steady-state errors, they demonstrate strong stability. Among the algorithms tested under different noise conditions, the RMEE algorithm, CR-RMC algorithm, and our proposed CR-RMEE algorithm perform the best. They exhibit not only fast convergence rates in different environments but also low steady-state errors. It is noted that, among these three algorithms, our proposed new algorithm performs the best in all four scenarios, which strongly proves the robustness of the algorithm against noise.
Finally, to demonstrate the performance of the CR-RMEE algorithm in practical applications, we consider an echo cancellation application, whose basic structure is as shown in Figure 5. The generation principle of acoustic echo is mainly due to the sound from room A being transmitted to room B through channels such as wired or wireless, played through a speaker. After a series of acoustic reflections, the sound carries the inherent noise in room B and is picked up by the microphone, then transmitted back to room A. Acoustic echo can cause speakers in room A to hear their own speech shortly after speaking, significantly impacting the communication quality. In such cases, echo cancellation aims to eliminate these effects as much as possible. The mathematical model of echo cancellation is as follows: the far-end signal u(i) from room A, when transmitted to room B, generates the echo signal r(i) after multiple reflections. Finally, the near-end signal, carrying the echo signal r(i) and the noise v(i) from room B, is combined to produce d(i). The Acoustic Echo Canceller (AEC) module processes the far-end signal u(i) to obtain y(i) (the echo signal). By taking the difference between d(i) and y(i), the error e(i) is obtained and passed to the adaptive filter (AEC module), leading to iterative updates of the filter coefficients. A smaller error indicates better echo cancellation effectiveness.
In this simulation, the input was changed to a real speech signal with a sampling rate of 8 kHz, as shown in Figure 6. The additive noise is impulsive with a mixed Gaussian distribution, v(n) ~ 0.95N(0,0.01) + 0.05N(0,5). The parameters used for each algorithm were the same as in Figure 1. From Figure 7, it can be observed that, when the input is real speech, the convergence curves of various algorithms exhibit noticeable fluctuations, especially the LMS algorithm. Although the RMC and RMEE algorithms also have relatively small steady-state errors, the fluctuations are still evident. Our proposed CR-RMEE algorithm not only has smaller fluctuations but also the minimum steady-state error, indicating its superior performance.

5. Conclusions

This paper introduces a Convex Regularized Recursive Minimum Error Entropy algorithm (CR-RMEE) to address the challenges faced by traditional adaptive filtering algorithms in sparse environments, such as poor recognition capability and susceptibility to impulse noise. Building upon the RLS algorithm, the CR-RMEE algorithm incorporates a series of improvements tailored to pulse noise and sparse environments. By introducing the Minimum Error Entropy (MEE) criterion into the cost function, this algorithm enhances resistance to impulse noise. Furthermore, regularization based on general convex functions is applied to the cost function, enabling the algorithm to leverage prior knowledge of sparsity, significantly improving its performance in sparse environments. Simulation results demonstrate that, in sparse system identification, the CR-RMEE algorithm outperforms the original RMEE algorithm. The new algorithm exhibits robustness in the presence of impulse noise and can adapt to system sparsity using prior knowledge. In future work, the application of the CR-RMEE algorithm may extend to complex-valued filtering and even nonlinear kernel filtering.

Author Contributions

Methodology, X.W. and S.O.; validation, Y.G. and X.W.; investigation, X.W. and S.O.; writing—original draft preparation, X.W., Y.G. and S.O.; writing—review and editing, X.W., S.O. and Y.G. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the Shandong Provincial Natural Science Foundation under Grant ZR2022MF314.

Data Availability Statement

Data are contained within the article.

Acknowledgments

The authors would like to thank all reviewers for their helpful comments and suggestions on this paper.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Zhao, H.; Gao, Y.; Zhu, Y. Robust Subband Adaptive Filter Algorithms-Based Mixture Correntropy and Application to Acoustic Echo Cancellation. IEEE/ACM Trans. Audio Speech Lang. Process. 2023, 31, 1223–1233. [Google Scholar] [CrossRef]
  2. Pauline, S.H.; Samiappan, D.; Kumar, R.; Anand, A.; Kar, A. Variable Tap-Length Non-Parametric Variable Step-Size NLMS Adaptive Filtering Algorithm for Acoustic Echo Cancellation. Appl. Acoust. 2020, 159, 107074. [Google Scholar] [CrossRef]
  3. Carbonelli, C.; Vedantam, S.; Mitra, U. Sparse Channel Estimation with Zero Tap Detection. IEEE Trans. Wirel. Commun. 2007, 6, 1743–1763. [Google Scholar] [CrossRef]
  4. Gu, Y.; Jin, J.; Mei, S. L0 Norm Constraint LMS Algorithm for Sparse System Identification. IEEE Signal Process. Lett. 2009, 16, 774–777. [Google Scholar] [CrossRef]
  5. Chen, Y.; Gu, Y.; Hero, A.O. Sparse LMS for System Identification. In Proceedings of the 2009 IEEE International Conference on Acoustics, Speech and Signal Processing, Taipei, Taiwan, 19–24 April 2009; pp. 3125–3128. [Google Scholar]
  6. Eksioglu, E.M.; Tanc, A.K. RLS Algorithm with Convex Regularization. IEEE Signal Process. Lett. 2011, 18, 470–473. [Google Scholar] [CrossRef]
  7. Chen, B.; Zhao, S.; Zhu, P.; Principe, J.C. Quantized Kernel Least Mean Square Algorithm. IEEE Trans. Neural Netw. Learn. Syst. 2012, 23, 22–32. [Google Scholar] [CrossRef] [PubMed]
  8. Chen, B.; Zhao, S.; Zhu, P.; Príncipe, J.C. Quantized Kernel Recursive Least Squares Algorithm. IEEE Trans. Neural Netw. Learn. Syst. 2013, 24, 1484–1491. [Google Scholar] [CrossRef] [PubMed]
  9. Chen, B.; Zhao, S.; Seth, S.; Principe, J.C. Online Efficient Learning with Quantized KLMS and L1 Regularization. In Proceedings of the 2012 International Joint Conference on Neural Networks (IJCNN), Brisbane, Australia, 10–15 June 2012; pp. 1–6. [Google Scholar]
  10. Wang, G.; Xue, R.; Wang, J. A Distributed Maximum Correntropy Kalman Filter. Signal Process. 2019, 160, 247–251. [Google Scholar] [CrossRef]
  11. Heravi, A.R.; Hodtani, G.A. A New Information Theoretic Relation Between Minimum Error Entropy and Maximum Correntropy. IEEE Signal Process. Lett. 2018, 25, 921–925. [Google Scholar] [CrossRef]
  12. Radmanesh, H.; Hajiabadi, M. Recursive Maximum Correntropy Learning Algorithm with Adaptive Kernel Size. IEEE Trans. Circuits Syst. II Express Briefs 2018, 65, 958–962. [Google Scholar] [CrossRef]
  13. Wu, Z.; Shi, J.; Zhang, X.; Ma, W.; Chen, B. Kernel Recursive Maximum Correntropy. Signal Process. 2015, 117, 11–16. [Google Scholar] [CrossRef]
  14. Chen, B.; Liu, X.; Zhao, H.; Principe, J.C. Maximum Correntropy Kalman Filter. Automatica 2017, 76, 70–77. [Google Scholar] [CrossRef]
  15. Chen, B.; Zhu, Y.; Hu, J. Mean-Square Convergence Analysis of ADALINE Training with Minimum Error Entropy Criterion. IEEE Trans. Neural Netw. 2010, 21, 1168–1179. [Google Scholar] [CrossRef] [PubMed]
  16. Shen, P.; Li, C. Minimum Total Error Entropy Method for Parameter Estimation. IEEE Trans. Signal Process. 2015, 63, 4079–4090. [Google Scholar] [CrossRef]
  17. Chen, B.; Dang, L.; Gu, Y.; Zheng, N.; Príncipe, J.C. Minimum Error Entropy Kalman Filter. IEEE Trans. Syst. Man Cybern. Syst. 2021, 51, 5819–5829. [Google Scholar] [CrossRef]
  18. Li, Z.; Xing, L.; Chen, B. Adaptive Filtering with Quantized Minimum Error Entropy Criterion. Signal Process. 2020, 172, 107534. [Google Scholar] [CrossRef]
  19. Ma, W.; Qu, H.; Gui, G.; Xu, L.; Zhao, J.; Chen, B. Maximum Correntropy Criterion Based Sparse Adaptive Filtering Algorithms for Robust Channel Estimation under Non-Gaussian Environments. J. Frankl. Inst. 2015, 352, 2708–2727. [Google Scholar] [CrossRef]
  20. Wang, G.; Peng, B.; Feng, Z.; Yang, X.; Deng, J.; Wang, N. Adaptive Filtering Based on Recursive Minimum Error Entropy Criterion. Signal Process. 2021, 179, 107836. [Google Scholar] [CrossRef]
  21. Zhang, X.; Li, K.; Wu, Z.; Fu, Y.; Zhao, H.; Chen, B. Convex Regularized Recursive Maximum Correntropy Algorithm. Signal Process. 2016, 129, 12–16. [Google Scholar] [CrossRef]
Figure 1. Transient NMSDs (dB) for different algorithms.
Figure 1. Transient NMSDs (dB) for different algorithms.
Electronics 13 00992 g001
Figure 2. Transient NMSDs (dB) of different algorithms in a non-stationary environment.
Figure 2. Transient NMSDs (dB) of different algorithms in a non-stationary environment.
Electronics 13 00992 g002
Figure 3. Transient NMSDs (dB) of CR-RMEE algorithm under different parameters σ , λ .
Figure 3. Transient NMSDs (dB) of CR-RMEE algorithm under different parameters σ , λ .
Electronics 13 00992 g003
Figure 4. Transient NMSD (dB) of different algorithms in four different types of noise (From (ad), the noise distributions are v(n) ~ N(0,0.25), v(n) ~ 0.95N(0,0.01) + 0.05N(0,1), v(n) ~ 0.95N(0,0.01) + 0.05N(0,9), and v(n) ~ 0.95N(0,0.01) + 0.05N(0,25)).
Figure 4. Transient NMSD (dB) of different algorithms in four different types of noise (From (ad), the noise distributions are v(n) ~ N(0,0.25), v(n) ~ 0.95N(0,0.01) + 0.05N(0,1), v(n) ~ 0.95N(0,0.01) + 0.05N(0,9), and v(n) ~ 0.95N(0,0.01) + 0.05N(0,25)).
Electronics 13 00992 g004
Figure 5. Principle diagram of echo cancellation.
Figure 5. Principle diagram of echo cancellation.
Electronics 13 00992 g005
Figure 6. Speech signal.
Figure 6. Speech signal.
Electronics 13 00992 g006
Figure 7. Transient NMSDs (dB) of different algorithms in Echo cancellation experiment.
Figure 7. Transient NMSDs (dB) of different algorithms in Echo cancellation experiment.
Electronics 13 00992 g007
Table 1. Transient NMSDs of CR-RMEE with different parameters after convergence.
Table 1. Transient NMSDs of CR-RMEE with different parameters after convergence.
σ = 0.2 σ = 0.5 σ = 1 σ = 2 σ = 3
λ 2 = 0.942−47.2−38.2−26.9−14.5−3.2
λ 2 = 0.970−52.1−44.1−32.2−20.8−14.0
λ 2 = 0.980−54.2−46.9−36.1−23.8−17.9
λ 2 = 0.992−60.4−54.8−43.5−31.1−24.3
λ 2 = 0.994−61.2−56.6−46.2−33.4−26.6
λ 2 = 0.996−64.2−60.3−49.6−36.7−29.6
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, X.; Ou, S.; Gao, Y. Convex Regularized Recursive Minimum Error Entropy Algorithm. Electronics 2024, 13, 992. https://doi.org/10.3390/electronics13050992

AMA Style

Wang X, Ou S, Gao Y. Convex Regularized Recursive Minimum Error Entropy Algorithm. Electronics. 2024; 13(5):992. https://doi.org/10.3390/electronics13050992

Chicago/Turabian Style

Wang, Xinyu, Shifeng Ou, and Ying Gao. 2024. "Convex Regularized Recursive Minimum Error Entropy Algorithm" Electronics 13, no. 5: 992. https://doi.org/10.3390/electronics13050992

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop