Next Article in Journal
Water-Air Interface Imaging: Recovering the Images Distorted by Surface Waves via an Efficient Registration Algorithm
Previous Article in Journal
Novel Simulation and Analysis of Mie-Scattering Lidar for Detecting Atmospheric Turbulence Based on Non-Kolmogorov Turbulence Power Spectrum Model
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

The Diagnosis of Congestive Heart Failure Based on Generalized Multiscale Entropy-Wavelet Leaders

1
Ocean College, Jiangsu University of Science and Technology, Zhenjiang 212100, China
2
College of Automation, Jiangsu University of Science and Technology, Zhenjiang 212100, China
*
Author to whom correspondence should be addressed.
Entropy 2022, 24(12), 1763; https://doi.org/10.3390/e24121763
Submission received: 13 November 2022 / Revised: 29 November 2022 / Accepted: 30 November 2022 / Published: 1 December 2022
(This article belongs to the Section Multidisciplinary Applications)

Abstract

:
Congestive heart failure (CHF) is a chronic heart condition associated with debilitating symptoms that can lead to mortality. The electrocardiogram (ECG) is a noninvasive and simple diagnostic method that can show detectable changes in CHF. However, manual diagnosis of ECG signals is often erroneous due to the small amplitude and duration of the ECG signals. This paper presents a CHF diagnosis method based on generalized multiscale entropy (MSE)-wavelet leaders (WL) and extreme learning machine (ELM). Firstly, ECG signals from normal sinus rhythm (NSR) and congestive heart failure (CHF) patients are pre-processed. Then, parameters such as segmentation time and scale factor are chosen, and the multifractal spectrum features and number of ELM hidden layer nodes are determined. Two different data sets (A, B) were used for training and testing. In both sets, the balanced data set (B) had the highest accuracy of 99.72%, precision, sensitivity, specificity, and F1 score of 99.46%, 100%, 99.44%, and 99.73%, respectively. The unbalanced data set (A) attained an accuracy of 99.56%, precision of 99.44%, sensitivity of 99.81%, specificity of 99.17%, and F1 score of 99.62%. Finally, increasing the number of ECG segments and different algorithms validated the probability of detection of the unbalanced data set. The results indicate that our proposed method requires a lower number of ECG segments and does not require the detection of R waves. Moreover, the method can improve the probability of detection of unbalanced data sets and provide diagnostic assistance to cardiologists by providing a more objective and faster interpretation of ECG signals.

1. Introduction

Heart failure is a serious global public health problem caused by structural or physical dysfunction. It is also considered to be the final stage in the development of most cardiovascular diseases. Over 26 million people worldwide currently suffer from heart failure, and 70% of CHF cases are caused by cardiovascular diseases, such as coronary artery disease [1]. Other causes of CHF include an elevated hemodynamic load, dysfunction related to ischemia, adverse ventricular remodeling, and genetic mutations [2]. Notably, the prevalence of heart failure increases significantly with increasing age. Therefore, there is a need for early detection of CHF in the aging population, a problem currently faced by many countries around the world. Early detection of CHF to avoid further structural or functional damage to the heart is essential and can save lives.
The diagnosis of CHF is a clinical diagnosis that requires a combination of signs and symptoms and conclusive evidence from investigative tests. Standard diagnostic tests for CHF include chest x-ray, magnetic resonance imaging (MRI), nuclear imaging, echocardiography, and invasive angiography, which can be time-consuming and expensive [3]. However, the electrocardiogram (ECG) is a non-invasive test that has been established as central to diagnosing cardiovascular pathology. It reflects the electrical activity of the heart, and ECG is inexpensive and widely available. The ECG of CHF may be subtly altered, but any alteration in normal is not specific to the diagnosis of CHF, and most academics currently use the ECG signals to diagnose CHF. Due to the small amplitude and duration of the ECG signals, manual diagnosis of the ECG signals is often erroneous, so quantifying ECG signals can improve the objectivity and reliability of heart failure ECG signals diagnosis.
In 2002, Costa et al. [4] improved the sample entropy by introducing the concept of multiscale entropy (MSE), which assesses the complexity of a time series by quantifying the entropy of the time series over a range of time scales. Since MSE was originally introduced, it has become a popular method for quantifying signal complexity and has been successfully applied to different research areas, such as biomedical time series. The basis and implementation of multiscale entropy were subsequently described in detail in 2005, showing its applicability to human heartbeat fluctuations under physiological and pathological conditions [5]. In 2014, Wu et al. [6] proposed a composite multiscale entropy that could solve the accuracy problem of MSE and improve the accuracy of MSE estimation. In 2015, Heurtier [7] elaborated on the MSE algorithm and an improved MSE estimation algorithm and extended MSE to higher-order moments. In the same year, Gao et al. [8] investigated the fundamental bi-scaling law of fractal time series based on MSE, distinguishing healthy individuals from patients with life-threatening congestive heart failure. Costa et al. [9] extended multiscale entropy to generalized multiscale entropy by using different moments of the distribution of random variables to coarse-grain the original time series. In 2017, Liu et al. [10] proposed an MSE analysis method for differential RR interval time series signals and distinguished normal sinus rhythm subjects from CHF patients. However, it has been suggested that entropy methods often do not perform satisfactorily when they are used to analyze the non-linear complexity in physiological signals, and the choice of coarse-grained time scales leads to inaccurate estimates of entropy.
Given the non-linear dynamism of the heart and its self-similarity features, the ECG signals present fractal properties, and the use of fractal methods can also quantify time series and investigate the characteristics of the ECG signals [11]. In 2007, Makowiec et al. [12] analyzed ECG signals during the RR interval using multifractal algorithms. They analyzed the scaling properties of average multifractal partition functions in these physiologically grounded interbeat intervals: low frequency (LF), very low frequency (VLF), and ultra-low frequency (ULV), following normal RR intervals in 39 healthy subjects. In 2016, Chakraborty [13] studied ECG signals collected from the MIT-BIH database in epileptic patients and healthy individuals using the monofractal approach and multifractal approach. The results of the multifractal detrended fluctuation analysis (MFDFA) confirmed that the degree of multifractality was higher in healthy normal individuals compared to people with epilepsy. In 2021, Rogelio et al. [14] introduced a new method based on a clever fusion of fractal dimensional (FD) algorithms and fuzzy logic systems for the automatic prediction of sudden cardiac death events with an accuracy of 91.54%. Serrano et al. [15] used Cantor Set and electroencephalogram (EEG) cases to show that estimation with wavelet leaders (WL) was more accurate than estimation with MFDFA, and the computational complexity of the WL algorithm is significantly lower than that of the MFDFA algorithm. In 2019, Jahmunah et al. [16] reviewed existing methods for automatic CHF diagnosis and suggested that the use of entropic and nonlinear features has greater benefit for the automatic diagnosis of CHF from ECG signals. Therefore, this paper proposes a CHF diagnosis method based on an extreme learning machine (ELM) by a feature extraction method using a combination of entropy and multifractal algorithm and verifies the possibility and effectiveness of the proposed algorithm on ECG signals.
The amplitude of the ECG signals is measured in millivolts, and many researchers extract the R-wave in the ECG signals before extracting the features for study. There is no doubt that R-peak detection is time-consuming. Inspired by Acharya et al. [17], they summarized the methods to automatically diagnose congestive heart failure. In order to improve the classification effect, the paper proposes a CHF diagnosis method based on generalized MSE-WL and ELM. Firstly, the ECG signal amplitude is pre-processed. Secondly, features are extracted, and suitable features are selected according to the proposed method. Next, a machine learning algorithm is used to differentiate between normal and heart failure patients. The dataset is augmented to verify the accuracy of the CHF diagnosis in the unbalanced dataset. Finally, a comparison is made with different algorithms that have been used in recent years to diagnose CHF. Compared with other algorithms, the combination of generalized MSE and WL algorithm features not only improves the diagnosis rate of CHF but also does not require the execution of R-peak detection.

2. Materials and Methods

2.1. Materials

In this study, we used data from three different databases of Physionet. These are from Beth Israel Deaconess Medical Center (BIDMC), Congestive Heart Failure Database (CHF), and MIT-BIH Normal Sinus Rhythm Database (NSR). The CHF and NSR databases are described in detail as follows:
BIDMC congestive heart failure database: The following database has more prolonged ECG recordings (NYHA Class 3 and Class 4) from 15 patients (11 men and 4 women) aged 22 to 71 with severe congestive heart failure [18]. The 20-h recording of each subject contains two ECG signals with a sampling frequency of 250 Hz. Two recordings of each subject constitute 15 × 2 = 30 recordings of the CHF database used in this study.
MIT-BIH normal sinus rhythm database: It contains 25 h of ECG recordings from 18 subjects aged 20 to 50, 5 men and 13 women, from the Arrhythmia Laboratory at Beth Israel Hospital in Boston [19]. Here, data were acquired at a sampling frequency of 128 Hz. Two ECG signals were available for each subject. Notably, subjects in this database were found to have no significant arrhythmias. Two recordings of each subject constitute 18 × 2 = 36 recordings of the NSR database used in this study.
This paper analyses 18 × 2 + 15 × 2 = 66 recordings of NSR data and CHF data, which were obtained from ECG recordings as raw data (in mV). They are then segmented into 5 s ECG (without R-wave extraction). A Normal ECG signal and the ECG signal of a CHF patient are shown in Figure 1a,b.

2.2. Methods

This study focuses on proposing an automatic CHF detection method based on the generalized multiscale entropy-wavelet leaders and extreme learning machine. The technique used for ECG signals classification and the steps can be divided into five parts in Figure 2. Firstly, the original ECG signals are extracted from the physiological database. Next, preprocessing is performed to remove noise and baseline drift. Then, the effect of parameter settings on the model is obtained according to the generalized MSE and the WL algorithms. Subsequently, suitable parameters are selected to obtain statistical features. Finally, the features are fed into the trained model for the classification and diagnosis of CHF patients.

2.2.1. Pre-Processing

In practice, ECG signals are negatively affected by many unfavorable factors during data acquisition and transmission, leading to signal bias and inaccurate diagnosis. Many different algorithms [20,21,22] have been proposed to suppress noise and obtain clean ECG signals. The wavelet transform, with its different scales and low entropy, is highly advantageous in non-stationary signal processing [23]. In this paper, high-frequency noise and baseline drift in the original ECG signals are removed by wavelet methods.
The raw ECG signals contain high-frequency noise and baseline drift, which can be removed by wavelet methods. In this paper, we have performed an 8-level decomposition of the ECG signals. This is because the decomposition levels were chosen to be high to ensure the presence of any low-frequency components of the main signals. We use the ‘bior2.6’ wavelet basis function to obtain a reconstructed denoised signal that has no baseline drift.
The waveforms before and after the wavelet denoising of NSR and CHF are shown in Figure 3. The ECG signals are split into segments of the time length of 5 s, and it is clear from Figure 3a,b that the small fluctuations in the waveform between the large fluctuations have been removed. Since we use the ‘bior2.6’ wavelet basis function with 8-layer decomposition, the length of the time series should be greater than 256.

2.2.2. Multiscale Entropy Algorithm with the First Moment

Considering a denoised time series { x 1 , x 2 , , x N } , the generalized MSE is calculated as follows: First, the original signals are divided into mutually disjoint segments of the length s . Second, selected moments are estimated for each of these segments to produce a coarse-grained time series of scale s . Finally, sample entropy is calculated for each coarse-grained time series.
The multiscale entropy algorithm with the first moment can be denoted as MSEN1; the steps are as follows [7,8]:
Step 1. For a time series { x 1 , x 2 , , x N } of length N , t = 1 , 2 , , N , construct a new sequence y j by dividing the original time series into non-overlapping segments of length s , try to use the equation y j = 1 s i = ( j 1 ) s j s x i to obtain the average of each interval of length s , we can obtain the new time series { y j , j = 1 , 2 , , N / s } with the length N s = N / s , N / s means the largest integer not greater than N / s . When s = 1 , the sequence is the original time signals, and the length of the coarse-graining signals is equal to the ratio of the length of original signals to the scale s .
Step 2. For the new times series { y j , j = 1 , 2 , , N s } , try to obtain the sample entropy. Construct the m -dimensional vector of sequent m points by using the moving window with length m . Divide the time series { y j , j = 1 , 2 , , N s ( N s = N / s ) } into N s m + 1 overlapped sequences, denoted as X m ( i ) = [ y i , y i + 1 , , y i + m 1 ] which represents a vector of m consecutive values of y starting from the i th point.
Step 3. Define d [ X m ( i ) , X m ( j ) ] as the maximum distance between the corresponding elements of vectors X m ( i ) and X m ( j ) , d [ X m ( i ) , X m ( j ) ] = max | y i + k y j + k | , k [ 0 , m 1 ] , i , j [ 1 , N s m + 1 ] , i j .
Step 4. For a given threshold value r , generally, 10–20% [24] of the standard deviation of the time series participating in the sample entropy calculation, and the standard deviation is σ = 1 ( N s 1 ) i = 1 N s ( y i y ¯ ) 2 , y ¯ = 1 ( N s 1 ) i = 1 N s y i . When i N s m + 1 , count the number of values d [ X m ( i ) , X m ( j ) ] < r (called template matches), and take the ratio of this number N m ( i ) to the total number N s m + 1 of distances, A r m ( i ) = N m ( i ) / ( N s m + 1 ) , average it over all i indexes, we can obtain A m ( r ) = [ 1 / ( N s m + 1 ) ] i = 1 N s m + 1 A r m ( i ) .
Step 5. Divide the new time series { y j , j = 1 , 2 , , N s } by moving widow with length m + 1 to obtain overlapped N s m segments, denoted as X m + 1 ( i ) = [ y i , y i + 1 , , y i + m ] , i = [ 1 , N s m ] , consisting of successive m + 1 values of y j from the i th point.
Repeat Steps 3–4 for Steps 6–7 to obtain the average of all i as A m + 1 ( r ) = [ 1 / ( N s m ) ] i = 1 N s m A r m + 1 ( i ) .
Step 8. Calculate the sample entropy:
M S E N 1 = S amp E n ( m , r , s ) = ln [ A m + 1 ( r ) / A m ( r ) ] .

2.2.3. Multiscale Entropy Algorithm with the Second Moment

The multiscale entropy algorithm with the second moment by using the unbiased estimator of variance can be denoted as MSEN2, the steps are as follows.
Step 1. For a time series { x 1 , x 2 , , x N } of length N , t = 1 , 2 , , N , construct a new sequence y j by dividing the original time series into non-overlapping segments of length s , try to use the equation y t = 1 s 1 j = ( t 1 ) s + 1 t s x i 1 s j = ( t 1 ) s + 1 t s x j 2   , t = 1 , 2 , , N s = N / s to obtain the second moment of each segment of length s by using the unbiased estimator of variance, we can obtain the new time series { y j , j = 1 , 2 , , N / s } with the length N s = N / s .
Steps 2–7 are the same as Section 2.2.2.
Step 8. Calculate the sample entropy:
M S E N 2 = S amp E n ( m , r , s ) = ln [ A m + 1 ( r ) / A m ( r ) ] .

2.2.4. Wavelet Leaders Method

The theoretical definition of the WL method of one-dimensional signals is as follows [15,25].
Step 1. Calculate the detailed coefficients d X ( j , k ) of the signals X = { X i , i = 1 , 2 , 3 , , N } by the discrete wavelet transform (DWT), where j and k are the scaling (dilation) index and the shifting (translation) index, respectively. Let X ( t ) t denote the signal X = { X i , i = 1 , 2 , 3 , , N } to be analyzed. Let ψ denote the mother wavelet, characterized by its uniform regularity index and number of vanishing moments N ψ , a strictly positive integer defined as ψ C N ψ 1 and n = 0 , , N ψ 1 , t k ψ ( t ) d t 0 t N ψ ψ ( t ) d t 0 . Let { ψ j , k ( t ) = 2 j / 2 ψ 0 ( 2 j t k ) , j , k } ( j , k ) 2 denote the collection of dilated and translated templates of ψ that form an orthonormal basis of L 2 ( ) . The coefficients of the discrete DWT of X are defined as c j , k = ψ j , k | X . The ( L 1 -normalized) discrete wavelet transform coefficients d X ( j , k ) = 2 j / 2 ψ j , k | X [26]. Note the use of an L 1 -normalization for the wavelet coefficients that better fits local regularity analysis and yields the correct self-similarity exponent of the wavelet coefficients for self-similar functions [27]. For a detailed introduction to wavelet transforms, readers are referred to, e.g., [28].
Step 2. Calculate wavelet leaders L X ( j , k ) . Let us define dyadic intervals as λ j , k = [ k 2 j , ( k + 1 ) 2 j ) . In addition, let 3 λ denote the union of the interval λ with its two adjacent dyadic intervals: 3 λ j , k = λ j , k 1 λ j , k λ j , k + 1 . We define wavelet leaders as
L X ( j , k ) = sup λ 3 λ d X , λ .
This definition means that the wavelet Leader L X ( j , k ) consists of the largest wavelet coefficient d X ( j , k ) computed at all finer scales 2 j 2 j within a narrow time neighborhood, ( k 1 ) 2 j 2 j k < ( k + 2 ) 2 j .
Step 3. Calculate the wavelet leaders structure functions S L ( j , q ) . For fixed analysis scales s = 2 j , we can form the time (space) averages of (the q -orders of) the L X ( j , k ) , referred to as the structure functions
S L ( j , q ) = 1 n j k = 1 n j L X ( j , k ) q .
where n j denotes the number of L X ( j , k ) available at scale 2 j .
If the wavelet leaders structure functions S L ( j , q ) possess power law behaviors with respect to scales in the limit of small scales s = 2 j 0
S L ( j , q ) s ζ L ( q ) .
where ζ L ( q ) is often referred to as the scaling function. This power law behavior establishes a clear and deep connection between the concepts of scale invariance and multifractal analysis [29].
Step 4. Calculate ζ L ( q ) with a given q . The scaling function ζ L ( q ) is then defined as
ζ L ( q ) = lim inf s 0 log 10 ( S L ( j , q ) ) log 10 ( s ) .
It is easy to obtain the singularity strength function α ( q ) and the multifractal spectrum f ( α ) via the Legendre transform
α ( q ) = d ζ L ( q ) / d q .
f ( α ) = inf q [ q α ( q ) ζ L ( q ) + 1 ] .
The methods based on the WL with different mother wavelets have different detrending capabilities [30]. In this paper, the wavelet basis function is ‘db3′, and the scaling range is s = [ 2 1 , 2 2 , 2 3 , 2 4 , , 2 log 2 ( N / ( 2 3 + 1 ) ) ] [26,27].
A schematic representation of the q ζ L ( q ) curve of one time series is shown in Figure 4, where an obvious inflection point appears near q = 0 . This means that ζ L ( q ) is not a linear function of q , so the time series is multifractal.
Figure 5 shows a schematic representation of the multifractal spectrum f ( α ) of the same time series based on WL in Figure 4. For the left endpoint of the multifractal spectrum, we use the singularity exponent α min to denote the corresponding horizontal axis of this point; it reflects the degree of large fluctuation of the signals. α 0 is the singularity exponent of the top point of the multifractal spectrum and corresponds to the most probable probability subset of the time series. The right endpoint with α max , reflects the degree of small fluctuation of the signals. The spectrum width Δ α = α max α min reflects the uneven degree of fluctuation of the signals.
For fast speed calculation, we use the matrix calculation method of entropy analysis [31].
F α ( q ) ( L X q ( j , k ) ) = k = 1 n j ( L X q ( j , k ) log 2 ( L X ( j , k ) ) ) ( k = 1 n j L X q ( j , k ) ) = k = 1 n j [ P j , k q log 2 ( L X ( j , k ) ) ] log 2 s α ( q ) .
F f ( α ( q ) ) ( L X q ( j , k ) ) = k = 1 n j ( P j , k q log 2 P j , k q ) + log 2 ( n j ) log 2 s f ( α ( q ) ) .
where P j , k q = L q ( j , k ) ( k = 1 n j L q ( j , k ) ) .
Similar to the estimation of ζ L ( q ) in WL method, α ( q ) and f ( α ( q ) ) are estimated as the linear regression slopes of the q -order entropies F α ( q ) ( L X q ( j , k ) ) and F f ( α ( q ) ) ( L X q ( j , k ) ) , respectively, and scale s = 2 j in log-log coordinates.

2.2.5. Extreme Learning Machine

The Extreme Learning Machine (ELM) is a single implicit layer feedforward neural network [32,33]. By setting the number of neurons in the implicit layer, the connection weights β between the implicit layer and the output layer are not adjusted iteratively but are determined once by solving a system of equations. From the perspective of learning efficiency, the extreme learning machine has the advantages of few training parameters, fast learning speed, and strong generalization ability. It consists of three parts: the input layer, the implicit layer, and the output layer, as shown in Figure 6. In this paper, we use the ELM classifier to diagnose CHF.

2.2.6. K-Fold Cross-Validation

In recent years, K-fold cross-validation has been commonly used in applied machine learning to compare and select models for a given predictive modeling problem. This is easy to understand and implement, resulting in skill estimates that typically have lower bias than other methods. In our experiments, the data were divided into K (K = 5) equal-sized parts, with one of the five parts selected for testing and the rest of the data for training. The results recorded in all five iterations are averaged and considered as the overall performance of our proposed system. For all class-oriented experiments, we used this cross-validation method.

2.2.7. Evaluation Criteria

Accuracy alone is not sufficient in the classification of data, as it is calculated by the ratio of accurately estimated data to the total data set. Sensitivity measures how often a test works properly, while specificity is a measure of a test’s ability to produce negative results for an untested disease. Therefore, these parameters should be assessed together. On the other hand, F1 score is the summed average of precision and recall rather than the arithmetic means to avoid ignoring extremes. Therefore, F1 score must also be included in the assessment metric. The equations for the assessment indicator are as follows:
•Accuracy:
A C C = T P + T N T P + T N + F P + F N × 100 %
•Precision:
P P V = T P T P + F P × 100 %
•Sensitivity:
S E N = T P T P + F N × 100 %
•Specificity:
S P E = T N F P + T N × 100 %
•F1 score:
F 1 = 2 T P 2 T P + F N + F P × 100 %
where TP indicates correct identification in the absence of disease, TN indicates correct detection of disease, FP indicates incorrect detection when the disease is present, and the detector is not detected, and FN indicates that disease is not present, but the detector detects disease.

3. Results

3.1. Optimization of Parameter Settings

When we use the MSE method, we first need to determine the parameter settings ( s , m , r ). Usually, the length of the data should be at least 10 m 20 m [34]. When MSE is used to analyze ECG signals, the length of the data N = t · f s , f s is the sampling frequency of the ECG signals, so we need to determine the segmentation time t of the ECG, The parameters r should be 10–20% [24] of the standard deviation, and we now use MSEN1 as an example to illustrate the effect of the parameter settings.

3.1.1. Embedded Dimensions

As with approximate entropy and sample entropy, the embedding dimension m = 2 is generally taken. The larger m is, the more detailed information can be obtained when dynamically reconstructing the joint probabilities, but the larger m is, the longer the length of data required, and the computation time will be longer. Therefore, we use the embedding dimension m = 2 .

3.1.2. Segmentation Time

In the analysis and processing of real signals, the length of the signal segmentation is an important factor affecting real-time analysis. If its features can be obtained with shorter signals, it will be important for the diagnosis of diseases. The sampling frequency of the NSR dataset is 128 Hz, and it is known that the data length N should be greater than 256, t = N / f s = 256 / 128 = 2 s, so the time t = [ 4 , 8 , 16 , 32 , 64 , 128 ] s is chosen for testing. After testing, it was found that there are a large number of INF points for calculating the multiscale entropy when segmentation time t is less than 32 s, which is because the multiscale entropy algorithm gives an inaccurate estimate of entropy, even leading to short time sequences with undefined entropy values. Therefore, the segmentation time t = [ 32 , 64 , 128 ] s were chosen.
The curves of MSEN1 for the ECG signals of NSR and CHF at t = [ 32 , 64 , 128 ] s are shown in Figure 7, where s = 1 : 1 : 50 , m = 2 , and r = 0.20 σ . As can be seen from Figure 7a,b, the differences between the three curves for different segmentation times are very small. Therefore, the length of the segmentation time has little effect on the multiscale entropy of the ECG signals. The multiscale entropy of the NSR data is unstable at t = 32 s and more stable at t = 64 s and t = 128 s. However, the larger the segmentation time, the more sampling points, and the longer the calculation time; therefore, the segmentation time is chosen to be t = 64 s.
The ECG signals for both data were chosen to be segmented at time t = 64 s. The datasets used in this study (Set A, Set B) are shown in Table 1. Both sets A and B contain full ECG data, with Set A being the unbalanced dataset and Set B being the balanced dataset.

3.1.3. Scale Factor

We plotted the MSEN1 curve for the ECG signals at scale s = 1 : 1 : 50 as shown in Figure 8, where t = 64 s and m = 2 . The entropy values of the coarse-grained time series from healthy subjects are significantly higher than those of the CHF. Therefore, coarse-grained time series from healthy subjects at large time scales may be more complex. This finding is consistent with 1 / f noise [4] ( 1 / f noise is generated as follows: we start with uniformly distributed white noise, we compute the fast Fourier transform (FFT), and after imposing an 1 / f distribution on the power spectrum, we compute the inverse FFT.) containing complex structures on multiple time scales. We use entropy methods to quantify the dynamics of each coarse-grained time series. We found an overall increasing trend in the magnitude of entropy values in the scale 1 to 20 range and stabilization to relatively constant values in the scale 20 to 50 range. However, the entropy values of coarse-grained time series from NSR were significantly higher than those of CHF. This suggests that the normal ECG signal data is more complex and the ECG signal from congestive heart failure is more regular.
When the selected sampling points are fixed, the larger the scale s , the less time is required for the calculation of MSEN1. However, with too large a scale value, the final data will become shorter, and the number of vector groups obtained in the coarse-grained process will be smaller. Therefore, the scale should be relatively small, and MSEN1 clearly varies linearly as the scale s . Therefore, we choose scale s = 10 : 1 : 20 .

3.1.4. Similarity Tolerance

The similarity tolerance r represents the width of the fuzzy function boundary. r is too large, and much statistical information is lost; r is too small, and the estimated statistical properties are unsatisfactory, increasing sensitivity to the noise of the results. Figure 9 shows the MSEN1 curves for the ECG signals of NSR and CHF for different r = [ 0.05 σ , 0.10 σ , 0.15 σ , 0.20 σ ] at the t = 64 s and s = 1 : 1 : 50 . For Figure 9a, the green dashed line with the green circle indicates that the mean of the MSEN1 of the signals with r = 0.05 σ ( σ denotes the standard deviation of the time series after coarse granulation) has an omission because r is too small and the statistical properties are not ideal. From Figure 9a,b, it can be seen that the larger the similarity tolerance r is, the larger the multiscale entropy value is, and r = 0.15 σ is chosen to maintain the smoothness of the data.

3.1.5. Multifractal Spectrum Features

The mother wavelets { ψ ( i ) ( x ) } i = 1 , , 2 d 1 are further required to possess additional regularity and localization properties. They are assumed to belong to C r ψ ( d ) with r ψ as large as possible. When r ψ 1 , all their partial derivatives of order at most r ψ have fast decay. For a one-dimensional signal, we have d = 1 . In addition, the number of vanishing moments N ψ is defined as a positive integer such that for any polynomial P of degree strictly smaller than N ψ , P ( x ) ψ ( i ) ( x ) d x = 0 . Both the regularity and the vanishing moment assumptions are required in order to obtain accurate wavelet characterizations of pointwise regularity: Let h max denote the largest smoothness order found in X , then a sufficient condition for choosing the mother wavelet reads: h max < min ( r ψ , N ψ ) .
The computation of the uniform Hölder exponent h min [35], using the following wavelet characterization: h min = lim inf j log 2 ( sup k d X ( j , k ) ) log 2 ( 2 j ) . Indeed, if h min > 0 , then X is a continuous function, whereas if h min < 0 , then X L l o c see References [36,37]. For numerous real-world applications, the restriction h min > 0 constitutes a severe limitation, c.f. Reference [37].
From a practical point of view, mother wavelets satisfying r ψ > h min are required for an accurate estimation of h min . Similarly, h max = lim inf j log 2 ( min k d X ( j , k ) ) log 2 ( 2 j ) .
In general, one does not have information concerning the a priori regularity of the data. Therefore, one does not know how smooth the analyzing wavelets should be. In practice, a rule of thumb consists of using smoother and smoother wavelets until the outcome no longer depends on the wavelet used, which is interpreted as meaning that sufficient regularity has been reached. This can afterward be confirmed using multifractal analysis tools (see Reference [36]), which allow us to determine the maximum regularity exponent present in the data. Further, with orthonormal wavelet bases (such as the so-called “Daubechies” compactly supported wavelets), widely used in applications), one necessarily has N ψ r ψ . A sufficient (and conservative) condition for accurate wavelet characterizations of pointwise regularity simplifiers to h max < r ψ .
We adhere to the convention that the finest available dyadic scale is labeled by j 1 = 1 . That means s = 2 1 , 1-layer wavelet decomposition. A Daubechies’ wavelet with N ψ = 3 vanishing moments is used, then j 2 = 10 = log 2 ( 8192 / ( 2 3 + 1 ) ) , this means s = 2 10 , 10-layer wavelet decomposition. Because α min > 0 , α max < 3 in Figure 10a,b, for h min > 0 and h max < 3 , so ‘db3′ is suitable. Figure 10 shows the simulation results for data sets A and B. In this paper, the WL-based multifractal analysis method is used, where the ‘db3’ wavelet basis functions are q max = 5 , Δ q = 0.25 , and the entropy analysis is carried out by matrix calculation methods, and the scale range can be s = [ 2 1 , 2 2 , 2 3 , 2 4 , 2 5 , 2 6 , 2 7 , 2 8 , 2 9 , 2 10 ] .
Figure 10a shows the WL-based multifractal spectrum for set A with 540 groups of NSR ECG signals of 8192 points and 360 groups of CHF ECG signals of 16,000 points, where the ‘db3’ wavelets are q max = 5 and Δ q = 0.25 . For ‘db3’, the number 3 refers to the number of vanishing moments. The NSR and CHF signals are sampled by different sample frequencies, so we choose the signal of 64 s (8192/128 = 64, 16,000/250 = 64) for NSR and CHF ECG signals, which results in the two different lengths of the signals 8192 and 16,000. When we do the simulations, we choose the same parameters for them, so the scale range is determined by the length 8192. The length of the wavelet filter is 2 × 3 = 6 of ‘db3’. Thus, when we choose the parameter s = [ 2 j 1 , , 2 j 2 ] for the range of scales, we fix the maximum scale 2 j 2 by choosing the largest integer j 2 that is not greater than log 2 ( N / ( 2 × 3 + 1 ) ) , i.e., log 2 ( 8192 / ( 2 × 3 + 1 ) )   13 3 = 10 . In the simulations, we use j 2 = 10 and j 1 = 1 . Figure 10b shows the WL-based multifractal spectrum for set B, where the ‘db3’ wavelet is q max = 5 , Δ q = 0.25 and there are differences between the two ECG signals of the same sample numbers 360 groups, respectively.
In Figure 3 after wavelet denoising, the small fluctuations are removed, and the slightly larger fluctuations are retained, as reflected by the singularity index α min = α ( q = 5 ) at the left endpoint of each multifractal spectrum. These large fluctuations in the ECG signals of a normal person are similar to each other, but the fluctuations in the ECG signals of an unhealthy person are not uniform and regular, so the slightly larger fluctuations result in slightly smaller singularity exponents of the left end point of the multifractal spectrum. When q > > 1 , the large fluctuations will dominate the statistics and result in a smaller singularity index. The singularity index α 0 corresponding to the top of the multifractal spectrum corresponds to the most probable or most likely subset (i.e., the subset with the largest number of line segments or elements, and this subset has a fractal dimension of 1, the ordinal number of points at the top of the multifractal singularity spectrum). The calculated multifractal profiles for normal and unhealthy individuals have some deviation in the value of the singularity index corresponding to the top point of the multifractal profile. When we extract the multifractal spectrum using the WL method, we can use the two parameters α min and α 0 to do the classification of the ECG signals of normal and unhealthy individuals.

3.1.6. Number of ELM Hidden Layer Nodes

In this study, ELM was selected as the classifier. To maintain good generalization performance, it was particularly important to determine l , which denotes the appropriate number of nodes of the hidden layer. The optimal number of nodes of the hidden layer was determined using five-fold cross-validation. The balanced dataset Set B was selected, and the 360 ECG segments were divided into five parts, of which four were used for training the model and one for testing the model. MSEN1, MSEN2, and the singularity indices α 0 and α min of the multifractal spectra of Group B were calculated according to the multiscale entropy algorithm for the first-order moment and second-order moment and WL. The number of nodes of the hidden layer was 1–100. The classification accuracy of the training set was obtained from five-fold cross-validation, as shown in Figure 11.
It can be seen from Figure 11 that the classification accuracy of the training and test sets gradually increases as the number of nodes of the hidden layer increases. The accuracy of the five-times cross-validation gradually increases to 100% when l > 20 . The classification accuracy of the training set with node numbers from 20 to 70 is shown in Table 2. It can be observed from Table 2 that the best results are obtained for the 5-fold training set when l 60 . Considering the running time, l = 60 is chosen as the optimal number of hidden layer nodes.

3.2. Training and Test of the CHF Classifier

3.2.1. Results of Classification

The number of ECG segments in sets A and B, the parameter settings of the generalized multiscale entropy algorithm, the WL-based multifractal spectrum parameters, and the number of ELM hidden layer nodes are determined in Section 3.1. Finally, M S E N 1 , M S E N 2 , α 0 , α min feature vectors are fed into the classifier ELM with 60 nodes of the hidden layer for classification. A five-fold cross-validation strategy is used to evaluate diagnostic algorithms using the accuracy, precision, sensitivity, specificity, and F1-score parameters obtained from the confusion matrix parameters.
The ECG segments of sets A and B were divided equally into five segments. For each iteration, four of five ECG segments were selected for training, and the rest were used for testing, which was repeated five times. Finally, the average of the five iterations was taken. Table 3 and Table 4 show the overall average performance of our proposed classification method for classifying normal and CHF categories for both sets A and B.
In Table 3, the parameters of the confusion matrix are shown for the unbalanced data, Set A. Our algorithm achieved significant results for CHF detection, with an accuracy of 99.56%, precision of 99.44%, sensitivity of 99.81%, specificity of 99.17%, and an F1 score of 99.62%. SEN value is 99.81%, which means 99.81% of the normal ECG segments were correctly classified as normal. SPEC is 99.17%, which means 99.17% of the CHF signals were correctly classified as CHF category, and only 0.19% and 0.83% of the ECG signals were incorrectly classified as CHF and normal. In Table 4, for Set B, the overall accuracy of 99.72%, precision of 99.46%, sensitivity of 100%, specificity of 99.44%, and an F1 score of 99.73%. Only 0.56% of ECG signals are incorrectly classified as CHF. It can also be seen that the accuracy of the CHF detection rate for the unbalanced data Set A in Table 3 is less than that of the CHF detection rate for the balanced data Set B in Table 4, with a variability of less than 0.3%.

3.2.2. Results of Adding Data Segments

In order to evaluate the effectiveness of the algorithm and model proposed in this paper for CHF detection and to validate the CHF detection rate of the unbalanced data set, the number of ECG signal segments was increased. 1800 ECG segments were taken for NSR, and 1500 ECG segments were taken for CHF, set as Set C.7200 ECG segments were taken for NSR, and 6000 ECG segments were taken for CHF, set as Set D. The overall confusion matrix of CHF detection with the same parameters selected for the same algorithm is shown in Table 5. For Set C, the overall accuracy of 99.24%, precision of 99.22%, sensitivity of 99.39%, specificity of 99.07%, and an F1 score of 99.30%. Set D attained an accuracy of 99.41%, precision of 99.57%, sensitivity of 99.35%, specificity of 99.48%, and an F1 score of 99.46%. The differences in the five diagnostic assessment indicators were smaller in Sets C and D compared with Set A, with the largest difference in specificity being 0.65%.
Table 6 shows the time required to run each dataset. The more ECG signal segments there are, the more time it takes to diagnose CHF. Therefore, the increase in data segments did not significantly reduce the accuracy of diagnosing CHF, but the training time increased, and the training time was shorter in Set A. Our proposed algorithm can therefore select a small number of ECG time series to diagnose CHF in the presence of a large amount of data and significantly improves the accuracy of diagnosing unbalanced data sets.

3.2.3. Comparison Results of Different Algorithms

Data Set A was selected, and the same five-fold cross-validation strategy was used; four of five ECG segments were selected for training, and the rest were used for testing. The generalized multiscale entropy extracted features and WL-extracted features are fed into ELM classification separately, and then the diagnostic results by using the classifiers Support Vector Machine (SVM) and K-Nearest Neighbor (KNN) are compared with the results of the proposed diagnose method in this paper. For the multi-classification of ECG signals by using SVM and Support Regression Machine (SVR), we used the kernel function as Radial Basis Function, and the gamma parameter coefficient was 0.05. For the KNN algorithm, the K value was chosen as the square root of the number of samples in the training set, which was 30.
The results of 5-fold cross-validation are shown in Table 7, and the associated box plot is shown in Figure 12 to display the influences of the different algorithms on the accuracy of diagnostic results. One can conclude that each classifier has high accuracy (>95%) because of the distinguishable feature extracted in the previous stage of work. The highest overall accuracy of 99.56% correct detection of CHF is obtained in the results of the generalized MSE-WL-based detection method with ELM classification. Therefore, based on the overall classification correctness, it is concluded that the generalized MSE-WL feature extraction method can be applied to CHF automatic diagnosis. The extracted features have a good performance on multiple classifiers.

4. Discussion

According to the results obtained in Table 7, the method is better for classifying ECG signals of NSR and CHF compared with the algorithms based on the physiological MIT-BIH database. Table 8 shows a comparison of the results obtained with the literature of recent years. In the literature, CHF and NSR conditions are classified using classical machine learning for the classification algorithm. There is literature to detect CHF ECG signals with accuracy greater than 99%, but most select a large number of ECG segments, and some detect R peaks. In contrast, we proposed a diagnosis method of CHF; it does not need to detect R-peaks on the ECG signals and extract a small number of ECG segments to obtain a CHF detection rate greater than 99% and to improve the accuracy of the unbalanced dataset. In this study, the time series of ECG signals used was short. Although the proposed feature extraction method did not achieve 100% accuracy in detecting CHF, it was the first implementation to combine generalized multiscale entropy and WL algorithms to classify ECG signals into normal and CHF categories using the ELM classifier, with 99.17% (Set A) and 99.44% (Set B) specificity in detecting CHF. With no reduction in CHF detection performance, the method does not require R-wave extraction. As a complexity analysis method, the generalized multiscale entropy method does not strongly depend on the data length when dealing with different complex signals. Therefore, the method requires fewer ECG segments to detect CHF and improves the timeliness of heart failure diagnosis.
In this work, we selected ECG signals with a duration of 64 s for automatic diagnosis of CHF using a feature binding approach. The programming implementations of the generalized multiscale entropy and WL algorithms are based on entropy theory. The former evaluates the complexity of a time series by quantifying its entropy over a range of temporal scales, and the coarse-grained procedure of different scale values results in a longer run time. However, the programming procedure of the linear fit of the power-law relationship of different scale values in the WL algorithm, which is realized with the help of matrices, runs fast. Hence, compared with other algorithms in the literature, the novelty of this work is to combine the generalized multiscale algorithm with the WL algorithm to obtain good feature vectors to ensure comparable good performance of the diagnosis of CHF, which will help cardiologists in the diagnosis and treatment of CHF.
The advantages of our proposed classification method are: (1) A feature classification method of CHF and NSR signals based on the generalized MSE-WL and ELM is proposed. (2) No R-peak detection is required. (3) Requires a small number of ECG segments. (4) Improves the detection rate of unbalanced datasets. The limitations of the proposed classification method are: (1) Requires ECG segmentation time greater than 32 s. (2) No tests for other cardiac diseases were performed.

5. Conclusions

CHF is a complex clinical condition in which the ability of the heart to fill and pump blood is impaired due to functional or structural disease. Early detection of CHF is of high importance to avoid death. In this paper, we propose a CHF feature classification method based on generalized MSE-WL and ELM, which does not require R-peak detection and uses ECG signals to diagnose CHF automatically. The ECG signals obtained from Physionet are used to determine the parameter settings for the first moment of MSE by simulations of different data segmentation, scaling ranges, and similarity tolerance. It provides guidance on feature selection and parameter settings in practical applications. The extracted feature factors are fed into the ELM for training and testing, and confusion matrices and accuracy values are given to evaluate the results obtained. Using our proposed method, we obtained the balanced data set (B) had an accuracy of 99.72%, and the unbalanced data set (A) attained an accuracy of 99.56%. The results of CHF classification suggest that a CHF detection method based on generalized MSE-WL can help doctors better diagnose CHF. The method requires fewer ECG segments to accurately distinguish between normal and CHF patients and can considerably reduce the workloads of doctors. It provides clinicians with a valuable reference for diagnosing CHF.
In the future, we will investigate a new method to automatically detect cardiac disease by imaging the ECG signal without removing noise. The ECG signal will be characterized using two-dimensional entropy theories and two-dimensional multifractal methods [43]. Then a classification method will be used to classify the ECG signal and diagnose cardiac disease.

Author Contributions

Conceptualization, methodology, and validation, J.Y. and C.X.; formal analysis and investigation, C.X.; writing—original draft preparation, J.Y.; writing—review and editing, C.X.; supervision, J.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This work was sponsored by the National Natural Science Foundation of China under Grant 61901195 and Grant 62276117.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

All data used in the experiments can be downloaded from the following links: http://www.physionet.org/physiobank/database/#ecg (accessed on 15 June 2022).

Acknowledgments

The authors are grateful to the anonymous reviewers for their helpful comments and valuable suggestions that led to a significant improvement of the manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Acharya, U.R.; Fujita, H.; Lih, O.S.; Adam, M.; Tan, J.H.; Chua, C.K. Automated Detection of Coronary Artery Disease Using Different Durations of ECG Segments with Convolutional Neural Network. Knowl. Based Syst. 2017, 132, 62–71. [Google Scholar] [CrossRef]
  2. Burgess, A.; Shah, K.; Hough, O.; Hynynen, K. Focused Ultrasound-Mediated Drug Delivery through the Blood–Brain Barrier. Expert Rev. Neurother. 2015, 15, 477–491. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Gladding, P.; Cave, A.; Zareian, M.; Smith, K.; Hussan, J.; Hunter, P.; Erogbogbo, F.; Aguilar, Z.; Martin, D.; Chan, E.; et al. Open Access Integrated Therapeutic and Diagnostic Platforms for Personalized Cardiovascular Medicine. JPM 2013, 3, 203–237. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Costa, M.; Goldberger, A.L.; Peng, C.-K. Multiscale Entropy Analysis of Complex Physiologic Time Series. Phys. Rev. Lett. 2002, 89, 068102. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Costa, M.; Goldberger, A.L.; Peng, C.-K. Multiscale Entropy Analysis of Biological Signals. Phys. Rev. E 2005, 71, 021906. [Google Scholar] [CrossRef] [Green Version]
  6. Wu, S.-D.; Wu, C.-W.; Lin, S.-G.; Lee, K.-Y.; Peng, C.-K. Analysis of Complex Time Series Using Refined Composite Multiscale Entropy. Phys. Lett. A 2014, 378, 1369–1374. [Google Scholar] [CrossRef]
  7. Humeau-Heurtier, A. The Multiscale Entropy Algorithm and Its Variants: A Review. Entropy 2015, 17, 3110–3123. [Google Scholar] [CrossRef] [Green Version]
  8. Gao, J.; Hu, J.; Liu, F.; Cao, Y. Multiscale Entropy Analysis of Biological Signals: A Fundamental Bi-Scaling Law. Front. Comput. Neurosci. 2015, 9, 64. [Google Scholar] [CrossRef] [Green Version]
  9. Costa, M.; Goldberger, A. Generalized Multiscale Entropy Analysis: Application to Quantifying the Complex Volatility of Human Heartbeat Time Series. Entropy 2015, 17, 1197–1203. [Google Scholar] [CrossRef]
  10. Liu, C.; Gao, R. Multiscale Entropy Analysis of the Differential RR Interval Time Series Signal and Its Application in Detecting Congestive Heart Failure. Entropy 2017, 19, 251. [Google Scholar] [CrossRef]
  11. Kiani, K.; Maghsoudi, F. Classification of 7 Arrhythmias from ECG Using Fractal Dimensions. J. Bioinform. Syst. Biol. 2019, 2, 053–065. [Google Scholar] [CrossRef]
  12. Makowiec, D.; Dudkowska, A.; Galaska, R.; Rynkiewicz, A. Multifractal Analysis of Normal RR Heart-Interbeat Signals in Power Spectra Range. 2007. Available online: https://doi.org/10.48550/arXiv.q-bio/0702047 (accessed on 28 November 2022).
  13. Chakraborty, M.; Das, T.; Ghosh, D. Comparative Analysis of Different Fractal Methods in Studying Post-Ictal ECG Signals of Epilepsy Patient. In Proceedings of the 2016 IEEE First International Conference on Control, Measurement and Instrumentation (CMI), Kolkata, India, 8–10 January 2016; IEEE: Kolkata, India, 2016; pp. 219–223. [Google Scholar]
  14. Piña-Vega, R.; Valtierra-Rodriguez, M.; Perez-Ramirez, C.A.; Amezquita-Sanchez, J.P. EARLY PREDICTION OF SUDDEN CARDIAC DEATH USING FRACTAL DIMENSION AND ECG SIGNALS. Fractals 2021, 29, 2150077. [Google Scholar] [CrossRef]
  15. Serrano, E.; Figliola, A. Wavelet Leaders: A New Method to Estimate the Multifractal Singularity Spectra. Phys. A Stat. Mech. Appl. 2009, 388, 2793–2805. [Google Scholar] [CrossRef]
  16. Jahmunah, V.; Oh, S.L.; Wei, J.K.E.; Ciaccio, E.J.; Chua, K.; San, T.R.; Acharya, U.R. Computer-Aided Diagnosis of Congestive Heart Failure Using ECG Signals—A Review. Phys. Med. 2019, 62, 95–104. [Google Scholar] [CrossRef] [Green Version]
  17. Acharya, U.R.; Fujita, H.; Oh, S.L.; Hagiwara, Y.; Tan, J.H.; Adam, M.; Tan, R.S. Deep Convolutional Neural Network for the Automated Diagnosis of Congestive Heart Failure Using ECG Signals. Appl. Intell. 2019, 49, 16–27. [Google Scholar] [CrossRef]
  18. Baim, D.S.; Colucci, W.S.; Monrad, E.S.; Smith, H.S.; Wright, R.F.; Lanoue, A.; Gauthier, D.F.; Ransil, B.J.; Grossman, W.; Braunwald, E. Survival of Patients with Severe Congestive Heart Failure Treated with Oral Milrinone. J. Am. Coll. Cardiol. 1986, 7, 661–670. [Google Scholar] [CrossRef] [Green Version]
  19. Goldberger, A.L.; Amaral, L.A.N.; Glass, L.; Hausdorff, J.M.; Ivanov, P.C.; Mark, R.G.; Mietus, J.E.; Moody, G.B.; Peng, C.-K.; Stanley, H.E. PhysioBank, PhysioToolkit, and PhysioNet: Components of a New Research Resource for Complex Physiologic Signals. Circulation 2000, 101, E215–E220. [Google Scholar] [CrossRef] [Green Version]
  20. Guo, X.; Shen, C.; Chen, L. Deep Fault Recognizer: An Integrated Model to Denoise and Extract Features for Fault Diagnosis in Rotating Machinery. Appl. Sci. 2016, 7, 41. [Google Scholar] [CrossRef] [Green Version]
  21. Al-Tmeme, A.; Woo, W.L.; Dlay, S.S.; Gao, B. Underdetermined Convolutive Source Separation Using GEM-MU With Variational Approximated Optimum Model Order NMF2D. IEEE/ACM Trans. Audio Speech Lang. Process. 2017, 25, 35–49. [Google Scholar] [CrossRef] [Green Version]
  22. Al-Nima, R.R.O.; Abdullah, M.A.M.; Al-Kaltakchi, M.T.S.; Dlay, S.S.; Woo, W.L.; Chambers, J.A. Finger Texture Biometric Verification Exploiting Multi-Scale Sobel Angles Local Binary Pattern Features and Score-Based Fusion. Digit. Signal Process. 2017, 70, 178–189. [Google Scholar] [CrossRef]
  23. Kumar, A.; Tomar, H.; Mehla, V.K.; Komaragiri, R.; Kumar, M. Stationary Wavelet Transform Based ECG Signal Denoising Method. ISA Trans. 2021, 114, 251–262. [Google Scholar] [CrossRef] [PubMed]
  24. Zurek, S.; Grabowski, W.; Kosmider, M.; Jurga, S.; Guzik, P.; Piskorski, J. Bootstrapping the Empirical Bounds on the Variability of Sample Entropy in 24-Hour ECG Recordings for 1 Hour Segments. J. Appl. Math. Comput. Mech. 2018, 17, 105–113. [Google Scholar] [CrossRef] [Green Version]
  25. Ouahabi, A.; Femmam, S. Wavelet-Based Multifractal Analysis of 1-D and 2-D Signals: New Results. Analog Integr. Circuits Signal Process. 2011, 69, 3–15. [Google Scholar] [CrossRef]
  26. Leonarduzzi, R.; Wendt, H.; Jaffard, S.; Roux, S.G.; Torres, M.E.; Abry, P. Extending Multifractal Analysis to Negative Regularity: P-Exponents and P-Leaders. In Proceedings of the 2014 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Florence, Italy, 4–9 May 2014; IEEE: Florence, Italy, 2014; pp. 305–309. [Google Scholar]
  27. Jaffard, S.; Melot, C.; Leonarduzzi, R.; Wendt, H.; Abry, P.; Roux, S.G.; Torres, M.E. P-Exponent and p-Leaders, Part I: Negative Pointwise Regularity. Phys. A Stat. Mech. Appl. 2016, 448, 300–318. [Google Scholar] [CrossRef] [Green Version]
  28. Mallat, S. A wavelet tour of signal processing. Phys. A Stat. Mech. Appl. 1998, 628, 998. [Google Scholar]
  29. Wendt, H.; Roux, S.G.; Jaffard, S.; Abry, P. Wavelet Leaders and Bootstrap for Multifractal Analysis of Images. Signal Process. 2009, 89, 1100–1114. [Google Scholar] [CrossRef] [Green Version]
  30. Kantelhardt, J.W. Fractal and Multifractal Time Series. 2008. Available online: https://doi.org/10.48550/arXiv.0804.0747 (accessed on 28 November 2022).
  31. Ihlen, E.A.F. Multifractal Analyses of Response Time Series: A Comparative Study. Behav. Res. 2013, 45, 928–945. [Google Scholar] [CrossRef]
  32. Huang, G.-B.; Zhu, Q.-Y.; Siew, C.-K. Extreme Learning Machine: Theory and Applications. Neurocomputing 2006, 70, 489–501. [Google Scholar] [CrossRef]
  33. Yang, Y.; Xi, C. Rolling Bearing Fault Diagnosis Based on MFDFA-SPS and ELM. Math. Probl. Eng. 2022, 2022, 4034477. [Google Scholar] [CrossRef]
  34. Richman, J.S.; Moorman, J.R. Physiological Time-Series Analysis Using Approximate Entropy and Sample Entropy. Am. J. Physiol. Heart Circ. Physiol. 2000, 278, H2039–H2049. [Google Scholar] [CrossRef] [Green Version]
  35. Leonarduzzi, R.; Wendt, H.; Abry, P.; Jaffard, S.; Melot, C.; Roux, S.G.; Torres, M.E. P-Exponent and p-Leaders, Part II: Multifractal Analysis. Relations to Detrended Fluctuation Analysis. Phys. A Stat. Mech. Appl. 2016, 448, 319–339. [Google Scholar] [CrossRef]
  36. Jaffard, S.; Abry, P.; Roux, S. Function Spaces Vs. Scaling Functions: Tools for Image Classification. In Mathematical Image Processing; Springer Proceedings in Mathematics; Bergounioux, M., Ed.; Springer: Berlin/Heidelberg, Germany, 2011; Volume 5, pp. 1–39. ISBN 978-3-642-19603-4. [Google Scholar]
  37. Abry, P.; Jaffard, S.; Wendt, H. Irregularities and Scaling in Signal and Image Processing: Multifractal Analysis. In Fractals and Dynamics in Mathematics, Science, and the Arts: Theory and Applications; World Scientific: Singapore, 2015; Volume 1, pp. 31–116. ISBN 978-981-4366-06-9. [Google Scholar]
  38. Daqrouq, K.; Dobaie, A. Wavelet Based Method for Congestive Heart Failure Recognition by Three Confirmation Functions. Comput. Math. Methods Med. 2016, 2016, 7359516. [Google Scholar] [CrossRef] [Green Version]
  39. Sudarshan, V.K.; Acharya, U.R.; Oh, S.L.; Adam, M.; Tan, J.H.; Chua, C.K.; Chua, K.P.; Tan, R.S. Automated Diagnosis of Congestive Heart Failure Using Dual Tree Complex Wavelet Transform and Statistical Features Extracted from 2 s of ECG Signals. Comput. Biol. Med. 2017, 83, 48–58. [Google Scholar] [CrossRef]
  40. Zhang, Y.; Xia, M. An Effective Method for CHF Diagnosis via Attention-Based RNN Using ECG Signals. In Proceedings of the 2020 5th International Conference on Communication, Image and Signal Processing (CCISP), Chengdu, China, 13–15 November 2020; IEEE: Chengdu, China, 2020; pp. 209–213. [Google Scholar]
  41. Liu, T.; Si, Y.; Yang, W.; Huang, J.; Yu, Y.; Zhang, G.; Zhou, R. Inter-Patient Congestive Heart Failure Detection Using ECG-Convolution-Vision Transformer Network. Sensors 2022, 22, 3283. [Google Scholar] [CrossRef]
  42. Liu, Z.; Chen, T.; Wei, K.; Liu, G.; Liu, B. Similarity Changes Analysis for Heart Rate Fluctuation Regularity as a New Screening Method for Congestive Heart Failure. Entropy 2021, 23, 1669. [Google Scholar] [CrossRef]
  43. Xi, C.; Liu, R. Detection of Small Floating Target on Sea Surface Based on Gramian Angular Field and Improved EfficientNet. Remote Sens. 2022, 14, 4364. [Google Scholar] [CrossRef]
Figure 1. Raw ECG signals. (a) Normal ECG signal; (b) CHF ECG signal.
Figure 1. Raw ECG signals. (a) Normal ECG signal; (b) CHF ECG signal.
Entropy 24 01763 g001
Figure 2. Block diagram of the ECG signals classification system proposed in this paper.
Figure 2. Block diagram of the ECG signals classification system proposed in this paper.
Entropy 24 01763 g002
Figure 3. ECG signals with attenuated noise. (a) Normal ECG signal; (b) CHF ECG signal. The wavelet basis function is ‘bior2.6’ and the number of wavelet layers is 8.
Figure 3. ECG signals with attenuated noise. (a) Normal ECG signal; (b) CHF ECG signal. The wavelet basis function is ‘bior2.6’ and the number of wavelet layers is 8.
Entropy 24 01763 g003
Figure 4. The scaling function ζ L ( q ) of the time series based on WL.
Figure 4. The scaling function ζ L ( q ) of the time series based on WL.
Entropy 24 01763 g004
Figure 5. The multifractal spectrum f ( α ) of the time series based on WL.
Figure 5. The multifractal spectrum f ( α ) of the time series based on WL.
Entropy 24 01763 g005
Figure 6. Structure of basic ELM.
Figure 6. Structure of basic ELM.
Entropy 24 01763 g006
Figure 7. MSEN1 curve at different segmentation time. (a) NSR ECG signal; (b) CHF ECG signal. The embedding dimension m = 2 , scale s = 1 : 1 : 50 , and similarity tolerance r = 0.2 σ ( σ is the standard deviation of the time series after coarse granulation).
Figure 7. MSEN1 curve at different segmentation time. (a) NSR ECG signal; (b) CHF ECG signal. The embedding dimension m = 2 , scale s = 1 : 1 : 50 , and similarity tolerance r = 0.2 σ ( σ is the standard deviation of the time series after coarse granulation).
Entropy 24 01763 g007
Figure 8. MSEN1 curves for two ECG signals at different scales. The embedding dimension m = 2 , segmentation time t = 64 s, and similarity tolerance r = 0.2 σ .
Figure 8. MSEN1 curves for two ECG signals at different scales. The embedding dimension m = 2 , segmentation time t = 64 s, and similarity tolerance r = 0.2 σ .
Entropy 24 01763 g008
Figure 9. MSEN1 curve at different similarity tolerance. (a) NSR ECG signal; (b) CHF ECG signal. The embedding dimension m = 2 , segmentation time t = 64 s, and scale s = 1 : 1 : 20 .
Figure 9. MSEN1 curve at different similarity tolerance. (a) NSR ECG signal; (b) CHF ECG signal. The embedding dimension m = 2 , segmentation time t = 64 s, and scale s = 1 : 1 : 20 .
Entropy 24 01763 g009
Figure 10. WL-based multifractal spectrum. (a) Set A; (b) Set b. The ‘db3’ wavelet basis functions q max = 5 , Δ q = 0.25 .
Figure 10. WL-based multifractal spectrum. (a) Set A; (b) Set b. The ‘db3’ wavelet basis functions q max = 5 , Δ q = 0.25 .
Entropy 24 01763 g010
Figure 11. Classification accuracy for 5-fold cross-validation.
Figure 11. Classification accuracy for 5-fold cross-validation.
Entropy 24 01763 g011
Figure 12. Block diagram of accuracy of different algorithms.
Figure 12. Block diagram of accuracy of different algorithms.
Entropy 24 01763 g012
Table 1. Total number of ECG segments used in the data set.
Table 1. Total number of ECG segments used in the data set.
Number of 64 s ECG
Segments
Type (Database)UnbalancedBalanced
AB
Normal (NSR)540360
CHF (BIDMC)360360
Table 2. Training accuracy of ELM with the number of nodes of the hidden layer.
Table 2. Training accuracy of ELM with the number of nodes of the hidden layer.
Number of ELM Hidden Layer Nodes (%)
K-Fold2025303540455055606570
198.6199.4898.9699.6599.6510099.6599.83100100100
298.7899.4899.4899.3199.4899.6599.8399.83100100100
398.9699.4899.6599.8399.8399.6599.6599.8310099.8399.65
499.3199.4899.6599.4899.4899.83100100100100100
598.6199.1398.9699.4899.6599.8399.6599.8399.83100100
Table 3. Confusion matrix for unbalanced data set-NSR/CHF (Set A).
Table 3. Confusion matrix for unbalanced data set-NSR/CHF (Set A).
Predicted
NormalCHFACC (%)PPV (%)SEN (%)SPEC (%)F1 (%)
OriginalNormal539199.5699.4499.8199.1799.62
CHF3357
Table 4. Confusion matrix for balanced data set-NSR/CHF (Set B).
Table 4. Confusion matrix for balanced data set-NSR/CHF (Set B).
Predicted
NormalCHFACC (%)PPV (%)SEN (%)SPEC (%)F1 (%)
OriginalNormal360099.7299.4610099.4499.73
CHF2358
Table 5. Classification results of CHF detection categories with different ECG fragments.
Table 5. Classification results of CHF detection categories with different ECG fragments.
DatasetTPTNFPFNACC (%)PPV (%)SEN (%)SPEC (%)F1 (%)
A5393573199.5699.4499.8199.1799.62
C17891486141199.2499.2299.3999.0799.30
D71535969314799.4199.5799.3599.4899.46
Table 6. Running time to complete a dataset.
Table 6. Running time to complete a dataset.
DatasetRunning Time (s)
A731.45
C3107.80
D11,825.24
Table 7. Accuracy of 5-fold classification for different algorithms.
Table 7. Accuracy of 5-fold classification for different algorithms.
AlgorithmFold1 (%)Fold2 (%)Fold3 (%)Fold4 (%)Fold5 (%)OA (%)
Generalized MSE + ELM92.2298.3397.2296.1195.0095.78
WL + ELM98.3399.4498.3397.7899.4498.67
Generalized MSE-WL + KNN95.0098.3395.0097.7897.2296.67
Generalized MSE-WL + SVM97.2297.7897.2296.6796.1197.00
Generalized MSE-WL + ELM (Set A)99.4410010098.8999.4499.56
Table 8. Literature Comparison.
Table 8. Literature Comparison.
ReferenceYearNumber of ECG DataMethodPerformance
Daqroup and Dobaie [38]2016CHF: 140
Normal: 152
▪Wavelet Packet Transform
▪Feature Extraction
Acc—92.60%
Sundarshan et al.
[39]
2017CHF: 25,328
Normal: 59,624
CHF: 25,328
Normal: 25,328
▪Denoising and baseline removal
▪Dual tree complex wavelet transform
▪KNN classifier
(2-s ECG segment)
Acc—98.42%
Sen—97.04%
Spec—99.01%
Acc—97.94%
Sen—98.19%
Spec—97.69%
Acharya et al.
[17]
2018CHF: 30,000
Normal: 70,308
CHF: 30,000
Normal: 30,000
▪KNN classifier
▪11-layer deep CNN
(2-s ECG segment)
Acc—95.98%
Sen—96.52%
Spec—95.75%
Acc—94.40%
Sen—94.68%
Spec—94.12%
Jahmunah et al.
[16]
2019CHF: 30,000
Normal: 70,308
CHF: 30,000
Normal: 30,000
▪Fuzzy entropy
▪Rényi entropy
▪Higuchi Fractal Dimension
▪Kraskov entropy, energy
▪Frequency localized filter banks
▪Quadratic support vector machine (QSVM)
▪10-fold cross validation
(2-s ECG segment)
Acc: > 99.66%
Sen: > 99.82%
Spec: > 99.28%
Yue Zhang and Ming Xia
[40]
2020CHF: 53,857
Normal: 58,675
▪Detected R peaks
▪RNN
Acc = 99.17%
Sen = 99.40%
Spec = 98.96%
Taotao Liu et al.
[41]
2022CHF: 36,000
Normal: 30,000
▪Feature Extraction
▪ECVT-Net
▪CNN
Acc: 98.88%
Pre: 98.84%
Sen: 98.94%
Zeming Liu et al.
[42]
20221 min length of RR segment▪Multi-feature—fApEn_IBS + IBS + LF/HF
▪Random Forces
Acc = 99.0%
Sen = 97.8%
Spec = 100.0%
Proposed Method2022CHF: 540
Normal: 360
CHF: 360
Normal: 360
▪The generalized multiscale entropy (MSE)
▪Wavelet leaders (WL)
▪ELM classifier
(64-s ECG segment)
Acc—99.56%
Sen—99.81%
Spec—99.17%
Acc—99.72%
Sen—100%
Spec—99.44%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Yang, J.; Xi, C. The Diagnosis of Congestive Heart Failure Based on Generalized Multiscale Entropy-Wavelet Leaders. Entropy 2022, 24, 1763. https://doi.org/10.3390/e24121763

AMA Style

Yang J, Xi C. The Diagnosis of Congestive Heart Failure Based on Generalized Multiscale Entropy-Wavelet Leaders. Entropy. 2022; 24(12):1763. https://doi.org/10.3390/e24121763

Chicago/Turabian Style

Yang, Juanjuan, and Caiping Xi. 2022. "The Diagnosis of Congestive Heart Failure Based on Generalized Multiscale Entropy-Wavelet Leaders" Entropy 24, no. 12: 1763. https://doi.org/10.3390/e24121763

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop