Next Article in Journal
Integration and Deployment of Cloud-Based Assistance System in Pharaon Large Scale Pilots—Experiences and Lessons Learned
Next Article in Special Issue
Blockchain Smart Contract to Prevent Forgery of Degree Certificates: Artificial Intelligence Consensus Algorithm
Previous Article in Journal
Isolation and Grading of Faults in Battery Packs Based on Machine Learning Methods
Previous Article in Special Issue
A Deep Learning-Based Action Recommendation Model for Cryptocurrency Profit Maximization
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Coronary Artery Disease Detection Model Based on Class Balancing Methods and LightGBM Algorithm

1
School of Computer Science (National Pilot Software Engineering School), Beijing University of Posts and Telecommunications, Beijing 100876, China
2
Key Laboratory of Trustworthy Distributed Computing and Service, Ministry of Education, Beijing 100876, China
3
School of Computer and Engineering, Fuyang Normal University, Fuyang 236000, China
4
Air Force Medical Center, Chinese People's Liberation Army, Beijing 100142, China
*
Author to whom correspondence should be addressed.
Electronics 2022, 11(9), 1495; https://doi.org/10.3390/electronics11091495
Submission received: 21 March 2022 / Revised: 1 May 2022 / Accepted: 2 May 2022 / Published: 6 May 2022
(This article belongs to the Special Issue Electronic Solutions for Artificial Intelligence Healthcare Volume II)

Abstract

:
Coronary artery disease (CAD) is a disease with high mortality and disability. By 2019, there were 197 million CAD patients in the world. Additionally, the number of disability-adjusted life years (DALYs) owing to CAD reached 182 million. It is widely known that the early and accurate diagnosis of CAD is the most efficient method to reduce the damage of CAD. In medical practice, coronary angiography is considered to be the most reliable basis for CAD diagnosis. However, unfortunately, due to the limitation of inspection equipment and expert resources, many low- and middle-income countries do not have the ability to perform coronary angiography. This has led to a large loss of life and medical burden. Therefore, many researchers expect to realize the accurate diagnosis of CAD based on conventional medical examination data with the help of machine learning and data mining technology. The goal of this study is to propose a model for early, accurate and rapid detection of CAD based on common medical test data. This model took the classical logistic regression algorithm, which is the most commonly used in medical model research as the classifier. The advantages of feature selection and feature combination of tree models were used to solve the problem of manual feature engineering in logical regression. At the same time, in order to solve the class imbalance problem in Z-Alizadeh Sani dataset, five different class balancing methods were applied to balance the dataset. In addition, according to the characteristics of the dataset, we also adopted appropriate preprocessing methods. These methods significantly improved the classification performance of logistic regression classifier in terms of accuracy, recall, precision, F1 score, specificity and AUC when used for CAD detection. The best accuracy, recall, F1 score, precision, specificity and AUC were 94.7%, 94.8%, 94.8%, 95.3%, 94.5% and 0.98, respectively. Experiments and results have confirmed that, according to common medical examination data, our proposed model can accurately identify CAD patients in the early stage of CAD. Our proposed model can be used to help clinicians make diagnostic decisions in clinical practice.

1. Introduction

With the development of economy and the aged tendency of population, the burden of disease and factors of death have changed dramatically all over the world. In the light of the analysis data of the World Health Statistics 2019, noncommunicable diseases have become the leading cause of death globally. In 2016, 41 million deaths, accounting for about 71% of global deaths, were caused by noncommunicable diseases. The main noncommunicable diseases incorporate cardiovascular disease, cancer, diabetes and chronic respiratory diseases [1]. Among them, cardiovascular disease (CVD) has become the leading cause of premature death and rising healthcare costs. In 2019, the worldwide cases of CVD were 523 million, and the deaths caused by CVD reached 18.5 million, which accounted for approximately one-third of all deaths universally [2,3,4]. CVD mainly includes ischemic heart disease (IHD) and stroke. By 2019, there were 197 million IHD patients in the world. Additionally, the number of disability-adjusted life years (DALYs) because of IHD reached 182 million [5,6]. The detailed analysis of the Global Burden of Diseases, Injuries, and Risk Factors Study (GBD) 2019 shows that (See Table 1), in all age groups, the top three causes of DALYs are neonatal disorders, IHD and stroke, respectively. In the 50 to 74 years age group and 75 years and older age group, the top two causes of DALYs are IHD and stroke. Additionally, in the 25 to 49 years age group, the top three reasons of DALYs are road injuries, HIV/AIDS and IHD [6]. It’s clear that IHD, a typical type of CVD, has become a major cause of death and disability.
IHD, namely, coronary atherosclerotic heart disease (coronary artery disease (CAD) for short), refers to one kind of heart disease that manifests myocardial ischemia, hypoxia or necrosis caused by coronary artery narrowing or occlusion, which is caused by atherosclerosis. To be consistent with other literature, coronary artery disease (CAD) is also used in the later part of the article. According to the anatomical structure of coronary artery, there are three main arteries supplying blood to the heart, namely, (1) left anterior descending coronary artery (LAD), (2) left circumflex coronary artery (LCX) and (3) right coronary artery (RCA). CAD occurs when the lumen of any one of the three coronary arteries is narrowed by 50% or more [7].
CAD is a highly lethal disease, so early detection and diagnosis is crucial for saving lives and improving prognosis. At present, the auxiliary examinations used in CAD diagnosis in medical include laboratory examination, electrocardiogram (ECG) examination, radionuclide examination, multislice spiral CT coronary angiography imaging, echocardiography and coronary angiography. Other examination methods include fractional flow reserve (FFR), optical coherence tomography and intravenous ultrasound (IVUS). Among them, coronary angiography is regarded as the gold standard for the diagnosis of CAD. Unfortunately, many developing and low-income countries do not have the equipment and specialized doctors to perform coronary angiography. Moreover, imaging examinations with high cost and technical requirements are difficult to be popularized in many countries and regions. Therefore, more and more doctors, scholars and researchers are devoted to finding other methods that can detect and diagnose CAD early. Machine learning and data mining technology are such a field that they focus on. In the past few years, machine learning and data mining technology have been broadly applied in various aspects of medical researches, for instance disease screening, disease risk stratification, disease prediction and assistant decision-making [8,9,10].
Similarly, machine learning technology has also played an important role in the prediction tasks of CAD. Numerous studies have been carried out. According to the data types used in study, these studies can be divided into studies based on the ECG signals [11,12,13,14,15,16,17,18,19,20,21,22], studies based on the imaging data [23,24,25,26,27,28,29], and studies based on the data of multiple routine examination items. As mentioned above, many imaging examinations are difficult to popularize in some countries and regions. Therefore, there are certain regional restrictions on the application of research based on the imaging data. The research based on the ECG signal that is easy to obtain can only identify the CAD with significant ECG changes. Nevertheless, for some hidden CAD, it is difficult to diagnose these CAD timely and accurately. The research based on the data of multiple routine examination items has the advantages of easy access to data and comprehensive reflection of CAD. Therefore, this study proposed a machine learning model for early and accurate detection of CAD based on conventional medical examination data. This study was carried out on Z-Alizadeh Sani dataset, which is the latest dataset with multiple examination indicators.
The rest of the paper is arranged as follows. Section 2, related work, summarizes the related research based on Z-Alizadeh Sani dataset in the past decades. Section 3 introduces the dataset used in our research. An introduction of our proposed method is exhibited in Section 4. Section 5 shows detailed information about the experiments and results, and provides a detailed analysis of the results. Section 6 discusses this study. The conclusion of the study is expressed in Section 7.

2. Related Work

In this section, we review the application and results of machine learning technology on CAD prediction tasks in the past ten years. We emphatically analyze the progress in technology and design of research based on the data of multiple routine examination items.
Alizadehsani et al. applied feature selection and feature construction techniques to process input data, and used multiple classification models on it. Finally, 94.08% classification accuracy was obtained on the sequence minimum optimization (SMO) model [30]. Elham et al. developed a technique called heterogeneous hybrid feature selection to select features that were related to CAD diagnosis from the Z-Alizadeh Sani dataset. Then, they applied two oversampling techniques, namely, synthetic minority oversampling technology (SMOTE) and adaptive synthetic (ADASYN) to deal with the imbalanced class problem in dataset. The classification accuracy gained was 92.58% [31]. Arabasadi et al. improved the classification performance of neural network on Z-Alizadeh Sani dataset by using a genetic algorithm to increase the initial weight value of the neural network. The classification performance with accuracy of 93.85%, specificity of 92% and sensitivity of 97% was achieved [32]. Alizadehsani et al. used the meta cost-sensitive algorithm to distinguish patients with CAD from healthy individuals. Naive bayes, support vector machines (SVM), k-nearest neighbors (KNN), SMO, and C4.5 algorithm were used for classification. The best classification result was obtained by SMO algorithm with accuracy of 92.09% and sensitivity of 97.22% [33]. Zomorodi-Moghadam et al. proposed a method discovering the rules of CAD classification. By using the feature selection method based on particle swarm optimization (PSO) and multi-objective evolutionary search, this method selected the features most related to CAD classification, and formed two different rule sets with 11 features and 13 features, respectively. The classification accuracy was evaluated on two rule sets. The results showed that this method has the ability to generate effective CAD detection rules [34].
Abdar et al. proposed a new nested integrated kernel support vector classification (NE-nu-SVC) model. The model improved the performance of traditional machine learning methods by applying ensemble learning technology, feature selection method and data balancing method. The model obtained 94.66% classification accuracy [35]. Abdar et al. introduced a method called N2Genetic optimizer to improve the performance of traditional algorithms. On the Z-Alizadeh Sani dataset, the method achieved an accuracy of 93.08% and F1 score of 91.51% [36]. Alizadehsani et al. realized the prediction of CAD by constructing a classifier for each coronary artery of LAD, LCX, and RCA. It was used to the extended Z-Alizadeh Sani dataset containing 500 patients and achieved 96.40% CAD detection accuracy [37]. Shahid et al. utilized four different feature selection means: Relief-F, Fisher, Weight by SVM, and Minimum Redundancy Maximum Relevance to improve the performance of emotional neural networks (EmNNs) on Z-Alizadeh Sani dataset. Finally, they obtained 88.34% classification accuracy [38]. Wang et al. developed a two-level stacking model based on stacking ensemble learning idea for CAD detection. This model showed 95.43% detection accuracy on Z-Alizadeh Sani dataset [39]. Tama et al. designed a two-tier integration framework. The first layer was feature selection, which integrated two methods of feature selection, namely correlation-based feature selection (CFS) and PSO. The second layer was classifier modeling, which mixed the class label of three integrated learners through the stacking architecture. The integration framework achieved 98.13% accuracy on the Z-Alizadeh Sani dataset [40]. Gupta et al. proposed a system (C-CADZ) for CAD detection. C-CADZ automatically realized feature extraction, feature selection, class balance and model prediction. This system used random forest (RF) and extreme tree (ET) models as machine learning classifiers and obtained 97.37% prediction accuracy on Z-Alizadeh Sani dataset [41]. Kolukisa et al. processed the Z-Alizadeh Sani dataset by applying linear discriminant analysis (LDA) dimensionality reduction technique and a hybrid feature selection technique that combined four feature selection methods, namely, gain ratio (GR), information gain (IG), Relief-F (RF) and chi -square (CS) test. Finally, a multiclass fisher linear discriminant analysis (FLDA) ensemble classifier was used to classify, and the accuracy was 92.07% [42].
Dekamin et al. adopted K-means algorithm to preprocess data and used naive bayes, KNN and decision tree to classify. This method achieved 90.91% efficiency [43]. Alizadehsani et al. performed CAD prediction research based on laboratory and echocardiography data. This research used SMO, naive Bayes, C4.5 and AdaBoost algorithms for classification prediction and achieved a classification accuracy of more than 82% [44]. Yadav et al. proposed a CAD prediction method based on association rule mining, and obtained 92.09% classification accuracy on SMO classification algorithm [45]. Ghiasi et al. applied a decision tree learning algorithm called CART to the detection of CAD, and achieved 92.41% accuracy [46]. Joloudari et al. proposed a hybrid machine learning model. The model took SVM as the basic classifier, used analysis of variance as the kernel function of SVM, and used genetic optimization algorithm for feature selection. This model obtained 89.45% classification accuracy [47]. In our previous research, we applied four feature processing techniques and two kinds of class balancing methods to develop a CAD prediction model based on Random Forest algorithm and XGBoost algorithm. The model effectively realized the early detection of CAD and achieved 94.7% prediction accuracy [48].
In the past ten years, researchers have carried out a large number of exploratory research on CAD prediction based on machine learning technology on the dataset containing the results of multiple routine examination items. In reference [49], Alizadehsani et al. have analyzed 149 literatures related to the subject of machine learning-based CAD prediction published from 1992 to 2019. Reference [49] mainly summarized the investigated papers from the aspects of classification algorithm, feature selection algorithm and model evaluation. At the level of classification algorithm, most studies have applied traditional machine learning algorithms, such as SVM, Naive Bayes, KNN, decision tree, SMO, C4.5 and artificial neural network. Among them, artificial neural network and decision tree algorithm were the two most widely used algorithms. In the application of feature selection algorithm, the most commonly used methods include information gain, Gini index, PCA and weight by SVM. In the model evaluation, the application proportions of accuracy, recall, specificity and precision were 96%, 68.8%, 63.2% and 65.6%, respectively. The application proportions of F measurement and AUC was less than 20%.
Through the analysis, it can be found that the previous studies have made some achievements and progress, but there are still some limitations. For example, firstly, in the application of classification algorithm, there is less research on the application of classical logistic regression algorithm suitable for small datasets. Secondly, these kinds of data sets often have the problem of class imbalance. Several articles used some common sampling methods and cost sensitive algorithms to solve this problem, such as SMOTE algorithm and ADASYN algorithm. However, there are few studies on the application of more sampling methods. In addition, in terms of model evaluation indicators, the application of AUC and F measurement indicators is insufficient. However, these two indexes are very important to evaluate the performance of the CAD prediction model. Especially in the dataset with class imbalance problem, the accuracy index often cannot accurately reflect the ability of the model.
In view of this, in this study, we proposed a machine learning model for early, accurate and rapid detection of CAD based on common medical detection data. The model used the classical logistic regression algorithm, which is the most commonly used in medical model research as the classifier. Additionally, we applied the advantages of feature selection and feature combination of tree models to solve the problem of manual feature engineering in logical regression algorithm. At the same time, five different sampling methods were applied to solve the class imbalance problem of Z-Alizadeh Sani dataset. Accuracy, recall, specificity, precision, F1 score, ROC and AUC were used to evaluate the performance of the model. Ten-fold cross-validation technology was also applied in the study. In addition, according to the characteristics of the dataset, we also adopted appropriate preprocessing methods. The frame diagram of our proposed machine learning model for CAD prediction is shown in Figure 1. The list of abbreviations used in this paper is shown in Table A1 of Appendix A.

3. Dataset

From UCI Machine Learning Repository we downloaded the Z-Alizadeh Sani dataset used by us. The Z-Alizadeh Sani dataset is the latest dataset with multiple examination indicators. This dataset contains 303 medical records derived from 303 cases who visited Shaheed rajaei hospital because of chest pain. Each record includes 55 features belonging to four categories. Four categories are demographic features; symptoms and physical examination; ECG; and laboratory tests and echocardiography features. A record is a sample. These 303 samples belong to two classes, namely, CAD class and normal class. When the stenosis of coronary arteries lumen of a sample reaches or exceeds 50%, this sample is classified as CAD class; otherwise it belongs to the normal class. Accordingly, in 303 samples, 216 instances accounting for 71.29% are CAD class, and 87 instances accounting for 28.71% are normal class [30]. Details of the Z-Alizadeh Sani dataset are shown in Table 2. To intuitively explain the features of Z-Alizadeh Sani dataset, in Figure 2, we showed an ECG waveform and an echocardiac image. The ECG features of Z-Alizadeh Sani dataset are obtained by professional doctors analyzing the ECG waveform in Figure 2a. Similarly, the echocardiography features of Z-Alizadeh Sani dataset are obtained by professional doctors detecting the echocardiac image in Figure 2b.
The 55 features of Z-Alizadeh Sani dataset can be split into two types: continuous features and categorical features. The statistical description of continuous features is shown in Table 3.
By analyzing Table 2 and Table 3, it can be seen that Z-Alizadeh Sani dataset has the following characteristics: (1) Exertional CP feature has the same values on 303 samples. In other words, External CP feature does not contribute to the classification prediction of CAD; (2) the dimensions of continuous features are not uniform, and the distribution of continuous features has a certain skewness; and (3) the dataset has class imbalance problem.

4. Proposed Method

This section describes the machine learning techniques used by the study in detail.

4.1. Preprocessing of Data

In view of the characteristics of Z-Alizadeh Sani dataset, we processed the dataset as follows: (1) delete the External CP feature directly; and (2) standardize the continuous features.
The method of data standardization is as follows:
Χ = { χ ( 1 ) , χ ( 2 ) , χ ( 3 ) , χ ( m ) }
where Χ is sample set, χ ( 1 ) , χ ( 2 ) , χ ( 3 ) , χ ( m ) are samples of Χ , m is the number of samples.
Suppose that sample χ ( i ) of Χ has n features, that is:
χ ( i ) = { χ 1 ( i ) , χ 2 ( i ) , χ 3 ( i ) , χ n ( i ) }
Then, for the continuous feature j there is:
μ j = 1 m i = 1 m χ j i
where χ j ( i ) is the value of feature j on sample χ ( i ) . μ j is the average value of feature j . Additionally, there is:
σ j = 1 m i = 1 m ( χ j i μ j ) 2
Thereinto, σ j is the standard deviation of feature j . Then:
χ j * i = χ j i μ j σ j
χ j * i is the value of feature j on sample χ ( i ) after standardization.

4.2. Methods of Balancing Classes

We applied five different methods of classes balancing on Z-Alizadeh Sani dataset. We agree that the sample set of minority class is S min , the sample set of majority class is S m a j , x 1 , x 2 , x 3 , x m are the samples of S min , m is the number of samples of S min , and x i refers to any sample of S min .

4.2.1. SMOTE

Synthetic Minority Oversampling Technique (SMOTE) [50,51,52,53,54] is one of the most commonly used and classic oversampling methods. Simply put, SMOTE synthesizes new samples artificially by analyzing the distance between minority samples and their nearest neighbors. The specific process of synthesizing samples is as follows: (1) distance calculation: calculate the distance between x i and all samples of S min ; thus, the k -nearest neighbors of sample x i are obtained; (2) nearest neighbors selection: set the number of oversampling of each sample of minority class according to the ratio of S min and S m a j , assuming this number is n . Then, n samples are randomly selected from the k -nearest neighbors of sample x i . Here, we assume that x j is one of the n -nearest neighbors samples; (3) new samples synthesis: the new sample x n e w between the sample x i and x j is synthesized according to the following formula:
x n e w = x i + r a n d ( 0 , 1 ) × | x i x j |
where r a n d ( 0 , 1 ) refers to generate a random number uniformly distributed between 0 and 1, | x i x j | is the distance between the sample x i and x j .
All new samples are synthesized in the same way until the class balance requirements are met. The essence of SMOTE algorithm is to select a random point on the connection line of two specific samples as a new sample. This method increases the number of minority samples effectively.

4.2.2. Borderline_SMOTE

Unlike SMOTE, which synthesizes new samples for each minority sample, Borderline_SMOTE algorithm [55] only resamples or strengthens the minority examples at the class borderline. Firstly, the k -nearest neighbors of sample x i are found. Then, the number of samples belonging to the majority class in the k -nearest neighbors of sample x i is analyzed, assuming this number is b . When b k 2 , sample x i is put into the SAFE set. When b k 2 , sample x i is put into the DANGER set. When b = k , sample x i is put into the NOISE set, as shown in Figure 3. Finally, the minority samples belonging to the DANGER set are resampled to synthesize new samples. The process of new sample synthesis is the same as SMOTE.
Borderline_SMOTE algorithm includes Borderline_SMOTE1 and Borderline_SMOTE2. Borderline_SMOTE1 will randomly select minority sample from the k -nearest neighbors of sample x i to synthesize new samples. While Borderline_SMOTE2 will randomly select any sample from the k -nearest neighbors of sample x i to synthesize new samples. In this article, we applied Borderline_SMOTE1.

4.2.3. SVM_SMOTE

SVM_SMOTE algorithm [56] is the combination of SVM algorithm and SMOTE algorithm. Similar to Borderline_SMOTE algorithm, SVM_SMOTE algorithm also focuses on the minority samples at the class borderline and only resamples or strengthens these samples. Firstly, the SVM classifier is applied on the training set to obtain the support vector, which is approximately the boundary region. Then, the new minority samples will be synthesized according to the following decision mechanism. The decision mechanism for synthesizing minority class samples depends on the distribution density of majority class samples around minority class support vector samples (See Figure 4). When more than half of the k -nearest neighbor samples of a minority class support vector sample belong to the minority class, the new minority class samples will be synthesized by an external interpolation mechanism. When more than half of the k -nearest neighbor samples of a minority class support vector sample belong to the majority class, the new minority class samples will be synthesized by an internal interpolation mechanism. When all the k -nearest neighbor samples of a minority class support vector sample belong to the majority class, it is considered that the minority class support vector sample is noise data and should be relabeled.
The essence of SVM_SMOTE algorithm is over-sampling based on support vector. Near the class boundary approximated by the support vector, different decision mechanisms are selected to synthesize the minority class samples according to the distribution density of the majority class samples around the minority class support vector. After applying SVM_SMOTE algorithm, the minority groups can be expanded to areas with low sample density of the majority class.

4.2.4. SMOTE_Tomek

The above three methods are all oversampling methods, which achieve class balance by increasing the number of minority class samples. However, the SMOTE_Tomek algorithm [57] introduced in this section is a comprehensive sampling method combining oversampling and undersampling. Before introducing the SMOTE_Tomek algorithm in depth, we need to understand the definition of Tomek links. Simply put, Tomek links is defined as a pair of connections between two nearest neighbor samples belonging to the opposite class. Assuming that x i and x j are two samples belonging to different class, d ( x i , x j ) is the distance between x i and x j . If there is no sample x e , which can make d ( x e , x i ) < d ( x i , x j ) or d ( x e , x j ) < d ( x i , x j ) , the sample pair ( x i , x j ) is a Tomek link. Two samples of a Tomek link, either one of them is noise, or both of them are in the boundary area. Therefore, Tomek links can be used for both data undersampling and data cleaning. When Tomek links are used as an undersampling method, samples belonging to the majority class in Tomek links are deleted. When Tomek links are used as a data cleaning method, two samples in Tomek links are deleted.
The SMOTE_Tomek algorithm applied in this paper takes advantage of the data cleaning effect of Tomek links. Firstly, SMOTE algorithm was applied to the original dataset, and a balanced dataset T S M O T E was obtained. Then, we found the Tomek links from T S M O T E and deleted the two samples in Tomek links. Finally, a new balanced dataset was obtained for training and testing.

4.2.5. SMOTENC

SMOTENC algorithm [51] is an oversampling method that can deal with categorical features. It is assumed that sample x i is a minority class sample containing both continuous and categorical features, sample x n e w is a new sample synthesized from sample x i by SMOTENC algorithm, and feature j is one of the categorical features. The value of feature j of sample x n e w is the value with the highest frequency among the values of feature j of the k -nearest neighbors of sample x i .

4.3. Feature Combination

In this study, logistic regression (lr) algorithm is used as a classifier for classification and prediction. Logistic regression is a classification model with the advantages of easy understanding, easy implementation and fast operation. However, when used for classification, a large number of feature engineering, such as feature extraction and feature combination, need to be carried out manually to improve the performance of lr classifier. Undoubtedly, it takes a lot of time to carry out feature engineering manually. Therefore, many researchers tried to improve the classification ability of lr through automatic feature engineering.
For a long time, decision tree model and neural network model have shown better performance in disease classification and prediction tasks [58,59,60,61]. The former has good interpretability, while the latter shows high efficiency on large data sets. The splitting process of the model based on decision tree algorithm in each node is the process of feature screening. Each leaf node on the decision tree is the output of the decision tree trained on a set of effective feature subsets. That is, in the decision tree, the path from the root node to each leaf node is a feature combination process. Each leaf node contains a set of unique feature combination information. Therefore, models based on decision tree algorithm often have good functions of feature selection and feature combination. Deepfm model is a classic model in neural network, which was proposed by Huawei Noah Ark laboratory in 2017 [62]. Deepfm model is the combination of factorization machines (FM) and neural network structure. Deepfm model can automatically learn low-order explicit feature combination and high-order implicit feature combination at the same time. Therefore, the Deepfm model can also be used for automatic feature combination.
Using the feature combination function of models based on decision tree algorithm and Deepfm model to improve the performance of lr classifier has been widely used in the prediction task of advertising click through rate, and has shown good performance. However, in the prediction task of CAD, the research on this method has not been carried out. Therefore, in this paper, we will utilize the feature combination advantages of decision tree model or Deepfm model to improve the performance of lr classifier on Z-Alizadeh Sani dataset.
To select the best tree model for feature combination, we applied five classification models based on ensemble learning technology to the original dataset. The five classification models are random forest (RF), extratrees, adaptive boosting (AdaBoost), eXtreme gradient boosting (XGBoost) and light gradient boosting machine (lightGBM). The performance results of the five models on the original dataset are shown in Table 4. It can be seen from Table 4 that on the original dataset the lightGBM model has the best performance. In addition, we also applied the DeepFM model to the original dataset. Unfortunately, the highest accuracy obtained by DeepFM model was only 83.52%. This may be related to the small sample size of the dataset. Therefore, we select the ligthGBM model with the best performance to play the function of feature combination for lr classifier.
LightGBM [63] is a machine learning algorithm framework based on boosting ensemble learning technology proposed by Microsoft in 2017. Compared with GBDT (Gradient Boosting Decision Tree) algorithm and XGBoost algorithm, lightGBM algorithm can not only achieve the same prediction performance, but also has more obvious advantages in training speed and memory consumption. On the basis of GBDT algorithm and XGBoost algorithm, lightGBM algorithm integrates several optimization strategies, such as histogram algorithm, grandient-based one-side sampling (GOSS) algorithm, exclusive feature bundling (EFB) strategy, leaf-wise strategy, supporting category feature strategy and supporting efficient parallel strategy. These optimization strategies make lightGBM become a classification model with high efficiency, low consumption, more accuracy and more convenient. The parameter boosting_type was set to GBDT (Gradient Boosting Decision Tree) [64,65,66,67] algorithm when we used the lightGBM algorithm.
The method where lightGBM algorithm plays the function of feature combination for lr classifier is to use the output of lightGBM algorithm as the input of lr algorithm for training to obtain the final prediction output. This is the application of stacking ensemble learning technology. It is worth noting that the input of lr algorithm is not the category labels (such as 0 or 1) predicted by lightGBM algorithm for each instance, but the index of leaf node of each instance on each decision tree. These indexes need to be processed by One-Hot Encoding before inputting lr algorithm. The specific process is as follows:
(1)
LightGBM model is applied to the training set, and the trained LightGBM classifier is obtained.
(2)
After the training, the index of leaf node on each decision tree in the lightGBM model is output for each instance of the training set. After all iterations, all indexes of each instance form a new set of features. At this time, a set of m × n -dimensional dataset is formed, where m is the size of the training set, and n is the number of weak estimators (decision trees) in the lightGBM model.
(3)
The m × n -dimensional dataset obtained in step (2) is encoded by One-Hot Encoding, and a sparse matrix M t r a i n with m × n × l -dimensional is obtained, where l is the number of leaf nodes per decision tree in the lightGBM model. The sparse matrix M t r a i n is the training set of lr algorithm.
(4)
lr model is applied to the sparse matrix M t r a i n , and the trained lr classifier is obtained.
(5)
Similarly, the testing set is processed by steps (1)–(3) to obtain the sparse matrix M t e s t , and the sparse matrix M t e s t is entered into the trained lr classifier to obtain the final prediction.
The above is the process of realizing automatic feature combination based on lightGBM model (see Figure 5). In addition, we also tried another feature combination method, that is, the method of combination the sparse matrix output by ightGBM model with the original feature set. In other words, in this study, we tried two methods of feature combination. One is to input the sparse matrix M t r a i n combined by lightGBM mode directly into the lr model for training. The other is to recombine the sparse matrix M t r a i n combined by lightGBM mode with the training set of the original feature set and then input them into the lr model for training. For the sake of distinction, we record the lr classifier trained by the former as lightGBM + lr, and the lr classifier trained by the latter as lightGBM + LR.

4.4. Classification Algorithm

Logistic Regression (lr) algorithm [68] is one of the most commonly used exploration tools in medical research, especially in the medical binary prediction task, lr model has a wide range of applications. The classification decision-making process of lr algorithm for a specific sample is shown in Equations (7)–(10).
For binary classification problem, suppose any sample X has:
X = { x 0 , x 1 , x 2 , x 3 , , x n }
where n is the number of features contained in sample X , x 0 is the bias term, x i is the value of sample X on the i th feature, i = 1 , 2 , 3 , n .
Then, the probability P ^ (or called it as decision function h θ ( X ) ) that the lr model predicts sample X as a positive class is calculated according to the following equation:
P ^ = h θ ( X ) = σ ( X T θ )
where θ is the weight vector and θ = { θ 0 , θ 1 , θ 2 , θ 3 , , θ n } , θ i is the weight value corresponding to x i , i = 0 , 1 , 2 , 3 , , n . σ is the sigmoid function i.e., σ ( t ) = 1 1 + exp ( t ) , then:
P ^ = h θ ( X ) = 1 1 + exp ( X T θ )
h θ ( X ) is the probability that the sample X belongs to the positive class calculated by the lr model. Combined with the classification threshold, the prediction class of sample X can be obtained. Assuming that the classification threshold is set to 0.5, the prediction function of lr model is:
y ^ ( X ) = { 1 ,   i f   h θ ( X ) 0.5 0 ,   i f   h θ ( X ) < 0.5
where y ^ ( X ) represents the prediction function, 0 and 1 represent the class codes of negative class and positive class, respectively, 0.5 is the classification threshold.
The above is the process of lr model classifying a specific sample. It can be summarized as follows: firstly, the probability that sample X in Equation (7) belongs to positive class is calculated through Equations (8) and (9). Then, the probability is compared with the classification threshold according to Equation (10). At this time, the prediction class of sample X can be obtained.
In the training process of the model, we directly take the classification evaluation index as the objective function of key hyperparametric optimization. In order to take into account the performance of the two classes at the same time, we take the evaluation index FI score as the objective function and use the learning curve to find the optimal hyperparameter. The hyperparameters setting of the proposed model and an example is shown in Table A2 of Appendix A.

5. Experiments Results

5.1. Evaluation Metrics

In order to comprehensively evaluate the classification performance and effectiveness of our proposed method, we applied accuracy, recall, F1 score, precision, specificity, ROC and AUC evaluation metrics. For the sake of expression of the significance and calculation formula of these evaluation metrics, we introduced the confusion matrix (See Table 5) first. The confusion matrix is a specific matrix used to visually present the performance of the algorithm. The confusion matrix of binary classification consists of two rows and two columns. Rows represent the true labels of the two classes in the dataset (denoted y true ). Columns represent the predicted label of the two classes acquired by the model (denoted as y pre ). As shown in Table 5, the confusion matrix of binary classification includes four indicators: TN, FN, FP and TP. The four indicators are defined as follows. We specified that the label of positive class is 1 and the label of negative class is 0.
TN (true negative) refers to the number of correctly predicted samples in the samples with the real class label of 0.
FN (false negative) refers to the number of incorrectly predicted samples in the samples with the real class label of 1.
FP (false positive) refers to the number of incorrectly predicted samples in the samples the real class label of 0.
TP (true positive) refers to the number of correctly predicted samples in the samples with the real class label of 1.

5.1.1. Accuracy

Accuracy refers to the proportion of samples that can be correctly predicted by the model in all samples. The calculation equation of accuracy is as follows. TN, TP, FN and FP refer to true negative, true positive, false negative and false positive, respectively.
Accuracy = TN + TP TN + TP + FN + FP
Accuracy is one of the most frequently used and most important model performance evaluation metrics. However, in the dataset with a class imbalance problem, due to the influence of majority class samples, the accuracy is often difficult to accurately measure the classification ability of the model. Therefore, in the dataset with a class imbalance problem, in addition to accuracy, more evaluation indicators need to be applied.

5.1.2. Recall

Recall refers to the proportion of samples that can be correctly predicted by the model in all samples with positive real class labels. Recall is an important indicator to measure the ability of model to identify positive samples. In medical models, it is necessary to pay attention to recall. Recall is calculated according to the following equation:
Recall = TP TP + FN
where TP and FN are true positive and false negative, respectively. In medical application, the cost of undiagnosed positive cases and wrongly diagnosed negative cases is different. The former may cause loss of life, while the latter may lead to excessive treatment. Compared with the former, the latter costs less. At the time of diagnosis, doctors and patients pay more attention to the detection of positive cases. Therefore, the recall is one of the important indicators to judge whether the model can be applied in practice.

5.1.3. Precision

Precision, like recall, is an important indicator to measure the ability of the model to correctly predict positive samples. Precision refers to the proportion of samples with positive real class labels among all samples predicted as positive by the model. According to the definition of precision, its calculation formula is as follows. TP and FP in the formula are true positive and false positive, respectively.
Precision = TP TP + FP

5.1.4. F1 Score

Sometimes the performance of the model evaluated by recall and precision may be show the opposite result, that is, one index has a good result but the other index has a poor result, so the ability of the model cannot be evaluated accurately. Therefore, F1 score is introduced. F1 score combines the results of recall and precision, and is the weighted harmonic mean of recall and precision. Only when the results of recall and precision are good, the F1 score will be higher. The higher the F1 score is, the better the model classification effect is. The following is the calculation formula of F1 score.
F 1 = 2 × Precision × Recall Precision + Recall

5.1.5. Specificity

Specificity refers to the proportion of samples that can be correctly predicted by the model in all samples with negative real class labels. Specificity measures the ability of the model to recognize negative samples. Specificity is calculated according to the following formula. TN and FP correspond to true negative and false positive, respectively.
Specificity = TN TN + FP

5.1.6. ROC and AUC

Area under curve (AUC) is the area under the receiver operating characteristic (ROC) curve. The ROC curve is drawn with the false positive rate (FPR) as x-axis and the true positive rate (TPR) as y-axis. ROC curve intuitively reflects the relationship between specificity and recall. The value of AUC is between 0 and 1, when the value of x-axis (i.e., the false positive rate (FPR) of the model) is closer to 0, and the value of y-axis (i.e., the true positive rate (TPR) of the model) is closer to 1, the value of AUC is closer to 1. The closer the AUC value is to 1, the higher the prediction performance of the classifier.

5.2. Experimental Results

In this section, we reported the classification results obtained by our proposed model on various data sets used in our study. The classification results are described from the accuracy, recall, precision, F1 score, specificity, ROC and AUC indicators. In order to explore the impact of data preprocessing method and class balancing methods on the classification performance of the model, we applied five class balancing methods on the original dataset and the standardized dataset, respectively. After the above processing, a total of 12 datasets were generated for research. At the same time, in order to verify whether our proposed method has competitive advantage in CAD prediction task, we also output the prediction results of lightGBM model and lr model without feature combination on each dataset during the experiment. Therefore, we obtained four groups of experimental results on each dataset. To facilitate narration, we record the classifiers that produce the four groups of experimental results as lightGBM, lr, lightGBM + lr and lightGBM + LR, respectively.

5.2.1. Results Obtained on Source Dataset

The performance results of the classification models used for CAD prediction on the source dataset are reported in this section. There is a certain class imbalance in the source dataset, which shows that 216 instances belong to CAD class (accounting for 71.29%) and 87 instances belong to the normal class (accounting for 28.71%). The average results of 10 tests in terms of accuracy, recall, precision, F1 score, specificity and AUC obtained by the classification models on the source dataset are shown in Table 6. Figure 6 is an intuitive display of Table 6. Figure 7 shows the ROC curve of the classification model on the 10 fold test set and the AUC value corresponding to each fold ROC curve.
It can be seen from Table 6 and Figure 6 that on the original dataset, the performance of lr model without feature combination processing is weaker than that of the ensemble algorithm ligthGBM. However, the classification performance of lr model is significantly improved after combining our proposed feature combination processing method. On the original dataset with new features combined by ligthGBM, lr model obtains the classification result better than ligthGBM model, which is also the best classification result on the dataset. The highest accuracy is 91.4%, recall is 92.5%, F1 score is 94.2%, precision is 96.3%, specificity is 88.6%, and AUC is 0.93. Figure 7 shows the ROC curve and corresponding AUC value of each fold in the 10-fold cross-validation obtained by classification model on the original dataset. It can be seen from the curve that, due to the distribution difference of 10 fold data, the AUC values obtained by the model on each fold test set are different. In addition, although the average AUC of the four classification models on 10 fold data are all about 0.93, the AUC disturbance of lightGBM + lr model, which has a standard deviation of 0.06, is slightly larger than that of the other three models.

5.2.2. Results Obtained on Dataset Processed by SMOTE

This part reports the performance results of the classification models for CAD prediction on the balanced dataset processed by SMOTE algorithm. The balanced dataset after SMOTE algorithm processing contains 432 sample instances, with CAD class and normal class accounting for 50%, respectively. The average results of 10 tests on accuracy, recall, precision, F1 score, specificity and AUC obtained by the classification models on the dataset processed by SMOTE are shown in Table 7. Figure 8 is a visual display of Table 7. Figure 9 shows the ROC curve of the classification model on the 10 fold test set of the dataset processed by SMOTE and the AUC value corresponding to each fold ROC curve.
As can be seen from Table 7 and Figure 8, compared with Table 6 and Figure 6, the classification performance of the four classification models in terms of accuracy, recall, specificity and AUC has been significantly improved on the balanced dataset processed by SMOTE. The performance of four classification models on the dataset used in this section is also different. Similarly, on the balanced dataset after SMOTE processing, the performance of lr model without feature combination processing is still worse than lightGBM. After the feature combination processing of lightGBM model, the performance of lr classifier in accuracy, recall, F1 score, precision, specificity and AUC has been significantly improved. From the point of view of all classification indicators, the best classification performance appears in the lightGBM + LR model, with accuracy of 94.0%, recall of 94.5%, F1 score 94.2%, precision of 94.4%, specificity of 93.5% and AUC of 0.97. However, from the recall alone, the best recall result appears in the lightGBM + lr model, which is 95.0%. Figure 9 shows the ROC curve and corresponding AUC value of each fold in the 10-fold cross-validation obtained by classification model on the dataset balanced by SMOTE. It can be seen from the curve that, due to the distribution difference of 10 fold data, the AUC values obtained by the model on each fold test set are different. However, it is obvious that model lightGBM + LR has the best AUC value.

5.2.3. Results Obtained on Dataset Processed by Borderline_SMOTE

In this part, we describe the classification results of the models on the balanced dataset processed by Borderline_SMOTE. The results are described based on six classification indicators: accuracy, recall, precision, F1 score, specificity and AUC. The balanced dataset processed by Borderline_SMOTE contains 432 sample instances, including 216 cases of CAD class and 216 cases of normal class. The average results of 10 tests on six classification indicators obtained by classification models on the balanced dataset processed by Borderline_SMOTE are exhibited in Table 8. Figure 10 is an intuitive display of Table 8. Figure 11 shows the ROC curve of the classification model on the 10 fold test set of the dataset processed by Borderline_SMOTE and the AUC value corresponding to each fold ROC curve.
By comparing Table 8 and Table 6, we can find that Borderline_SMOTE class balancing method improves the classification performance of models in terms of accuracy, recall, specificity and AUC. On this dataset, the feature combination function of lightGBM model can still significantly improve the performance of lr model in terms of accuracy, recall, F1 score, precision, specificity and AUC. The best classification results in this dataset are obtained by lr model on the dataset combining the combined features output by lightGBM and the original feature set, with accuracy of 93.8%, recall of 93.8%, F1 score 94.0%, precision of 94.8%, specificity of 93.8% and AUC of 0.97. The best recall result in this section appears in the lightGBM + lr model, which is 94.9%. Similarly, when comparing the classification ability of lightGBM and lr model without feature combination processing separately, the classification ability of lightGBM model is stronger. Figure 11 shows the ROC curve and corresponding AUC value of each fold in the 10-fold cross-validation obtained by classification model on the dataset balanced by Borderline_SMOTE. It can be seen from the curve that, due to the distribution difference of 10 fold data, the AUC values obtained by the model on each fold test set are different. The lr model without any feature processing in the four models obtained the lowest AUC average value on the 10 fold data. After the feature combination processing of ligthGBM algorithm, the AUC value obtained by lr model has been improved.

5.2.4. Results Obtained on Dataset Processed by SMOTE_SVM

The classification performance of the classifiers on the dataset after SMOTE_SVM processing is described in this part. The balanced dataset processed by SMOTE_SVM contains the same number of CAD class instances and normal class instances. Each class includes 216 samples, and the entire dataset contains 432 sample instances. The average results of 10 tests on accuracy, recall, precision, F1 score, specificity and AUC obtained by the classifiers on the dataset processed by SMOTE_SVM are displayed in Table 9. Figure 12 is the visual presentation of Table 9. Figure 13 shows the ROC curve of the classification model on the 10 fold test set of the dataset processed by SMOTE_SVM and the AUC value corresponding to each fold ROC curve.
It can be seen from Table 9 and Figure 12 that when compared with Table 6 and Figure 6, the classification performance of the models on the dataset processed by the SMOT_SVM method is significantly improved, mainly reflected in accuracy, recall, specificity and AUC evaluation indicators. On the dataset used in this section, the feature combination method proposed by us has significantly improved the classification ability of lr model. However, the improved results do not exceed the results obtained by the lightGBM model. Therefore, on the dataset studied in this part, lightGBM classifier produced the best classification results, with accuracy of 93.1%, recall of 93.7%, F1 score of 93.3%, precision of 93.5%, specificity of 92.5% and AUC of 0.97. Figure 13 shows the ROC curve and corresponding AUC value of each fold in the 10-fold cross-validation obtained by classification model on the dataset balanced by SMOTE_SVM. It can be seen from the curve that, due to the distribution difference of 10 fold data, the AUC values obtained by the model on each fold test set are different. Among the four models, the lr model without any feature processing has the lowest average AUC value on the 10 fold data. On the dataset processed by ligthGBM, the AUC value of lr model increases slightly, but it is lower than the AUC value obtained by ligthGBM model itself.

5.2.5. Results Obtained on Dataset Processed by SMOTE_Tomek

The dataset used in this section is processed by SMOTE_Tomek method. The dataset consists of 390 sample instances, among them 195 cases belong to CAD class and 195 cases belong to normal class. The average results of 10 tests on accuracy, recall, precision, F1 score, specificity and AUC obtained by the classification models on the dataset processed by SMOTE_Tomek are expressed in Table 10. Figure 14 is an intuitive display of Table 10. Figure 15 shows the ROC curve of the classification model on the 10-fold test set of the dataset processed by SMOTE_Tomek and the AUC value corresponding to each fold ROC curve.
The results of Table 10 and Table 6 show that the classification performance of the four models on the balanced dataset processed by SMOTE_Tomek method has been greatly improved in addition to the precision index. The best classification results are generated by lr model on the feature set combined by lightGBM, and the best accuracy, recall, F1 score, precision, specificity and AUC are 94.6%, 95.1%, 94.8%, 94.9%, 94.1% and 0.97, respectively. In particular, the highest recall in this section appears in the lightGBM + LR model, which is 95.4%. Similarly, our proposed feature combination method has significantly improved the classification ability of lr model. Figure 15 shows the ROC curve and corresponding AUC value of each fold in the 10-fold cross-validation obtained by classification model on the dataset balanced by SMOTE_Tomek. It can be seen from the curve that, due to the distribution difference of 10 fold data, the AUC values obtained by the model on each fold test set are different. Moreover, the average AUC value obtained by lr model on the dataset without feature combination processing is the lowest. However, on the dataset processed by ligthGBM model, the AUC result of lr model is improved.

5.2.6. Results Obtained on Dataset Processed by SMOTENC

The dataset processed by the SMOTENC method in this section contains 432 sample instances, of which 216 are classified as CAD class and 216 as normal class. The average results of 10 tests on accuracy, recall, precision, F1 score, specificity and AUC of the classification models on the dataset processed by SMOTENC are shown in Table 11. Figure 16 is a visual representation of Table 11. Figure 17 shows the ROC curve of the classification model on the 10-fold test set of the dataset processed by SMOTENC and the AUC value corresponding to each fold ROC curve.
Comparing Table 11 and Table 6, it can be seen that on the dataset processed by the SMOTENC method, the performance of the classification models in accuracy, recall, specificity and AUC has been improved. The feature combination function of lightGBM model can significantly improve the classification performance of lr model and produce the best classification results. Overall, the best classification results are generated by lightGBM + LR model. The best accuracy, recall, F1 score, precision, specificity and AUC are 93.1%, 93.3%, 93.4%, 94.4%, 92.9% and 0.97, respectively. However, it is worth noting that the highest accuracy and recall of this section appear in the lightGBM + lr model, with the highest accuracy of 93.3% and the highest recall of 94.6%. Figure 17 shows the ROC curve and corresponding AUC value of each fold in the 10-fold cross-validation obtained by classification model on the dataset balanced by SMOTENC. It can be seen from the curve that, due to the distribution difference of 10 fold data, the AUC values obtained by the model on each fold test set are different. In addition, the lr model obtained the highest average AUC value on the dataset without feature combination processing. Additionally, the AUC result obtained by lr model on the dataset processed by ligthGBM is reduced.

5.2.7. Results Obtained on Standardized Dataset

This section reports the performance results of the classification models used for CAD prediction in terms of accuracy, recall, precision, F1 score, specificity and AUC on the standardized dataset. Data standardization only eliminates the dimension of continuous features, and does not change the size of the dataset. Therefore, there is still a certain class imbalance in the standardized dataset, showing that 71.29% of the samples belong to the CAD class, and 28.71% of the samples belong to the normal class. The average results of 10 tests in terms of accuracy, recall, precision, F1 score, specificity and AUC obtained by the four classification models on the standardized dataset are listed in Table 12. The intuitive exhibition of Table 12 is drawn in Figure 18. Figure 19 shows the ROC curve of the classification model on the 10-fold test set of the standardized dataset and the AUC value corresponding to each fold ROC curve.
By analyzing Table 12 and Table 6, we can conclude that on the standardized dataset, the classification performance of lr model has been improved in general. Especially on the dataset processed by lightGBM, lr model obtains the best classification results. The best classification results are accuracy 91.4%, recall 93.1%, F1 score 94.2%, precision 95.8%, specificity 87.1% and AUC 0.93. However, when the lightGBM classifier is applied on a standardized dataset separately, the results are inferior to those obtained on the original dataset. Figure 19 shows the ROC curve and corresponding AUC value of each fold in the 10-fold cross-validation obtained by classification model on the dataset processed by data standardization. It can be seen from the curve that, due to the distribution difference of 10-fold data, the AUC values obtained by the model on each fold test set are different. Additionally, the lr model obtained the highest average AUC value on the dataset without feature combination processing.

5.2.8. Results Obtained on Dataset Processed by Standardization and SMOTE

The research in this section is developed based on the dataset processed by standardization and SMOTE algorithm. The dataset after standardization and SMOTE processing contains 216 samples of CAD class and 216 samples of normal class. The average results of 10 tests on accuracy, recall, precision, F1 score, specificity and AUC obtained by the classification models on this dataset are shown in Table 13. Figure 20 is an intuitive display of Table 13. Figure 21 shows the ROC curve of the classification model on the 10-fold test set of the dataset processed by standardization and SMOTE and the AUC value corresponding to each fold ROC curve.
It can be seen from Table 13 and Figure 20 that when compared with Table 12 and Figure 18, the performance of the classification models on the dataset processed by SMOTE has been significantly improved, mainly reflected in accuracy, recall, specificity and AUC indicators. The best classification performance is obtained by lr model on the dataset combined by lightGBM, with accuracy of 93.5%, recall of 94.6%, F1 score of 93.6%, precision of 93.0%, specificity of 92.4% and AUC of 0.97. Figure 21 shows the ROC curve and corresponding AUC value of each fold in the 10-fold cross-validation obtained by classification model on the dataset processed by data standardization and SMOTE. It can be seen from the curve that, due to the distribution difference of 10 fold data, the AUC values obtained by the model on each fold test set are different. Additionally, the proposed feature combination method has improved the AUC value of lr model on the dataset.

5.2.9. Results Obtained on Dataset Processed by Standardization and Borderline_SMOTE

In this section, the Borderline_SMOTE class balancing method is applied to the standardized dataset to obtain a new dataset. Four classifiers, namely lightGBM, lr, lightGBM + lr and lightGBM + LR, are applied on this dataset. The dataset processed by standardization and Borderline_SMOTE contains 432 samples, with CAD class and normal class accounting for 50 %, respectively. The average results of 10 tests on accuracy, recall, precision, F1 score, specificity and AUC obtained by the classification models on dataset used in this part are displayed in Table 14. Figure 22 is a visual representation of Table 14. Figure 23 shows the ROC curve of the classification model on the 10 fold test set of the dataset processed by standardization and Borderline_SMOTE and the AUC value corresponding to each fold ROC curve.
The comparison of Table 14 and Table 12 shows that Borderline_SMOTE significantly improves the ability of the classifiers. Specifically, in addition to precision, the performance of the four classifiers in other evaluation indicators is significantly improved. The best results consistently appear in the lightGBM + LR classification model. The best accuracy, recall, F1 score, precision, specificity and AUC are 94.7%, 96.1%, 94.7%, 93.5%, 93.2% and 0.97, respectively. From the results, it can be concluded that our proposed feature combination method has significantly improved the classification performance of lr model. Figure 23 shows the ROC curve and corresponding AUC value of each fold in the 10-fold cross-validation obtained by classification model on the dataset processed by data standardization and Borderline_SMOTE. It can be seen from the curve that, due to the distribution difference of 10 fold data, the AUC values obtained by the model on each fold test set are different. Additionally, the proposed feature combination method has improved the AUC value of lr model on this dataset.

5.2.10. Results Obtained on Dataset Processed by Standardization and SMOTE_SVM

This section shows the performance of classification models combined with data standardization and SMOTE_SVM processing methods. The number of sample instances in CAD class and normal class in the dataset processed by standardization and SMOTE_SVM method is the same, and both contain 216 cases. The average results of 10 tests in terms of accuracy, recall, precision, F1 score, specificity and AUC obtained by the classification models on this dataset are listed in Table 15. Figure 24 is a graphical representation of Table 15. Figure 25 shows the ROC curve of the classification model on the 10 fold test set of the dataset processed by standardization and SMOTE_SVM and the AUC value corresponding to each fold ROC curve.
As shown in Table 15 and Figure 24, when compared with Table 12 and Figure 18, it can be found that SMOTE_SVM processing significantly improves the ability of the classifiers. Specifically, in addition to precision, the performance of the four classifiers in other evaluation indexes is significantly improved. Overall, the best results are produced by the lightGBM + LR model. The best accuracy, recall, F1 score, precision, specificity and AUC are 94.5%, 95.1%, 94.5%, 94.4%, 93.9% and 0.96, respectively. However, on the dataset used in this section, the highest recall and AUC appear in lightGBM + lr model and the lightGBM model, respectively. The highest recall and AUC are 96.3% and 0.97. It can be seen that the feature combination of lightGBM model can significantly improve the classification ability of lr model. Figure 25 shows the ROC curve and corresponding AUC value of each fold in the 10-fold cross-validation obtained by classification model on the dataset processed by data standardization and SMOTE_SVM. It can be seen from the curve that, due to the distribution difference of 10 fold data, the AUC values obtained by the model on each fold test set are different. Additionally, the proposed feature combination method does not improve the AUC value of lr model on this dataset.

5.2.11. Results Obtained on Dataset Processed by Standardization and SMOTE_Tomek

In this section, we applied SMOTE_Tomek class balancing method to the standardized dataset, and obtained a balanced dataset. The balanced dataset contains 430 samples, of which 215 belong to CAD class and 215 belong to normal class. Table 16 shows the average results of 10 tests on accuracy, recall, precision, F1 score, specificity and AUC obtained by the classification models on this dataset. Figure 26 is an intuitive display of Table 16. Figure 27 shows the ROC curve of the classification model on the 10 fold test set of the dataset processed by standardization and SMOTE_Tomek and the AUC value corresponding to each fold ROC curve.
By comparing Table 16 and Table 12, we can infer that the classification performance of the four models has significantly improved on the dataset balanced by the SMOTE_Tomek class balancing method. On the dataset used in this section, four models show good classification ability. In particular, on the feature set combined by lightGBM, lr model achieved the best classification results in this study, with accuracy of 94.7%, recall of 94.8%, F1 score of 94.8%, precision of 95.3%, specificity of 94.5% and AUC of 0.98. This shows that the method of using the feature combination function of lightGBM model to improve the classification ability of lr model is effective. This also fully proves the effectiveness of our proposed method. Figure 27 shows the ROC curve and corresponding AUC value of each fold in the 10-fold cross-validation obtained by classification model on the dataset processed by data standardization and SMOTE_Tomek. It can be seen from the curve that, due to the distribution difference of 10 fold data, the AUC values obtained by the model on each fold test set are different. Additionally, the proposed feature combination method has significantly improved the AUC value of lr model on this dataset. Additionally, the highest AUC value of this study was produced.

5.2.12. Results Obtained on Dataset Processed by Standardization and SMOTENC

The performance results of the four classification models for CAD prediction in terms of accuracy, recall, precision, F1 score, specificity and AUC on the dataset processed by data standardization and SMOTENC method are reported in this section. The dataset used in this part consists of 432 samples, including 216 instances of CAD class and 216 instances of normal class. Table 17 exhibits the average results of 10 tests obtained by the four classification models on this dataset. Figure 28 is a histogram of Table 17. Figure 29 shows the ROC curve of the classification model on the 10 fold test set of the dataset processed by standardization and SMOTENC and the AUC value corresponding to each fold ROC curve.
It can be seen from Table 17 and Figure 28, compared with Table 12 and Figure 18, SMOTENC method improves the performance of the classification models in terms of accuracy, recall, specificity and AUC. On the dataset used in this section, the feature combination of lightGBM also has significantly improved the classification ability of lr model. The lightGBM + lr model obtained the best results with accuracy of 94.5%, recall of 94.2%, F1 score 94.7%, precision of 95.8%, specificity of 94.7% and AUC of 0.97. Figure 29 shows the ROC curve and corresponding AUC value of each fold in the 10-fold cross-validation obtained by classification model on the dataset processed by data standardization and SMOTENC. It can be seen from the curve that, due to the distribution difference of 10-fold data, the AUC values obtained by the model on each fold test set are different. Additionally, the proposed feature combination method has significantly improved the AUC value of lr model on this dataset.

5.3. Results Analysis

5.3.1. Analysis of the Influence of Class Balancing Methods on the Performance of Classification Models

This section reports the impact trend of five class balancing methods on the performance of classification models. In order to eliminate the impact of the data standardization processing method on the performance of the model, we analyze it based on the same basic dataset and the same feature processing. Therefore, study of this section is divided into two parts, one is based on the original dataset, and the other is based on the standardized dataset.
  • Research Based on Original Dataset
In this part, we describe in detail the change trend of the performance of the classification models after applying five class balancing methods on the original dataset, as shown in Figure 30a.
The three child figures from top to bottom in Figure 30a are the trend charts of the performance evaluation indexes of the three classification models with the five class balancing methods, respectively. The three classification models are lr, lightGBM + lr and lightGBM + LR. The mark ‘None’ in Figure 30 refers to the dataset that does not apply the class balancing methods. The marks ‘SMOTE’, ‘BorderLine_SMOTE’, ‘SMOTE_SVM’, ‘SMOTE_Tomek’ and ‘SMOTENC’ correspond to the datasets that applied the corresponding class balancing methods, respectively.
By analyzing the results of three classification models on the ‘None’ dataset, it can be found that the distribution of results is scattered, that is, the model obtains good results on some indicators, but it is poor on other indicators. This may be related to the skewness and class imbalance of the original dataset. After applying the class balancing methods, the results of the model are improved, and the distribution of the results tends to be concentrated. For lr and lightGBM + lr classification models, the most concentrated and best results appear on the dataset balanced by SMOTE_Tomek method. For the lightGBM + LR model, the most concentrated results appear on the dataset balanced by the BorderLine_SMOTE method, and the best results also appear on the dataset balanced by the SMOTE_Tomek method. The reason for this phenomenon in the lightGBM + LR model may be related to the addition of the original dataset to the recombined dataset. The change trend of the results of the three classification models in Figure 30a confirms that the five class balancing methods can effectively improve the classification performance and stability of the model on the original dataset. Especially, SMOTE_Tomek algorithm is a more effective method.
2.
Research Based on Standardized Dataset
The change trend of the performance of the three classification models after applying five class balancing methods on the standardized dataset are described in Figure 30b.
By analyzing Figure 30b, we can find that after applying the class balancing methods, the results of the models are improved, and the distribution of the results tends to be concentrated. For lightGBM + lr model, the most concentrated and best results appear on the dataset balanced by SMOTE_Tomek method. For the lr model, the most concentrated results appear on the dataset balanced by the SMOTE_Tomek method, and the best results appear on the dataset balanced by the SMOTE_SVM method. For the lightGBM + LR model, the most concentrated results appear on the dataset balanced by the SMOTE_SVM method, and the best results appear on the dataset balanced by the BorderLine_SMOTE method. The variation trend of the results of the three classification models in Figure 30b confirms that on the standardized dataset, the five class balancing methods can effectively improve the classification performance and stability of the model.

5.3.2. Analysis of the Influence of Data Standardization Method on the Performance of Classification Models

In this part, we analyze the impact trend of data standardization method on the performance of classification models. According to the foregoing, the best classification results on each dataset often come from the lightGBM + lr model or the lightGBM + LR model. Therefore, this part mainly analyzes the influence of data standardization method on lightGBM + lr model and lightGBM + LR model. As shown in Figure 31 and Figure 32, respectively. The study of this part adopts the matching method, that is, the application of other processing methods is the same for the two sets of datasets used for comparison, except whether data standardization processing method is applied or not. Therefore, each model has six groups of comparative datasets, which are the datasets without balancing and the datasets processed by SMOTE, BorderLine_SMOTE, SMOTE_SVM, SMOTE_Tomek and SMOTENC, respectively. Figure 31 shows the performance comparison of lightGBM + lr model on six groups of datasets. Figure 32 shows the performance comparison of lightGBM + LR model on six groups of datasets. The (a) to (f) of Figure 31 and Figure 32 correspond to the datasets without balancing and the datasets processed by SMOTE, BorderLine_SMOTE, SMOTE_SVM, SMOTE_Tomek and SMOTENC, respectively.
From the 6 groups comparison diagram of (a)–(f) in Figure 31, it can be find that the standardization of datasets can significantly improve the classification ability of lightGBM + lr model.
From the 6 groups comparison diagram of (a)–(f) in Figure 32, it can be find that, on the whole, the data standardization method can significantly improve the classification ability of lightGBM + LR model. Although the data standardization method does not improve the performance of lightGBM + LR model on the dataset processed by SMOTE method, the recognition ability of lightGBM + LR model to positive samples, namely the recall index, has been improved on the standardized dataset, which is also considered to be very important.

5.3.3. Analysis of Ablation Study

The composition of the CAD prediction model studied in this paper can be divided into four modules, namely data preprocessing, class balance processing, feature combination and class prediction. The modules of data preprocessing, class balance processing and feature combination are designed to optimize the performance of classification algorithm. To understand which modules are critical to detection performance, we analyzed the ablation study of the model. That is, after removing one module in turn, the change of the model’s detection performance will be observed. This can intuitively reflect the impact of the module on the classification performance of model. In this study, five different class balancing algorithms are tried. The experimental results show that the best classification results are obtained by lr model on the dataset processed by SMOTE_Tomek method and lightGBM algorithm. Therefore, the class balance processing module studied in this part applies SMOTE_Tomek method, and the feature combination module does not involve the recombination of the original dataset. Table 18, Table 19 and Table 20 show the classification results of the prediction model after removing the modules of data preprocessing, class balance processing and feature combination in turn. Figure 33 compares the improvement effects of different modules on the performance results of model.
By analyzing Table 18, Table 19 and Table 20, it can be seen that the feature combination module can significantly improve the results of all performance indicators. In addition to the precision, the class balance processing module also shows a significant improvement in other performance indexes. Additionally, the data preprocessing module also shows an improvement in other performance indexes, in addition to the recall and F1 score. Figure 33 intuitively shows the comparison of the three processing modules on the improvement of model prediction performance. It can be seen from Figure 33 that the three processing modules are all effective methods to improve the classification ability of the model, but they have different effects on the performance improvement of the classification model. Among them, the feature combination has the strongest effect on the improvement of model performance. The second is class balance processing. The data preprocessing module has the weakest effect on the improvement of model performance. Specifically, the class balance processing has the strongest ability to improve recall, specificity and AUC. Feature combination processing has the strongest ability to improve accuracy, F1 score and precision. Therefore, the combination of class balance and feature combination is an effective method to realize the early and accurate diagnosis of CAD.

5.3.4. Analysis of Model Loss

In this section, we analyze the loss of the proposed model on the training set and test set. As mentioned earlier, lr model obtains the best classification performance in this study on the dataset processed by data preprocessing, SMOTE_Tomek method and ligthGBM algorithm. This model is the complete embodiment of our proposed method. This section is analyzed and discussed based on this model, namely ligthGBM + lr classifier. At the same time, we also calculate the loss of lr model on the dataset without feature combination processing, that is, lr classifier. In the development process of the model, we applied the 10-fold cross-validation technology. Therefore, the loss in each epoch is the average of the 10-fold data. Figure 34 and Figure 35, respectively, show the change trend of the loss of the two models on the training set and the test set with the update of the number of epochs. It can be seen from Figure 34 and Figure 35 that, on the training set, with the increase of epoch times, the loss of both models gradually decreases and finally approaches a stable value. On the test set, with the update of the number of epochs, the loss of lr model gradually decreases and finally tends to be stable. The loss of lightGBM + lr model shows a trend of decreasing first, then increasing finally stabilizing. The lowest loss occurs when epoch = 4. These indicate that the model shows a certain over fitting when epochs increases to 5. Therefore, we take epochs = 4 to predict the test set.

6. Discussion

In this study, we proposed a model for early, accurate and rapid detection of CAD. The model is based on the classical logistic regression algorithm. Logistic regression algorithm with good interpretability is the most commonly used research method in medical problems. Especially in medical binary classification problems, logistic regression algorithm has irreplaceable advantages. However, the disadvantage of logistic regression algorithm is also obvious, that is, it needs a lot of manual feature engineering. There is no doubt that this will take a lot of time. Tree model algorithm has been widely used in feature selection and feature combination, because of its efficient node splitting mechanism. Therefore, in this study, we used the tree model algorithm to realize the feature engineering automation of lr algorithm. At the same time, in order to improve the classification performance of the proposed model, we applied five sampling methods to solve the class imbalance problem in the dataset. The corresponding solutions to the problem of feature distribution skewness in the dataset were proposed. In addition, the 10 fold cross validation technique was used to test the robustness of the model. Seven important evaluation indexes, namely, accuracy, recall, F1 score, precision, specificity, ROC curve and AUC, were used for the evaluation and analysis of the model. The experimental results shown that the data preprocessing, class balance processing and feature combination methods can significantly improve the classification ability of lr model. However, different processing methods have different effects on the classification performance of the model. In general, the feature combination processing based on lighthGBM algorithm has the strongest effect on improving the performance of the model, followed by class balance processing. The data standardization method has the weakest effect on the improvement of model performance. When it comes to the performance evaluation metrics of the model, we can find that class balance processing has the strongest ability to improve recall, specificity and AUC. Feature combination processing based on lighthGBM algorithm has the strongest ability to improve accuracy, F1 score and precision. Besides, the data standardization method does not contribute to the recall rate and F1 score index of the model, but also improved the performance of the model in other indicators. Therefore, the combination of data preprocessing method, class balance method and feature combination method can improve the performance of the classification model in all evaluation metrics. Our experimental results also confirm this. The best classification result of this study is generated by lr model on the dataset after standardization, resampling and feature combination. The best results we obtained were accuracy of 94.7%, recall of 94.8%, F1 score of 94.8%, precision of 95.3%, specificity of 94.5% and AUC of 0.98.
In order to more intuitively show whether our proposed method has the ability to improve lr model and whether the results of the proposed model can compete with the ensemble learning algorithm, which has obvious advantages in classification task. During the experiment, we applied lr model without feature combination effect on the same dataset. Additionally, the prediction results of lighthGBM model on each dataset were output. By comparing the results on each dataset, it can be seen that the best classification results on each dataset are almost produced by the lr model with the feature combination effect of lighthGBM. This showed that our proposed method has the ability to improve the classification performance of lr model. Additionally, our proposed method has a competitive advantage.
In addition, we compared the results obtained by our proposed method with those obtained on Z-AlizadehSani dataset reported in previous literature (See Table 21). As shown in Table 21, our proposed method is very competitive. On the Z-Alizadeh Sani dataset, our proposed method achieves better results than existing research. It is worth noting that there are many cell values labeled as ‘N’ in Table 21, which means that the corresponding indicators are not reported in the literature. However, these indicators are very valuable for evaluating the performance and stability of medical models, especially recall, F1 score and AUC indicators.
To sum up the above, the main contributions of this paper in technology and experimental design are as follows: (1) in the prediction task of CAD, we applied the automatic feature combination function based on ligthGBM model to improve the performance of lr classifier; (2) five sampling methods were applied to Z-Alizadeh Sani dataset to solve the class imbalance problem of dataset; and (3) recall, F1 score, ROC curve and AUC of metrics, which are critical to the evaluation of medical models, were used to evaluate the models. At the same time, our article also has some limitations, such as: (1) more feature combination methods were not tried; and (2) more classification algorithms were not tried. These are also our future research directions.

7. Conclusions

CAD is the major cause of global health burden and death. The early, accurate and rapid diagnosis of CAD is the only effective way to reduce the damage of CAD. In this paper, we proposed a model that can accurately detect CAD according to the results of clinical routine examination in the early stage. The model took the logistic regression algorithm as the base classifier, and combined the feature combination processing based on lighthGBM algorithm and the class balance method of resampling. In addition, according to the characteristics of the dataset, the corresponding data preprocessing method was applied. A 10-fold cross validation technique was used to test the robustness of the model. Accuracy, recall, specificity, precision, F1 score, ROC curve and AUC indicators were used to evaluate the model. The experimental results and analysis show that our proposed model has strong advantages in the models used for CAD detection. Our proposed model can be used to assist doctors in making decisions.

Author Contributions

Conceptualization, Y.Y., J.T. and X.W.; methodology, S.Z., Y.Y., J.Y. and Z.Y.; software, S.Z.; validation, S.Z.; formal analysis, S.Z., J.T., X.W., J.Y. and Z.Y.; writing—original draft preparation, S.Z.; writing—review and editing, S.Z. and Y.Y.; supervision, Y.Y., J.T. and X.W.; funding acquisition, J.T. and X.W. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by two funds: Basic Research of the Ministry of Science and Technology, China (2013FY114000), and National Key Basic Research Development Program, China (2015CB351900).

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Table A1. A list of abbreviations used in this paper.
Table A1. A list of abbreviations used in this paper.
AbbFull NameAbbFull Name
CADcoronary artery diseaseADASYNadaptive synthetic
DALYsdisability-adjusted life yearsSVMsupport vector machines
CVDcardiovascular diseaseKNNk-nearest neighbors
IHDischemic heart diseasePSOparticle swarm optimization
HIVhuman immunodeficiency virusPCAprincipal component analysis
AIDSacquired immunodeficiency syndromeRFrandom forest
LADleft anterior descending coronary arteryETextreme tree
LCXleft circumflex coronary arteryLDAlinear discriminant analysis
RCAright coronary arteryGRgain ratio
ECGelectrocardiogramIGinformation gain
FFRfractional flow reserveFLDAfisher linear discriminant analysis
IVUSintravenous ultrasoundCSchi-square
COPDchronic obstructive pulmonary diseaseCARTclassification and regression tree
SMOsequence minimum optimizationROCreceiver operating characteristic
SMOTEsynthetic minority oversampling technologyAUCarea under curve
Abb, Abbreviations.
Table A2. The hyperparameters setting of the proposed model on the dataset processed by data standardization and SMOTE_Tomek method.
Table A2. The hyperparameters setting of the proposed model on the dataset processed by data standardization and SMOTE_Tomek method.
ModelAlgorithmsHyperparametersValue of Hyperparameters
lightGBM + lrlightGBMboosting_type‘gbdt’
objective‘binary’
learning_rate0.52488778
n_estimators231
max_depth2
num_leaves3
max_bin74
min_data_in_leaf21
bagging_fraction0.676450
bagging_freq2
feature_fraction0.610
lrC1.79532
class_weight{0:0.5,1:0.5}
dualFalse
fit_interceptTrue
intercept_scaling1
max_iter4
multi_class‘ovr’
n_jobs1
penalty‘l1’
random_state69
solver“liblinear”
tol0.01
verbose0
warm_startFalse

References

  1. World Health Organization. World Health Statistics 2019: Monitoring Health for the SDGs, Sustainable Development Goals; World Health Organization: Geneva, Switzerland, 2019. [Google Scholar]
  2. Mensah, G.A.; Roth, G.A.; Fuster, V. The Global Burden of Cardiovascular Diseases and Risk Factors 2020 and Beyond. JACC 2019, 74, 2529–2532. [Google Scholar] [CrossRef] [PubMed]
  3. Roth, G.A.; Mensah, G.A.; Fuster, V. The Global Burden of Cardiovascular Diseases and Risks: A Compass for Global Action. J. Am. Coll. Cardiol. 2020, 76, 2980–2981. [Google Scholar] [CrossRef] [PubMed]
  4. GBD 2019 Risk Factors Collaborators. Global Burden of 87 Risk Factors in 204 Countries and Territories, 1990–2019: A Systematic Analysis for the Global Burden of Disease Study 2019. Lancet 2020, 396, 1223–1249. [Google Scholar] [CrossRef]
  5. Roth, G.A.; Mensah, G.A.; Johnson, C.O.; Addolorato, G.; Ammirati, E.; Baddour, L.M.; Barengo, N.C.; Beaton, A.Z.; Benjamin, E.J.; Benziger, C.P.; et al. Global Burden of Cardiovascular Diseases and Risk Factors, 1990–2019: Update From the GBD 2019 Study. J. Am. Coll. Cardiol. 2020, 76, 2982–3021. [Google Scholar] [CrossRef] [PubMed]
  6. GBD. Diseases and Injuries Collaborators. Global Burden of 369 Diseases and Injuries in 204 Countries and Territories, 1990–2019: A Systematic Analysis for the Global Burden of Disease Study 2019. Lancet 2020, 396, 1204–1222. [Google Scholar] [CrossRef]
  7. Zipes, D.P.; Libby, P.; Bonow, R.O. Braunwald’s Heart Disease E-Book: A Textbook of Cardiovascular Medicine; Elsevier Health Sciences: Amsterdam, The Netherlands, 2018. [Google Scholar]
  8. Jayaraman, V.; Sultana, H.P. Artifificial Gravitational Cuckoo Search Algorithm along with Particle Bee Optimized Associative Memory Neural Network for Feature Selection in Heart Disease Classification. J. Ambient Intell. Humaniz. Comput. 2019, 1–10. [Google Scholar] [CrossRef]
  9. Liu, M.; Kim, Y. Classification of Heart Diseases Based on ECG Signals Using Long Short-Term Memory. In Proceedings of the 40th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Honolulu, HI, USA, 18–21 July 2018; pp. 2707–2710. [Google Scholar]
  10. Vijayashree, J.; Sultana, H.P. Heart Disease Classification Using Hybridized Ruzzo-Tompa Memetic Based Deep Trained Neocognitron Neural Network. Health Technol. 2019, 10, 207–216. [Google Scholar] [CrossRef]
  11. Acharya, U.R.; Fujita, H.; Oh, S.L.; Hagiwara, Y.; Tan, J.H.; Adam, M. Application of Deep Convolutional Neural Network for Automated Detection of Myocardial Infarction Using ECG Signals. Inf. Sci. 2017, 415, 190–198. [Google Scholar] [CrossRef]
  12. Gupta, V.; Mittal, M. Arrhythmia Detection in ECG Signal Using Fractional Wavelet Transform with Principal Component Analysis. J. Inst. Eng. India Ser. B 2020, 101, 451–461. [Google Scholar] [CrossRef]
  13. Liu, W.; Huang, Q.; Chang, S.; Wang, H.; He, J. Multiple-Feature-Branch Convolutional Neural Network for Myocardial Infarction Diagnosis Using Electrocardiogram. Biomed. Signal Process. Control 2018, 45, 22–32. [Google Scholar] [CrossRef]
  14. Tan, J.H.; Hagiwara, Y.; Pang, W.; Lim, I.; Oh, S.L.; Adam, M.; Tan, R.S.; Chen, M.; Acharya, U.R. Application of Stacked Convolutional and Long Short-Term Memory Network for Accurate Identification of CAD ECG Signals. Comput. Biol. Med. 2018, 94, 19–26. [Google Scholar] [CrossRef] [PubMed]
  15. Acharya, U.R.; Fujita, H.; Oh, S.L.; Adam, M.; Tan, J.H.; Chua, K.C. Automated Detection of Coronary Artery Disease Using Different Durations of ECG Segments with Convolutional Neural Network. Knowl. Base Syst. 2017, 132, 62–71. [Google Scholar] [CrossRef]
  16. Zihlmann, M.; Perekrestenko, D.; Tschannen, M. Convolutional Recurrent Neural Networks for Electrocardiogram Classification. arXiv 2017, arXiv:1710.06122v1. [Google Scholar]
  17. Gupta, V.; Mittal, M.; Mittal, V. A Novel FrWT Based Arrhythmia Detection in ECG Signal Using YWARA and PCA. Wirel. Pers. Commun. 2021, 1, 1–18. [Google Scholar] [CrossRef]
  18. Gupta, V.; Mittal, M.; Mittal, V.; Gupta, A. An Efficient AR Modelling-Based Electrocardiogram Signal Analysis for Health Informatics. Int. J. Med. Eng. Inform. 2022, 14, 74–89. [Google Scholar]
  19. Patidar, S.; Pachori, R.B.; Acharya, U.R. Automated Diagnosis of Coronary Artery Disease Using Tunable-Q Wavelet Transform Applied on Heart Rate Signals. Knowl.-Based Syst. 2015, 82, 1–10. [Google Scholar] [CrossRef]
  20. Sridhar, C.; Acharya, U.R.; Fujita, H.; Bairy, G.M. Automated Diagnosis of Coronary Artery Disease using Nonlinear Features Extracted from ECG Signals. In Proceedings of the 2016 IEEE International Conference on Systems, Man, and Cybernetics (SMC), Budapest, Hungary, 9–12 October 2016; pp. 545–549. [Google Scholar] [CrossRef]
  21. Altan, G.; Allahverdi, N.; Kutlu, Y. Diagnosis of Coronary Artery Disease Using Deep Belief Networks. Eur. J. Eng. Nat. Sci. 2017, 2, 29–36. [Google Scholar]
  22. Sharma, M.; Acharya, U.R. A New Method to Identify Coronary Artery Disease with ECG Signals and Time-Frequency Concentrated Antisymmetric Biorthogonal Wavelet Filter Bank. Pattern Recognit. Lett. 2019, 125, 235–240. [Google Scholar] [CrossRef]
  23. Zreik, M.; van Hamersvelt, R.W.; Wolterink, J.M.; Leiner, T.; Viergever, M.A.; Išgum, I. A Recurrent CNN for Automatic Detection and Classification of Coronary Artery Plaque and Stenosis in Coronary CT Angiography. IEEE Trans. Med. Imaging 2018, 38, 1588–1598. [Google Scholar] [CrossRef] [Green Version]
  24. Sprem, J.; de Vos, B.D.; de Jong, P.A.; Viergever, M.A.; Išgum, I. Classification of Coronary Artery Calcifications According to Motion Artifacts in Chest CT Using a Convolutional Neural Network. In Proceedings of the Medical Imaging 2017: Image Processing, Orlando, FL, USA, 11–16 February 2017; International Society for Optics and Photonics: Orlando, FL, USA, 2017; Volume 10133. [Google Scholar]
  25. Kim, G.Y.; Lee, J.H.; Hwang, Y.N.; Kim, S.M. A Novel Intensity-Based Multi-Level Classification Approach for Coronary Plaque Characteriza-Tion in Intravascular Ultrasound Images. Biomed. Eng. Online 2018, 17, 200–213. [Google Scholar] [CrossRef] [Green Version]
  26. Gessert, N.; Lutz, M.; Heyder, M.; Latus, S.; Leistner, D.M.; Abdelwahed, Y.S.; Schlaefer, A. Automatic Plaque Detection in IVOCT Pullbacks Using Convolutional Neural Networks. IEEE Trans. Med. Imaging 2018, 38, 426–434. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  27. Yang, J.L.; Zhang, B.; Wang, H.R.; Lin, F.; Han, Y.C.; Lin, X.L. Automated Characterization and Classification of Coronary Atherosclerotic Plaques for Intravascular Optical Coherence Tomography. Biocybern. Biomed. Eng. 2019, 39, 719–727. [Google Scholar] [CrossRef]
  28. Martin-Isla, C.; Campello, V.M.; Izquierdo, C.; Raisi-Estabragh, Z.; Baessler, B.; Petersen, S.E.; Lekadir, K. Image-Based Cardiac Diagnosis With Machine Learning: A Review. Front. Cardiovasc. Med. 2020, 7, 1. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  29. Chen, C.; Qin, C.; Qiu, H.Q.; Tarroni, G.; Duan, J.M.; Bai, W.J.; Rueckert, D. Deep Learning for Cardiac Image Segmentation: A Review. Front. Cardiovasc. Med. 2020, 7, 25. [Google Scholar] [CrossRef] [PubMed]
  30. Alizadehsani, R.; Habibi, J.; Hosseini, M.J.; Mashayekhi, H.; Boghrati, R.; Ghandeharioun, A.; Bahadorian, B.; Sani, Z.A. A Data Mining Approach for Diagnosis of Coronary Artery Disease. Comput. Methods Programs Biomed. 2013, 111, 52–61. [Google Scholar] [CrossRef] [PubMed]
  31. Nasarian, E.; Abdar, M.; Fahami, M.A.; Alizadehsani, R.; Hussaind, S.; Basiri, M.E.; Zomorodi-Moghadam, M.; Zhou, X.J.; Pławiak, P.; Acharya, U.R.; et al. Association between Work-Related Features and Coronary Artery Disease: A Heterogeneous Hybrid Feature Selection Integrated with Balancing Approach. Pattern Recognit. Lett. 2020, 133, 33–40. [Google Scholar] [CrossRef]
  32. Arabasadi, Z.; Alizadehsani, R.; Roshanzamir, M.; Moosaei, H.; Yarifard, A.A. Computer Aided Decision Making for Heart Disease Detection Using Hybrid Neural Net-Work-Genetic Algorithm. Comput. Methods Programs Biomed. 2017, 141, 19–26. [Google Scholar] [CrossRef]
  33. Alizadehsani, R.; Hosseini, M.J.; Sani, Z.A.; Ghandeharioun, A.; Boghrati, R. Diagnosis of Coronary Artery Disease Using Cost-Sensitive Algorithms. In Proceedings of the IEEE 12th International Conference on Data Mining Workshops, Brussels, Belgium, 10 December 2012; pp. 9–16. [Google Scholar]
  34. Zomorodi-moghadam, M.; Abdar, M.; Davarzani, Z.; Zhou, X.J.; Pławiak, P.; Acharya, U.R. Hybrid Particle Swarm Optimization for Rule Discovery in the Diagnosis of Coronary Artery Disease. Expert Syst. 2019, 38, e12485. [Google Scholar] [CrossRef]
  35. Abdar, M.; Acharya, U.R.; Sarrafzadegan, N.; Makarenkov, V. Ne-nu-svc: A New Nested Ensemble Clinical Decision Support System for Effective Diagnosis of Coronary Artery Disease. IEEE Access 2019, 7, 167605–167620. [Google Scholar] [CrossRef]
  36. Abdar, M.; Książek, W.; Acharya, U.R.; Tan, R.S.; Makarenkov, V.; Pławiak, P. A New Machine Learning Technique for an Accurate Diagnosis of Coronary Artery Disease. Comput. Methods Programs Biomed. 2019, 179, 104992. [Google Scholar] [CrossRef]
  37. Alizadehsani, R.; Hosseini, M.J.; Khosravi, A.; Khozeimeh, F.; Roshanzamir, M.; Sarrafzadegan, N.; Nahavandi, S. Non-invasive Detection of Coronary Artery Disease in High-Risk Patients Based on the Stenosis Prediction of Separate Coronary Arteries. Comput. Methods Programs Biomed. 2018, 162, 119–127. [Google Scholar] [CrossRef] [PubMed]
  38. Shahid, A.H.; Singh, M.P. A Novel Approach for Coronary Artery Disease Diagnosis using Hybrid Particle Swarm Optimization based Emotional Neural Network. Biocybern. Biomed. Eng. 2020, 40, 1568–1585. [Google Scholar] [CrossRef]
  39. Wang, J.K.; Liu, C.C.; Li, L.P.; Li, W.; Yao, L.K.; Li, H.; Zhang, H. A Stacking-Based Model for Non-Invasive Detection of Coronary Heart Disease. IEEE Access 2020, 8, 37124–37133. [Google Scholar] [CrossRef]
  40. Tama, B.A.; Im, S.; Lee, S. Improving an Intelligent Detection System for Coronary Heart Disease Using a Two-Tier Classifier Ensemble. BioMed Res. Int. 2020, 2020, 9816142. [Google Scholar] [CrossRef] [PubMed]
  41. Gupta, A.; Kumar, R.; Arora, H.S.; Raman, B. C-CADZ: Computational Intelligence System for Coronary Artery Disease Detection Using Z-Alizadeh Sani Dataset. Appl. Intell. 2021, 52, 2436–2464. [Google Scholar] [CrossRef]
  42. Kolukisa, B.; Hacilar, H.; Goy, G.; Kus, M.; Bakir-Gungor, B.; Aral, A.; Gungor, V.C. Evaluation of Classification Algorithms, Linear Discriminant Analysis and a New Hybrid Feature Selection Methodology for the Diagnosis of Coronary Artery Disease. In Proceedings of the 2018 IEEE International Conference on Big Data (Big Data), Seattle, WA, USA, 10–13 December 2018; pp. 2232–2238. [Google Scholar]
  43. Dekamin, A.; Sheibatolhamdi, A. A Data Mining Approach for Coronary Artery Disease Prediction in Iran. J. Adv. Med. Sci. Appl. Technol. 2017, 3, 29–38. [Google Scholar] [CrossRef] [Green Version]
  44. Alizadehsani, R.; Habibi, J.; Alizadehsani, Z.; Mashayekhi, H.; Boghrati, R.; Ghandeharioun, A.; Bahadorian, B. Diagnosis of Coronary Artery Disease Using Data Mining Based on Lab Data and Echo Features. J. Med. Bioeng. 2012, 1, 26–29. [Google Scholar] [CrossRef]
  45. Yadav, C.; Lade, S.; Suman, M.K. Predictive Analysis for the Diagnosis of Coronary Artery Disease using Association Rule Mining. Int. J. Comput. Appl. 2014, 87, 9–13. [Google Scholar] [CrossRef]
  46. Ghiasi, M.M.; Zendehboudi, S.; Mohsenipour, A.A. Decision Tree-Based Diagnosis of Coronary Artery Disease: CART Model. Comput. Methods Programs Biomed. 2020, 192, 105400. [Google Scholar] [CrossRef]
  47. Joloudari, J.H.; Azizi, F.; Nematollahi, M.A.; Alizadehsani, R.; Hassannatajjeloudari, E.; Nodehi, I.; Mosavi, A. GSVMA: A Genetic Support Vector Machine ANOVA Method for CAD Diagnosis. Front. Cardiovasc. Med. 2022, 8, 760178. [Google Scholar] [CrossRef]
  48. Zhang, S.; Yuan, Y.; Yao, Z.; Wang, X.; Lei, Z. Improvement of the Performance of Models for Predicting Coronary Artery Disease Based on XGBoost Algorithm and Feature Processing Technology. Electronics 2022, 11, 315. [Google Scholar] [CrossRef]
  49. Alizadehsani, R.; Abdar, M.; Roshanzamir, M.; Khosravi, A.; Kebria, P.M.; Khozeimeh, F.; Nahavandi, S.; Sarrafzadegan, N.; Acharya, U.R. Machine Learning-Based Coronary Artery Disease Diagnosis: A Comprehensive Review. Comput. Biol. Med. 2019, 111, 103346. [Google Scholar] [CrossRef] [PubMed]
  50. Liu, T.; Moore, A.W.; Gray, A.; Yang, K. An Investigation of Practical Approximate Nearest Neighbor Algorithms. In Proceedings of the 17th International Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 1 December 2004; pp. 825–832. [Google Scholar]
  51. Chawla, N.V.; Bowyer, K.W.; Hall, L.O.; Kegelmeyer, W.P. Smote: Synthetic Minority Over-Sampling Technique. J. Artif. Intell. Res. 2002, 16, 321–357. [Google Scholar] [CrossRef]
  52. Xu, X.L.; Chen, W.; Sun, Y.F. Over-Sampling Algorithm for Imbalanced Data Classification. J. Syst. Eng. Electron. 2019, 30, 1182–1191. [Google Scholar] [CrossRef]
  53. Lee, H.S.; Jung, S.; Kim, M.; Kim, S. Synthetic Minority Over-Sampling Technique based on Fuzzy C-means Clustering for Imbalanced Data. In Proceedings of the 2017 International Conference on Fuzzy Theory and Its Applications (iFUZZY), Taiwan, China, 12–15 November 2017. [Google Scholar]
  54. Gosain, A.; Sardana, S. Handling Class Imbalance Problem using Oversampling Techniques: A Review. In Proceedings of the 2017 International Conference on Advances in Computing, Communications and Informatics (ICACCI), Udupi, India, 13–16 September 2017; pp. 79–85. [Google Scholar]
  55. Han, H.; Wang, W.Y.; Mao, B.H. Borderline-SMOTE: A New Over-Sampling Method in Imbalanced Data Sets Learning. Lect. Notes Artif. Intell. 2005, 3644, 878–887. [Google Scholar]
  56. Nguyen, H.M.; Cooper, E.W.; Kamei, K. Borderline Over-sampling for Imbalanced Data Classification. In Proceedings of the Fifth International Workshop on Computational Intelligence & Applications, Hiroshima University, Hiroshima, Japan, 10–12 November 2009. [Google Scholar]
  57. Batista, G.E.A.P.A.; Prati, R.C.; Monard, M.C. A Study of the Behavior of Several Methods for Balancing Machine Learning Training Data. ACM SIGKDD Explor. Newsl. 2004, 6, 20–29. [Google Scholar] [CrossRef]
  58. Wang, M.Y.; Wei, Z.H.; Jia, M.; Chen, L.Z.; Ji, H. Deep Learning Model for Multi-Classification of Infectious Diseases from Unstructured Electronic Medical Records. BMC Med. Inform. Decis. Mak. 2022, 22, 41. [Google Scholar] [CrossRef] [PubMed]
  59. Santos, L.I.; Camargos, M.O.; Silveira, M.F.; D’Angelo, M.F.S.V.; Mendes, J.B.; Medeiros, E.E.C.d.; Guimarães, A.L.S.; Palhares, R.M. Decision Tree and Artificial Immune Systems for Stroke Prediction in Imbalanced Data. Expert Syst. Appl. 2022, 191, 116221. [Google Scholar] [CrossRef]
  60. Sapra, V.; Saini, M.L.; Verma, L. Identification of Coronary Artery Disease using Artificial Neural Network and Case-Based Reasoning. Recent Adv. Comput. Sci. Commun. 2021, 14, 2651–2661. [Google Scholar] [CrossRef]
  61. Chen, X.; Fu, Y.; Lin, J.; Ji, Y.; Fang, Y.; Wu, J. Coronary Artery Disease Detection by Machine Learning with Coronary Bifurcation Features. Appl. Sci. 2020, 10, 7656. [Google Scholar] [CrossRef]
  62. Guo, H.; Tang, R.; Ye, Y.; Li, Z.; He, X. DeepFM: A Factorization-Machine based Neural Network for CTR Prediction. In Proceedings of the Twenty-Sixth International Joint Conference on Artificial Intelligence (IJCAI-17), Melbourne, Australia, 19–25 August 2017; pp. 1725–1731. [Google Scholar]
  63. Ke, G.L.; Meng, Q.; Finley, T.; Wang, T.F.; Chen, W.; Ma, W.D.; Ye, Q.W.; Liu, T.Y. LightGBM: A Highly Efficient Gradient Boosting Decision Tree. Adv. Neural Inf. Processing Syst. 2017, 30, 1–9. [Google Scholar]
  64. Friedman, J.H. Greedy Function Approximation: A Gradient Boosting Machinegreedy. Ann. Stat. 2001, 29, 1189–1232. [Google Scholar] [CrossRef]
  65. Tian, Z.; Chen, C.Y.; Fan, Y.M.; Ou, X.J.; Wang, J.; Ma, X.L.; Xu, J.G. Glioblastoma and Anaplastic Astrocytoma: Differentiation Using MRI Texture Analysis. Front. Oncol. 2019, 9, 876. [Google Scholar] [CrossRef] [PubMed]
  66. Qing, Y.; Zhi, J.C.; Ying, L.T. Prediction of Aptamer–Protein Interacting Pairs Based on Sparse Autoencoder Feature Extrac-Tion and an Ensemble Classifier. Math. Biosci. 2019, 311, 103–108. [Google Scholar]
  67. Tian, X.; Wang, J.; Wen, Y.; Ma, H. Multi-Attribute Scientific Documents Retrieval and Ranking Model Based on GBDT and LR. Math. Biosci. Eng. 2022, 19, 3748–3766. [Google Scholar] [CrossRef] [PubMed]
  68. Zabor, E.C.; Reddy, C.A.; Tendulkar, R.D.; Patil, S. Logistic Regression in Clinical Studies. Int. J. Radiat. Oncol. Biol. Phys. 2021, 112, 271–277. [Google Scholar] [CrossRef] [PubMed]
  69. Li, H.; Wang, X.P.; Li, Y.; Qin, C.J.; Liu, C.C. Comparison between Medical Knowledge Based and Computer Automated Feature Selection for Detection of Coronary Artery Disease Using Imbalanced Data. In Proceedings of the BIBE 2018, International Conference on Biological Information and Biomedical Engineering, Shanghai, China, 6–8 June 2018; pp. 1–4. [Google Scholar]
  70. Cüvitoğlu, A.; Işik, Z. Classification of Cad Dataset by Using Principal Component Analysis and Machine Learning Approaches. In Proceedings of the 5th International Conference on Electrical and Electronic Engineering (ICEEE), Istanbul, Turkey, 3–5 May 2018; pp. 340–343. [Google Scholar]
Figure 1. The frame diagram of our proposed machine learning model.
Figure 1. The frame diagram of our proposed machine learning model.
Electronics 11 01495 g001
Figure 2. An example of an ECG waveform and an echocardiac image (pictures from the Internet).
Figure 2. An example of an ECG waveform and an echocardiac image (pictures from the Internet).
Electronics 11 01495 g002
Figure 3. Standard for Borderline_SMOTE algorithm to assign minority samples to SAFE set, DANGER set and NOISE set.
Figure 3. Standard for Borderline_SMOTE algorithm to assign minority samples to SAFE set, DANGER set and NOISE set.
Electronics 11 01495 g003
Figure 4. The decision mechanism of SVM_SMOTE algorithm for synthesizing new minority samples. A, B and C are minority class support vector samples.
Figure 4. The decision mechanism of SVM_SMOTE algorithm for synthesizing new minority samples. A, B and C are minority class support vector samples.
Electronics 11 01495 g004
Figure 5. The process of realizing automatic feature combination based on lightGBM model.
Figure 5. The process of realizing automatic feature combination based on lightGBM model.
Electronics 11 01495 g005
Figure 6. The histogram of Table 6.
Figure 6. The histogram of Table 6.
Electronics 11 01495 g006
Figure 7. The ROC curve and AUC value of each fold in the 10-fold cross-validation obtained by classification models on the original dataset. (ad) are the ROC curves of lightGBM, lr, lightGBM + lr and lightGBM + LR classifiers, respectively.
Figure 7. The ROC curve and AUC value of each fold in the 10-fold cross-validation obtained by classification models on the original dataset. (ad) are the ROC curves of lightGBM, lr, lightGBM + lr and lightGBM + LR classifiers, respectively.
Electronics 11 01495 g007
Figure 8. The histogram of Table 7.
Figure 8. The histogram of Table 7.
Electronics 11 01495 g008
Figure 9. The ROC curve and AUC value of each fold in the 10-fold cross-validation obtained by classification models on the dataset processed by SMOTE. (ad) are the ROC curves of lightGBM, lr, lightGBM + lr and lightGBM + LR classifiers, respectively.
Figure 9. The ROC curve and AUC value of each fold in the 10-fold cross-validation obtained by classification models on the dataset processed by SMOTE. (ad) are the ROC curves of lightGBM, lr, lightGBM + lr and lightGBM + LR classifiers, respectively.
Electronics 11 01495 g009
Figure 10. The histogram of Table 8.
Figure 10. The histogram of Table 8.
Electronics 11 01495 g010
Figure 11. The ROC curve and AUC value of each fold in the 10-fold cross-validation obtained by classification models on the dataset processed by Borderline_SMOTE. (ad) are the ROC curves of lightGBM, lr, lightGBM + lr and lightGBM + LR classifiers, respectively.
Figure 11. The ROC curve and AUC value of each fold in the 10-fold cross-validation obtained by classification models on the dataset processed by Borderline_SMOTE. (ad) are the ROC curves of lightGBM, lr, lightGBM + lr and lightGBM + LR classifiers, respectively.
Electronics 11 01495 g011
Figure 12. The histogram of Table 9.
Figure 12. The histogram of Table 9.
Electronics 11 01495 g012
Figure 13. The ROC curve and AUC value of each fold in the 10-fold cross-validation obtained by classification models on the dataset processed by SMOTE_SVM. (ad) are the ROC curves of lightGBM, lr, lightGBM + lr and lightGBM + LR classifiers, respectively.
Figure 13. The ROC curve and AUC value of each fold in the 10-fold cross-validation obtained by classification models on the dataset processed by SMOTE_SVM. (ad) are the ROC curves of lightGBM, lr, lightGBM + lr and lightGBM + LR classifiers, respectively.
Electronics 11 01495 g013
Figure 14. The histogram of Table 10.
Figure 14. The histogram of Table 10.
Electronics 11 01495 g014
Figure 15. The ROC curve and AUC value of each fold in the 10-fold cross-validation obtained by classification models on the dataset processed by SMOTE_Tomek. (ad) are the ROC curves of lightGBM, lr, lightGBM + lr and lightGBM + LR classifiers, respectively.
Figure 15. The ROC curve and AUC value of each fold in the 10-fold cross-validation obtained by classification models on the dataset processed by SMOTE_Tomek. (ad) are the ROC curves of lightGBM, lr, lightGBM + lr and lightGBM + LR classifiers, respectively.
Electronics 11 01495 g015
Figure 16. The histogram of Table 11.
Figure 16. The histogram of Table 11.
Electronics 11 01495 g016
Figure 17. The ROC curve and AUC value of each fold in the 10-fold cross-validation obtained by classification models on the dataset processed by SMOTENC. (ad) are the ROC curves of lightGBM, lr, lightGBM + lr and lightGBM + LR classifiers, respectively.
Figure 17. The ROC curve and AUC value of each fold in the 10-fold cross-validation obtained by classification models on the dataset processed by SMOTENC. (ad) are the ROC curves of lightGBM, lr, lightGBM + lr and lightGBM + LR classifiers, respectively.
Electronics 11 01495 g017
Figure 18. The histogram of Table 12.
Figure 18. The histogram of Table 12.
Electronics 11 01495 g018
Figure 19. The ROC curve and AUC value of each fold in the 10-fold cross-validation obtained by classification models on the dataset processed by data standardization. (ad) are the ROC curves of lightGBM, lr, lightGBM + lr and lightGBM + LR classifiers, respectively.
Figure 19. The ROC curve and AUC value of each fold in the 10-fold cross-validation obtained by classification models on the dataset processed by data standardization. (ad) are the ROC curves of lightGBM, lr, lightGBM + lr and lightGBM + LR classifiers, respectively.
Electronics 11 01495 g019
Figure 20. The histogram of Table 13.
Figure 20. The histogram of Table 13.
Electronics 11 01495 g020
Figure 21. The ROC curve and AUC value of each fold in the 10-fold cross-validation obtained by classification models on the dataset processed by data standardization and SMOTE. (ad) are the ROC curves of lightGBM, lr, lightGBM + lr and lightGBM + LR classifiers, respectively.
Figure 21. The ROC curve and AUC value of each fold in the 10-fold cross-validation obtained by classification models on the dataset processed by data standardization and SMOTE. (ad) are the ROC curves of lightGBM, lr, lightGBM + lr and lightGBM + LR classifiers, respectively.
Electronics 11 01495 g021
Figure 22. The histogram of Table 14.
Figure 22. The histogram of Table 14.
Electronics 11 01495 g022
Figure 23. The ROC curve and AUC value of each fold in the 10-fold cross-validation obtained by classification models on the dataset processed by data standardization and Borderline_SMOTE. (ad) are the ROC curves of lightGBM, lr, lightGBM + lr and lightGBM + LR classifiers, respectively.
Figure 23. The ROC curve and AUC value of each fold in the 10-fold cross-validation obtained by classification models on the dataset processed by data standardization and Borderline_SMOTE. (ad) are the ROC curves of lightGBM, lr, lightGBM + lr and lightGBM + LR classifiers, respectively.
Electronics 11 01495 g023
Figure 24. The histogram of Table 15.
Figure 24. The histogram of Table 15.
Electronics 11 01495 g024
Figure 25. The ROC curve and AUC value of each fold in the 10-fold cross-validation obtained by classification models on the dataset processed by data standardization and SMOTE_SVM. (ad) are the ROC curves of lightGBM, lr, lightGBM + lr and lightGBM + LR classifiers, respectively.
Figure 25. The ROC curve and AUC value of each fold in the 10-fold cross-validation obtained by classification models on the dataset processed by data standardization and SMOTE_SVM. (ad) are the ROC curves of lightGBM, lr, lightGBM + lr and lightGBM + LR classifiers, respectively.
Electronics 11 01495 g025
Figure 26. The histogram of Table 16.
Figure 26. The histogram of Table 16.
Electronics 11 01495 g026
Figure 27. The ROC curve and AUC value of each fold in the 10-fold cross-validation obtained by classification models on the dataset processed by data standardization and SMOTE_Tomek. (ad) are the ROC curves of lightGBM, lr, lightGBM + lr and lightGBM + LR classifiers, respectively.
Figure 27. The ROC curve and AUC value of each fold in the 10-fold cross-validation obtained by classification models on the dataset processed by data standardization and SMOTE_Tomek. (ad) are the ROC curves of lightGBM, lr, lightGBM + lr and lightGBM + LR classifiers, respectively.
Electronics 11 01495 g027
Figure 28. The histogram of Table 17.
Figure 28. The histogram of Table 17.
Electronics 11 01495 g028
Figure 29. The ROC curve and AUC value of each fold in the 10-fold cross-validation obtained by classification models on the dataset processed by data standardization and SMOTENC. (ad) are the ROC curves of lightGBM, lr, lightGBM + lr and lightGBM + LR classifiers, respectively.
Figure 29. The ROC curve and AUC value of each fold in the 10-fold cross-validation obtained by classification models on the dataset processed by data standardization and SMOTENC. (ad) are the ROC curves of lightGBM, lr, lightGBM + lr and lightGBM + LR classifiers, respectively.
Electronics 11 01495 g029
Figure 30. The trend charts of the performance evaluation indexes with class balancing methods on original dataset and standardized dataset.
Figure 30. The trend charts of the performance evaluation indexes with class balancing methods on original dataset and standardized dataset.
Electronics 11 01495 g030
Figure 31. The performance comparison of lightGBM + lr model on six groups of datasets. (af) correspond to the datasets without balancing and the datasets processed by SMOTE, BorderLine_SMOTE, SMOTE_SVM, SMOTE_Tomek and SMOTENC, respectively.
Figure 31. The performance comparison of lightGBM + lr model on six groups of datasets. (af) correspond to the datasets without balancing and the datasets processed by SMOTE, BorderLine_SMOTE, SMOTE_SVM, SMOTE_Tomek and SMOTENC, respectively.
Electronics 11 01495 g031
Figure 32. The performance comparison of lightGBM + LR model on six groups of datasets. (af) correspond to the datasets without balancing and the datasets processed by SMOTE, BorderLine_SMOTE, SMOTE_SVM, SMOTE_Tomek and SMOTENC, respectively.
Figure 32. The performance comparison of lightGBM + LR model on six groups of datasets. (af) correspond to the datasets without balancing and the datasets processed by SMOTE, BorderLine_SMOTE, SMOTE_SVM, SMOTE_Tomek and SMOTENC, respectively.
Electronics 11 01495 g032
Figure 33. Comparison of improvement effects of different modules on performance results.
Figure 33. Comparison of improvement effects of different modules on performance results.
Electronics 11 01495 g033
Figure 34. The change trend of the loss of the two models on the training set.
Figure 34. The change trend of the loss of the two models on the training set.
Electronics 11 01495 g034
Figure 35. The change trend of the loss of the two models on the test set.
Figure 35. The change trend of the loss of the two models on the test set.
Electronics 11 01495 g035
Table 1. Top three causes of whole world DALYs and percentage of all DALYs in 2019.
Table 1. Top three causes of whole world DALYs and percentage of all DALYs in 2019.
OrderCauses 1Percentag 1Causes 2Percentag 2Causes 3Percentag 3Causes 4Percentag 4
1Neonatal disorders7.3Road injuries5.1IHD11.8IHD16.2
2IHD7.2HIV/AIDS4.8Stroke9.3Stroke13.0
3Stroke5.7IHD4.7Diabetes5.1COPD8.5
1 In all age groups. 2 In 25 to 49 years age group. 3 In 50 to 74 years age group. 4 In 75 years and older age group. DALY, disability-adjusted life-year. IHD, ischemic heart disease. COPD, chronic obstructive pulmonary disease. HIV, human immunodeficiency virus. AIDS, acquired immunodeficiency syndrome. The data in the table are from [6].
Table 2. Detailed information of each feature and feature attribute in Z-Alizadeh Sani dataset.
Table 2. Detailed information of each feature and feature attribute in Z-Alizadeh Sani dataset.
CategoryFeature NameRangeType
Demographic featuresAge30–86continuous
Weight48–120continuous
Length140–188continuous
SexMale, Femalecategorical
BMI18.12–40.90continuous
DM0, 1categorical
HTN0, 1categorical
Current smoker0, 1categorical
Ex-smoker0, 1categorical
FH0, 1categorical
Obesity (Yes (BMI > 25), else No)Y, Ncategorical
CRFY, Ncategorical
CVAY, Ncategorical
Airway diseaseY, Ncategorical
Thyroid diseaseY, Ncategorical
CHFY, Ncategorical
DLPY, Ncategorical
Symptoms and Physical examinationBP90.0–190.0continuous
PR50.0–110.0continuous
Edema0, 1categorical
Weak peripheral pulseY, Ncategorical
Lung ralesY, Ncategorical
Systolic murmurY, Ncategorical
Diastolic murmurY, Ncategorical
Typical chest pain0, 1categorical
DyspneaY, Ncategorical
Function class1–4categorical
AtypicalY, Ncategorical
NonanginalY, Ncategorical
Exertional CPNcategorical
LowTH AngY, Ncategorical
ElectrocardiographyQ Wave0, 1categorical
St elevation0, 1categorical
St depression0, 1categorical
T inversion0, 1categorical
LVHY, Ncategorical
Poor R progressionY, Ncategorical
BBBN, LBBB, RBBBcategorical
Laboratory Tests and EchocardiographyFBS62.0–400.0continuous
CR0.5–2.2continuous
TG37.0–1050.0continuous
LDL18.0–232.0continuous
HDL15.9–111.0continuous
BUN6.0–52.0continuous
ESR1–90continuous
HB8.9–17.6continuous
K3.0–6.6continuous
Na128.0–156.0continuous
WBC3700–18,000continuous
Lymph7.0–60.0continuous
Neut32.0–89.0continuous
PLT25.0–742.0continuous
EF-TTE15.0–60.0continuous
Region RWMA0, 1, 2, 3, 4categorical
VHDMild, N, moderate, severecategorical
Table 3. Statistical description of continuous features of Z-Alizadeh Sani dataset.
Table 3. Statistical description of continuous features of Z-Alizadeh Sani dataset.
FeatureMinMaxAveMed1th5th10th50th90th95th99th
Age308658.958.0036.0843.0047.0058.0073.0076.0081.00
Weight4812073.8374.0050.0055.0060.0074.0089.0094.80107.88
Length140188164.72165.00145.00150.00152.00165.00176.60179.00185.96
BMI18.1240.9027.2526.7818.8321.1022.3126.7833.2134.8838.24
BP90190129.55130.0090.00100.00110.00130.00160.00160.00180.00
PR5011075.1470.0060.0064.0070.0070.0087.2090.00109.60
FBS62400119.1898.0069.0477.0080.0098.00193.20223.80358.40
CR0.502.201.061.000.600.700.801.001.401.501.90
TG371050150.34122.0043.1267.4076.00122.00250.00309.00469.20
LDL18232104.64100.0030.2452.6064.40100.00154.20170.00212.60
HDL15.9111.040.2339.0018.1625.2028.0039.0053.0055.8081.40
BUN65217.5016.008.0010.0011.0016.0025.6031.8042.92
ESR19019.4615.001.043.004.0015.0041.0051.0079.84
HB8.917.613.1513.209.0010.0211.0013.2015.1015.6817.19
K3.06.64.234.203.203.503.704.204.805.005.40
Na128156141.00141.00130.04135.00137.00141.00145.00147.00153.00
WBC370018,0007562.057100.003812.004700.005100.007100.0010,700.0012,100.0016,960.00
Lymph76032.4032.009.0415.2019.0032.0044.0049.0058.96
Neut328960.1560.0035.1244.0049.0060.0073.6078.0084.96
PLT25742221.49210.00118.44158.20170.00210.00293.60331.60391.52
EF-TTE156047.2350.0020.0026.0035.0050.0055.0055.0060.00
Min, minimum value. Max, maximum value. Ave, average value. Med, median. th, th percentile.
Table 4. The performance results obtained by the five models on the original dataset.
Table 4. The performance results obtained by the five models on the original dataset.
AlgorithmsAccuracyRecallF1PrecisionSpecificityAUC
RF0.887 ± 0.0560.909 ± 0.0510.923 ± 0.0380.940 ± 0.0510.834 ± 0.1170.92 ± 0.05
Extratrees0.891 ± 0.0450.932 ± 0.0450.923 ± 0.0330.916 ± 0.0550.788 ± 0.1230.91 ± 0.06
AdaBoost0.904 ± 0.0590.929 ± 0.0670.934 ± 0.0380.944 ± 0.0500.842 ± 0.1110.93 ± 0.05
XGBoost0.894 ± 0.0680.909 ± 0.0720.929 ± 0.0440.954 ± 0.0460.856 ± 0.1060.93 ± 0.05
lightGBM0.911 ± 0.0600.922 ± 0.0680.940 ± 0.0380.963 ± 0.0410.881 ± 0.0950.93 ± 0.05
Table 5. Confusion matrix.
Table 5. Confusion matrix.
y pre = 0 y pre = 1
y true = 0 TNFP
y true = 1 FNTP
Table 6. The average results of 10 tests obtained by classification models on the source dataset.
Table 6. The average results of 10 tests obtained by classification models on the source dataset.
ClassifiersAccuracyRecallF1PrecisionSpecificityAUC
lightGBM0.911 ± 0.0600.922 ± 0.0680.940 ± 0.0380.963 ± 0.0410.881 ± 0.0950.93 ± 0.05
lr0.884 ± 0.0520.911 ± 0.0590.921 ± 0.0320.935 ± 0.0420.819 ± 0.0990.93 ± 0.05
lightGBM + lr0.907 ± 0.0470.921 ± 0.0590.937 ± 0.0280.958 ± 0.0380.873 ± 0.0810.93 ± 0.06
lightGBM + LR0.914 ± 0.0450.925 ± 0.0590.942 ± 0.0290.963 ± 0.0350.886 ± 0.0740.93 ± 0.05
Table 7. The average results of 10 tests obtained by classification models on the dataset processed by SMOTE.
Table 7. The average results of 10 tests obtained by classification models on the dataset processed by SMOTE.
ClassifiersAccuracyRecallF1PrecisionSpecificityAUC
lightGBM0.931 ± 0.0550.930 ± 0.0900.934 ± 0.0480.944 ± 0.0350.932 ± 0.0390.96 ± 0.03
lr0.915 ± 0.0570.932 ± 0.0960.917 ± 0.0490.911 ± 0.0580.897 ± 0.0470.96 ± 0.03
lightGBM + lr0.933 ± 0.0540.950 ± 0.0830.934 ± 0.0470.926 ± 0.0520.916 ± 0.0530.97 ± 0.04
lightGBM + LR0.940 ± 0.0480.945 ± 0.0780.942 ± 0.0430.944 ± 0.0410.935 ± 0.0370.97 ± 0.03
Table 8. The average results of 10 tests obtained by classification models on the dataset processed by Borderline_SMOTE.
Table 8. The average results of 10 tests obtained by classification models on the dataset processed by Borderline_SMOTE.
ClassifiersAccuracyRecallF1PrecisionSpecificityAUC
lightGBM0.933 ± 0.0410.938 ± 0.0810.935 ± 0.0350.939 ± 0.0420.928 ± 0.0330.97 ± 0.03
lr0.903 ± 0.0520.919 ± 0.0870.904 ± 0.0460.897 ± 0.0590.887 ± 0.0490.96 ± 0.03
lightGBM + lr0.933 ± 0.0470.949 ± 0.0680.933 ± 0.0460.921 ± 0.0570.917 ± 0.0560.97 ± 0.03
lightGBM + LR0.938 ± 0.0530.938 ± 0.0810.940 ± 0.0480.948 ± 0.0580.938 ± 0.0580.97 ± 0.03
Table 9. The average results of 10 tests obtained by classification models on the dataset processed by SMOTE_SVM.
Table 9. The average results of 10 tests obtained by classification models on the dataset processed by SMOTE_SVM.
ClassifiersAccuracyRecallF1PrecisionSpecificityAUC
lightGBM0.931 ± 0.0520.937 ± 0.0840.933 ± 0.0460.935 ± 0.0430.925 ± 0.0440.97 ± 0.03
lr0.903 ± 0.0590.926 ± 0.0870.903 ± 0.0550.888 ± 0.0680.880 ± 0.0620.95 ± 0.04
lightGBM + lr0.926 ± 0.0590.936 ± 0.0870.928 ± 0.0520.926 ± 0.0430.916 ± 0.0500.95 ± 0.04
lightGBM + LR0.928 ± 0.0620.933 ± 0.0980.932 ± 0.0540.939 ± 0.0470.924 ± 0.0490.96 ± 0.03
Table 10. The average results of 10 tests obtained by classification models on the dataset processed by SMOTE_Tomek.
Table 10. The average results of 10 tests obtained by classification models on the dataset processed by SMOTE_Tomek.
ClassifiersAccuracyRecallF1PrecisionSpecificityAUC
lightGBM0.946 ± 0.0440.950 ± 0.0650.948 ± 0.0400.948 ± 0.0400.942 ± 0.0440.97 ± 0.03
lr0.926 ± 0.0510.943 ± 0.0830.927 ± 0.0440.917 ± 0.0430.904 ± 0.0460.96 ± 0.04
lightGBM + lr0.946 ± 0.0490.951 ± 0.0710.948 ± 0.0450.949 ± 0.0400.941 ± 0.0450.97 ± 0.04
lightGBM + LR0.941 ± 0.0380.954 ± 0.0640.942 ± 0.0340.933 ± 0.0340.925 ± 0.0360.97 ± 0.04
Table 11. The average results of 10 tests obtained by classification models on the dataset processed by SMOTENC.
Table 11. The average results of 10 tests obtained by classification models on the dataset processed by SMOTENC.
ClassifiersAccuracyRecallF1PrecisionSpecificityAUC
lightGBM0.931 ± 0.0400.938 ± 0.0800.932 ± 0.0350.935 ± 0.0520.924 ± 0.0430.97 ± 0.03
lr0.912 ± 0.0510.936 ± 0.0870.912 ± 0.0460.897 ± 0.0600.888 ± 0.0540.97 ± 0.02
lightGBM + lr0.933 ± 0.0480.946 ± 0.0750.934 ± 0.0440.926 ± 0.0370.920 ± 0.0390.96 ± 0.04
lightGBM + LR0.931 ± 0.0520.933 ± 0.0920.934 ± 0.0440.944 ± 0.0460.929 ± 0.0380.97 ± 0.03
Table 12. The average results of 10 tests obtained by classification models on the dataset processed by data standardization.
Table 12. The average results of 10 tests obtained by classification models on the dataset processed by data standardization.
ClassifiersAccuracyRecallF1PrecisionSpecificityAUC
lightGBM0.897 ± 0.0490.909 ± 0.0580.931 ± 0.0300.958 ± 0.0440.869 ± 0.0930.92 ± 0.06
lr0.884 ± 0.0590.900 ± 0.0620.922 ± 0.0370.949 ± 0.0450.844 ± 0.1050.93 ± 0.06
lightGBM + lr0.914 ± 0.0400.922 ± 0.0580.942 ± 0.0250.968 ± 0.0360.895 ± 0.0740.92 ± 0.06
lightGBM + LR0.914 ± 0.0600.931 ± 0.0740.942 ± 0.0380.958 ± 0.0380.871 ± 0.0860.93 ± 0.06
Table 13. The average results of 10 tests obtained by classification models on the dataset processed by data standardization and SMOTE.
Table 13. The average results of 10 tests obtained by classification models on the dataset processed by data standardization and SMOTE.
ClassifiersAccuracyRecallF1PrecisionSpecificityAUC
lightGBM0.924 ± 0.0670.931 ± 0.1000.927 ± 0.0580.930 ± 0.0430.917 ± 0.0480.97 ± 0.03
lr0.912 ± 0.0490.923 ± 0.0810.913 ± 0.0440.911 ± 0.0580.902 ± 0.0530.96 ± 0.03
lightGBM + lr0.935 ± 0.0350.942 ± 0.0660.936 ± 0.0330.935 ± 0.0380.929 ± 0.0350.97 ± 0.03
lightGBM + LR0.935 ± 0.0450.946 ± 0.0710.936 ± 0.0420.930 ± 0.0490.924 ± 0.0490.97 ± 0.03
Table 14. The average results of 10 tests obtained by classification models on the dataset processed by data standardization and Borderline_SMOTE.
Table 14. The average results of 10 tests obtained by classification models on the dataset processed by data standardization and Borderline_SMOTE.
ClassifiersAccuracyRecallF1PrecisionSpecificityAUC
lightGBM0.935 ± 0.0480.944 ± 0.0790.937 ± 0.0440.935 ± 0.0480.926 ± 0.0440.97 ± 0.03
lr0.914 ± 0.0520.934 ± 0.0810.914 ± 0.0480.902 ± 0.0650.895 ± 0.0640.96 ± 0.02
lightGBM + lr0.940 ± 0.0360.957 ± 0.0540.939 ± 0.0350.926 ± 0.0480.923 ± 0.0470.97 ± 0.03
lightGBM + LR0.947 ± 0.0360.961 ± 0.0540.947 ± 0.0340.935 ± 0.0430.932 ± 0.0420.97 ± 0.03
Table 15. The average results of 10 tests obtained by classification models on the dataset processed by data standardization and SMOTE_SVM.
Table 15. The average results of 10 tests obtained by classification models on the dataset processed by data standardization and SMOTE_SVM.
ClassifiersAccuracyRecallF1PrecisionSpecificityAUC
lightGBM0.938 ± 0.0580.944 ± 0.0800.939 ± 0.0520.939 ± 0.0520.932 ± 0.0560.97 ± 0.03
lr0.917 ± 0.0470.936 ± 0.0820.917 ± 0.0430.906 ± 0.0610.898 ± 0.0560.96 ± 0.03
lightGBM + lr0.942 ± 0.0400.963 ± 0.0650.942 ± 0.0370.926 ± 0.0380.922 ± 0.0370.96 ± 0.04
lightGBM + LR0.945 ± 0.0360.951 ± 0.0620.945 ± 0.0330.944 ± 0.0410.939 ± 0.0360.96 ± 0.03
Table 16. The average results of 10 tests obtained by classification models on the dataset processed by data standardization and SMOTE_Tomek.
Table 16. The average results of 10 tests obtained by classification models on the dataset processed by data standardization and SMOTE_Tomek.
ClassifiersAccuracyRecallF1PrecisionSpecificityAUC
lightGBM0.937 ± 0.0480.939 ± 0.0850.940 ± 0.0410.948 ± 0.0390.936 ± 0.0320.98 ± 0.02
lr0.912 ± 0.0510.926 ± 0.0890.913 ± 0.0450.911 ± 0.0660.898 ± 0.0550.96 ± 0.03
lightGBM + lr0.947 ± 0.0500.948 ± 0.0730.948 ± 0.0460.953 ± 0.0470.945 ± 0.0480.98 ± 0.02
lightGBM + LR0.944 ± 0.0480.952 ± 0.0750.946 ± 0.0430.944 ± 0.0460.936 ± 0.0450.97 ± 0.03
Table 17. The average results of 10 tests obtained by classification models on the dataset processed by data standardization and SMOTENC.
Table 17. The average results of 10 tests obtained by classification models on the dataset processed by data standardization and SMOTENC.
ClassifiersAccuracyRecallF1PrecisionSpecificityAUC
lightGBM0.931 ± 0.0580.934 ± 0.0890.933 ± 0.0520.939 ± 0.0560.928 ± 0.0550.97 ± 0.03
lr0.917 ± 0.0460.926 ± 0.0860.919 ± 0.0420.921 ± 0.0610.908 ± 0.0520.96 ± 0.03
lightGBM + lr0.945 ± 0.0500.942 ± 0.0800.947 ± 0.0450.958 ± 0.0490.947 ± 0.0470.97 ± 0.03
lightGBM + LR0.938 ± 0.0430.944 ± 0.0740.939 ± 0.0400.940 ± 0.0550.931 ± 0.0500.97 ± 0.02
Table 18. The classification results of the prediction model after removing the modules of data preprocessing.
Table 18. The classification results of the prediction model after removing the modules of data preprocessing.
ModulesAccuracyRecallF1PrecisionSpecificityAUC
ALL0.947 ± 0.0500.948 ± 0.0730.948 ± 0.0460.953 ± 0.0470.945 ± 0.0480.98 ± 0.02
NP0.946 ± 0.0490.951 ± 0.0710.948 ± 0.0450.949 ± 0.0400.941 ± 0.0450.97 ± 0.04
P0.001−0.0030.0000.0040.0040.01
NP, no data preprocessing. P, data preprocessing.
Table 19. The classification results of the prediction model after removing the modules of class balance processing.
Table 19. The classification results of the prediction model after removing the modules of class balance processing.
ModulesAccuracyRecallF1PrecisionSpecificityAUC
ALL0.947 ± 0.0500.948 ± 0.0730.948 ± 0.0460.953 ± 0.0470.945 ± 0.0480.98 ± 0.02
NBC0.914 ± 0.0400.922 ± 0.0580.942 ± 0.0250.968 ± 0.0360.895 ± 0.0740.92 ± 0.06
BC0.0330.0260.006−0.0150.0500.06
NBC, no methods of balancing classes. BC, balancing classes.
Table 20. The classification results of the prediction model after removing the modules of feature combination.
Table 20. The classification results of the prediction model after removing the modules of feature combination.
ModulesAccuracyRecallF1PrecisionSpecificityAUC
ALL0.947 ± 0.0500.948 ± 0.0730.948 ± 0.0460.953 ± 0.0470.945 ± 0.0480.98 ± 0.02
NFC0.912 ± 0.0510.926 ± 0.0890.913 ± 0.0450.911 ± 0.0660.898 ± 0.0550.96 ± 0.03
FC0.0350.0220.0350.0420.0470.02
NFC, no feature combination. FC, feature combination.
Table 21. The comparison of classification results between our study and other studies on the Z-Alizadeh Sani dataset.
Table 21. The comparison of classification results between our study and other studies on the Z-Alizadeh Sani dataset.
MethodAccuracy %Recall %F1%Precision %Specificity %AUC
SMO [33]92.0997.22NN79.31N
SMO + IG [30]94.0896.30NN88.51N
KNN [43]90.9193.3393.3393.3385.71N
NN + genetic algorithm [32]93.8597NN92N
NB + genetic algorithm [69]88.1688.00NN87.78N
Ensemble [70]86.4973.610.75N91.670.83
SVM + feature engineering + 500 examples [37]96.4100NN88.10.92
NE-nu-SVC [35]94.6694.7094.7094.70N0.966
N2GC-nuSVM [36]93.08N91.51NNN
XGBoost + hybrid FSA + FA + ETCA + SMOTE [31]92.5892.9990.6292.59NN
Hybrid PSO-EmNN coupled with feature selection [38]88.3491.8592.1292.3778.98N
XGBoost + feature construction + SMOTE [48]94.796.194.693.493.20.98
GSVMA [47]89.4581.2280.49N100100
C-CADZ [41]97.3798.15NN95.45N
CART [46]92.4198.61NN77.01N
lightGBM + lr + SMOTE_Tomek *94.794.894.895.394.50.98
* Our proposed methods.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhang, S.; Yuan, Y.; Yao, Z.; Yang, J.; Wang, X.; Tian, J. Coronary Artery Disease Detection Model Based on Class Balancing Methods and LightGBM Algorithm. Electronics 2022, 11, 1495. https://doi.org/10.3390/electronics11091495

AMA Style

Zhang S, Yuan Y, Yao Z, Yang J, Wang X, Tian J. Coronary Artery Disease Detection Model Based on Class Balancing Methods and LightGBM Algorithm. Electronics. 2022; 11(9):1495. https://doi.org/10.3390/electronics11091495

Chicago/Turabian Style

Zhang, Shasha, Yuyu Yuan, Zhonghua Yao, Jincui Yang, Xinyan Wang, and Jianwei Tian. 2022. "Coronary Artery Disease Detection Model Based on Class Balancing Methods and LightGBM Algorithm" Electronics 11, no. 9: 1495. https://doi.org/10.3390/electronics11091495

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop