Next Article in Journal
Epidemiological Investigation: Important Measures for the Prevention and Control of COVID-19 Epidemic in China
Next Article in Special Issue
GERPM: A Geographically Weighted Stacking Ensemble Learning-Based Urban Residential Rents Prediction Model
Previous Article in Journal
Development of a New Zeta Formula and Its Role in Riemann Hypothesis and Quantum Physics
Previous Article in Special Issue
Some Modified Ridge Estimators for Handling the Multicollinearity Problem
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Early Identification of Risk Factors in Non-Alcoholic Fatty Liver Disease (NAFLD) Using Machine Learning

by
Luis Rolando Guarneros-Nolasco
1,
Giner Alor-Hernández
2,
Guillermo Prieto-Avalos
3 and
José Luis Sánchez-Cervantes
4,*
1
Tecnologías de la Información y Comunicación Área Desarrollo de Software, Universidad Tecnológica del Centro de Veracruz, Av. Universidad No. 350, Carretera Federal Cuitláhuac-La Tinaja, Loc. Dos Caminos, Cuitláhuac C.P. 94910, Veracruz, Mexico
2
Tecnológico Nacional de México Campus Orizaba, Av. Oriente 9 No. 852 Col. Emiliano Zapata, Orizaba C.P. 94320, Veracruz, Mexico
3
Departamento de Ingeniería Eléctrica—Electrónica, Tecnológico Nacional de México Campus Mexicali, Av. Instituto Tecnológico S/N, Col. Plutarco Elías Calles, Mexicali C.P. 21376, Baja California, Mexico
4
CONACYT-Instituto Tecnológico de Orizaba, Av. Oriente 9 No. 852 Col. Emiliano Zapata, Orizaba C.P. 94320, Veracruz, Mexico
*
Author to whom correspondence should be addressed.
Mathematics 2023, 11(13), 3026; https://doi.org/10.3390/math11133026
Submission received: 28 May 2023 / Revised: 4 July 2023 / Accepted: 4 July 2023 / Published: 7 July 2023
(This article belongs to the Special Issue Advances in Machine Learning and Applications)

Abstract

:
Liver diseases are a widespread and severe health concern, affecting millions worldwide. Non-alcoholic fatty liver disease (NAFLD) alone affects one-third of the global population, with some Latin American countries seeing rates exceeding 50%. This alarming trend has prompted researchers to explore new methods for identifying those at risk. One promising approach is using Machine Learning Algorithms (MLAs), which can help predict critical factors contributing to liver disease development. Our study examined nine different MLAs across four datasets to determine their effectiveness in predicting this condition. We analyzed each algorithm’s performance using five important metrics: accuracy, precision, recall, f1-score, and roc_auc. Our results showed that these algorithms were highly effective when used individually and as part of an ensemble modeling technique such as bagging or boosting. We identified alanine aminotransferase (ALT), aspartate aminotransferase (AST), alkaline phosphatase (ALP), and albumin as the top four attributes most strongly associated with non-alcoholic fatty liver disease risk across all datasets. Gamma-glutamyl transpeptidase (GGT), hemoglobin, age, and prothrombin time also played significant roles. In conclusion, this research provides valuable insights into how we can better detect and prevent non-alcoholic fatty liver diseases by leveraging advanced machine learning techniques. As such, it represents an exciting opportunity for healthcare professionals seeking more accurate diagnostic tools while improving patient outcomes globally.

1. Introduction

According to the National Institute of Statistics and Geography of Mexico (INEGI, Spanish acronym), liver disease is the sixth leading cause of death in Mexico [1], mainly affecting people over 25. At least in 2020, Mexico recorded 41,492 deaths related to liver disease, so the prevalence of NAFLD exceeds 50% and is considered a national health problem [2].
The liver has many important functions, including food digestion and nutrient processing and distribution. A possible sign of liver disease is when the skin turns yellow, which is known as jaundice. Among the many different liver diseases, some, such as hepatitis, are caused by viruses, whereas others may be the effect of excessive alcohol consumption or a long-term injury to the liver (i.e., cirrhosis). NAFLD is one of the most important liver diseases [3]. It can be described as a prolonged and progressive disorder characterized by hepatic steatosis in people who consume less than 20g of alcohol per day. Mexico has been attributed to the escalating prevalence of NAFLD risk factors, with more than 50% of its population showing at least one of these factors. Therefore, the medium-term for NAFLD sufferers is very pessimistic if immediate actions are not taken to reduce what is already considered a federal health problem [4].
Several authors have identified biomarkers to be used as a roadmap for diagnosing NAFLD. Among the most important are alanine aminotransferase, gamma-glutamyl transpeptidase [5], aspartate aminotransferase [6], alkaline phosphatase, age, and albumin [7]. Therefore, this analysis aims to identify the main risk factors for NAFLD in four datasets of clinical data from patients with liver disease by applying machine learning (ML) techniques, which is a promising technology for medical data analysis and prediction for disease diagnosis and early detection, to contribute to healthcare professionals improving better preventive diagnosis of NAFLD and more accurate and timely treatment. The algorithms are tested using boosting, bagging (ensemble methods), and non-ensemble methods.
The remainder of this article is organized as follows: Section 2 discusses current research on MLA in clinical datasets, MLA performance metrics, ensembles, and clinical and data sets are available in the data science community repository. Section 3 presents our MLA evaluation methodology. In Section 4, we present and discuss our results. Finally, in Section 5, we conclude and provide suggestions for future work.

2. Related Work

In this section, we analyze a series of research works that use ensemble learning for disease diagnosis and prevention. We classify the examined research into five primary ensemble learning trends: bagging, boosting, stacking, bagging-boosting, and bagging-boosting-stacking. These ensembles are widely implemented for the prediction and diagnosis of diseases such as schizophrenia, breast cancer, cardiovascular diseases, and liver diseases, among others.

2.1. Bagging Ensembles

In their work, Lin et al. [8] established a bagging ensemble and compared it with other algorithms to determine their efficiency in terms of schizophrenia detection. Next, the researchers found that a bagging ensemble with attribute selection could be an applicable method to support software tools for predicting the efficient results of schizophrenia. Ponnaganti and Anitha [9] suggested an Ensemble Bagging Weighted Voting Classification (EBWvc) method for the categorization of breast cancer. The researchers comparatively evaluated five metrics, and their results showed an increased performance of the EBWvc method if compared to similar existing classification methods. Chicco and Jurman [10] analyzed a set of EHRs and relied on MLA to predict the diagnosis of a series of liver diseases. The results confirmed the utility of ensemble learning for predicting the diagnosis of hepatitis C and cirrhosis. In the study of Anisha and Saranya [11], the researchers proposed a system for the early diagnosis of stroke disorder that relies on a homogeneous logistic regression ensemble classifier. The results of the experiment revealed a higher accuracy of the system than simple logistic regression initiatives. In turn, Devi et al. [12] proposed a method to analyze the clinical features influencing liver activity. The researchers applied Gradient Boost Regressor, Extra Tree Regressor, and Random Forest Regressor to the analyzed clinical data to find the highest feature importance. Then, the data were fitted to a set of classifiers to analyze performance metrics before and after feature scaling. As the main findings, Bagging Classifier was found to retain the accuracy of 72% before and after feature scaling with the top features from Random Forest Regressor.
In their work, Lin et al. [13] compared a bagging ensemble with other algorithms and found that a bagging ensemble framework may provide a suitable approach to building tools for predicting cognitive function in schizophrenia. From a similar perspective, the research of Ejiofor and Ochei [14] relied on the strong capability of the ensemble bagging machine learning technique for predicting breast cancer. The bagging ensemble was implemented on two MLAs as base learners. The validation exhibited an accuracy value of 74% and 51% for Linear Regression (LR) and Decision Tree (DT), respectively. Rahman and Mahmood [15] proposed a model for identifying key biomarkers for heart disease. To this end, the researchers implemented three classification algorithms: Random Forest (RF), K-Nearest Neighbors (KNN), and Naïve Bayes, and three approaches: Lasso, Mutual Information, and Recursive Feature Elimination. Additionally, the researchers employed the bagging ensemble approach with RF to improve the results of their model, which exhibited 85.18% of accuracy through Recursive Feature Elimination with Bagging (Random Forest). In their work, Thomgkam et al. [16] explored the performance and effectiveness of machine learning methods. Their findings revealed that RF is superior to bagging in all the analyzed criteria. In turn, S. Yadav and Singh [17] implemented a model with classifiers to categorize patients affected by Parkinson’s disease. As their main findings, Support Vector Classifier exhibited the highest accuracy (93.83%) on the basic classifiers, whereas bagging exhibited 73.28% accuracy from the ensemble technique.

2.2. Boosting Ensembles

In their work, Buyrukoglu [18] suggested ensemble learning methods to improve classification performance compared to single MLAs. The best classification performance of ensemble methods was achieved by boosting the ensemble (AdaBoost) (92.7%). The study found that the classification rate with ensemble learning methods increased between 3.2% and 7.2% compared to the AdaBoost ensemble method with single-based ML approaches. Researchers A. Singh et al. [19] constructed an intelligent hybrid approach for the identification of hepatitis. To avoid clinical experience, the approach relies on ensemble learning to reduce estimation time using many learners to solve a liver disease problem. After comparing different metrics, the findings revealed that both the k-means clustering and improved ensemble learning methodology achieved improved prediction results than other existing individual and integrated models. In Sarvestany et al. [20], the researchers trained six MLAs to identify advanced fibrosis. The MLA exhibiting the highest accuracy was an ensemble algorithm of classical algorithms. In turn, Dutta et al. [21] discussed liver disease diagnosis through different data mining algorithms. The best algorithm for liver disease detection was found to be DT, achieving an accuracy of 99.96%. The study of Verma and Mentha [22] suggested a novel ensemble learning algorithm for the classification of five datasets of the University of California Irvine (UCI). The researchers concluded that the suggested ensemble learning method is remarkably appropriate for handling the classification problem in the bioinformatics domain.

2.3. Stacking Ensembles

In their work, Meng [23] built a stacking ensemble model to predict the disorder evolution and medical results of Alpha-1-antitrypsin deficiency-associated liver disease (AATD-LD). The model uses meta-learning by mixing several supervised MLAs and can be implemented to predict the clinical outcome of other similar diseases. Al Telaq and Hewahi [24] proposed a method to predict liver disease by applying multiple MLAs on a public liver disease dataset. The results revealed that ensemble learning of different classifiers exhibited the highest accuracy (88%). In Gupta and Gupta [25], the researchers recommended a stacking architecture for efficiently predicting the diagnosis of cervical cancer. In Ponnaganti and Anitha [9], the researchers built a decision support system using the ensemble model with different metrics to evaluate the performance of the model. The researchers showed that the proposed method recorded a notable accuracy of 97% in classifying breast cancer data. Pouriyeh et al. [26] research aimed to evaluate the accuracy of different data mining classifiers for heart disease prediction using ensemble ML. The study used the Cleveland heart disease dataset, different classifiers, and applied classifier ensemble prediction, bagging, boosting, and stacking to the dataset. The results of the experiment revealed that the Support Vector Machines (SVM) method with boosting outperformed the other methods. Kabir and Ludwing [27] attempted to improve the classification performance of multiple MLAs using super learning or stacked ensembles. The experiment results showed that super learning had a better classification performance than the individual base learners. Doganer et al. [28] compared the performance of different MLAs and found that stacking ensembles outperformed boosting, voting, bagging ensembles, and ML methods.

2.4. Bagging and Boosting Ensembles

Hakim et al. [29] evaluated the implementation of two ensemble-based MLAs of bagging and boosting with five different base classifiers for predicting myocardial infarction. The results showed that bagging with RF achieved higher accuracy. The research of Yadav and Pal [30] applied rule-based classification algorithms on a prepared dataset with three selected algorithms by bagging and boosting ensemble methods and calculated four metrics on a diabetes dataset. The research revealed that bagging exhibited the highest accuracy, namely, 98%. In Gao et al. [31], the researchers used different ensemble learning methods to improve heart disease prediction and compared these methods with MLAs. The results revealed that the bagging ensemble learning method with DT achieved the best performance. In the study by Taser [32], bagging and boosting methods were implemented using six different decision tree-based (DTB) classifiers on experimental data to predict diabetes. The experimental results showed that bagging and boosting outperformed the individual DTB classifiers. In Niranjana et al. [33], researchers evaluated the performance of different classification ensembles in the early detection of coronary heart disease based on risk factors. The Bagged Trees Ensemble Classifier was found to have the highest classification accuracy using four performance metrics. The research of Fraiwan et al. [34] studies various DTB ensemble methods to tackle the challenge of multi-class classification, and with their proposed approach, these methods achieved the highest classification accuracy (99.17%). The work of Dhilsath et al. [35] relied on two boosting classifiers and one bagging classifier to build a model for heart disease prediction. The research aims to evaluate the efficiency of grid search algorithms and random search algorithms by tuning the gradient boost parameters (GB), Adaboost, and RF. The findings revealed that the tunning strategy increased the ensemble learner’s efficiency.

2.5. Bagging, Boosting, and Stacking Ensembles

In their work, Khanam et al. [36] applied an ExtraTreeClassifier (ETC) method to find the highly significant features of cervical cancer using different ensemble methods, including bagging, boosting, and stacking. The results of the experiment revealed that stacking combined with RF, SVM, ETC, Extreme Gradient Boosting (XGBoost), and bagging exhibited the highest accuracy (94.4%). In Niranjana et al. [33], the research aimed at predicting coronary heart disease by applying a risk factor method. Using predictive techniques, the researchers evaluated only three metrics, and their results showed that the stacked ensemble was the most effective in terms of accuracy. In Bang et al. [37], the researchers used 18 ML classifiers to determine prediction models of curative resection with different variables on early gastric cancer (U-EGC). As the main finding, XGBoost exhibited the best performance.
According to our literature review, the top eight MLAs used in different ensembles to detect and diagnose diseases such as heart disease, diabetes, breast cancer, liver disease, and schizophrenia include DT, RF, KNN, LR SVM, Artificial Neural Network (ANN), GB, and AdaBoost. In addition, current initiatives to detect and diagnose these chronic diseases are mainly based on the use of bagging and boosting ensembles with KNN, SVM, AdaBoost, RF, DT, NN, and logistic regression algorithms.

3. Materials and Methods

The following section describes how we analyzed the performance of nine MLAs independently and with two ML ensembles (bagging and boosting) on the four liver disease datasets.

3.1. Datasets

We identified four main datasets on clinical open access for liver diseases: the BUPA liver disorders (BLD) dataset, the Hepatocellular Carcinoma (HCC) Survival dataset, the Indian Liver Patient Dataset (ILPD), and the Cirrhosis Prediction Dataset (CPD). Table 1 summarizes the characteristics of each of these datasets.
The BUPA liver disease database [38] is an open access database which is maintained in the Irvine University of California (UCI) online repository. It is often used to predict liver disorders based on blood tests and alcohol consumption. In addition, the BUPA dataset contains two classes, six numerical attributes, and 345 records (Table 2).
The data in the HCC Survival Dataset were collected from a university hospital in Portugal [39]. The dataset is open access, and it is stored in the online repository of UCI [40]. It contains accurate clinical data on 165 HCC patients. In addition, the HCC Survival Dataset contains two classes, 49 numerical attributes, and 165 records (Table 3).
The ILPD [41] contains 146 liver patient records and 167 non-liver patient records. The dataset was collected from patients in the northeast of Andhra Pradesh, India (Table 4).
The CPD [42] contains data collected from de Mayo Clinic trial in primary biliary cirrhosis (PBC) with a total of 424 PBC patients and 20 attributes. It is an open access dataset stored in the online repository of Kaggle (Table 5).

3.2. Machine Learning Classifiers

We have employed nine distinct classifying procedures from various domains of machine learning. The classifiers adopt a linear statistical approach, including LR [43], three tree-based techniques: RF [44], ExtraTrees (ETT) [45], and DT [46]; one SVM model [47]; an instance-based learning algorithm [48]. Additionally, we utilized three ensemble boosting methods consisting of GB [49], LightGBM (LGBM) [50], and AdaBoost [51]. We evaluated the performance of each method independently as well as with two ML ensembles- bagging technique [52] and boosting methodology [53]. Finally, all results were accurately documented for further analysis purposes.

3.3. Methodology

In order to evaluate the main risk factors of liver disease on clinical datasets, we employed a six-staged methodology as outlined in Guarneros-Nolasco et al. [54]. To enhance our analysis, we incorporated ensemble techniques such as bagging and boosting (as depicted in Figure 1) into this process. The Python programming language [55] and the Scikit–Learn [56] library algorithms were used. The stages involved were: (1) loading data dataset; (2) pre-processing data; (3) selecting attributes; (4) running ML models with bagging and boosting ensembles; (5) applying evaluation metrics; and finally; (6) processing MLA/classifier/ensemble performance results.
Each strategy arrangement can be depicted as follows:
  • Load data dataset. Select and stack information from the dataset containing clinical records of patients with liver diseases;
  • Pre-process dataset. Audit stacked information to get their content. At that point, select the classification variable to obtain the best results.
  • Select attributes or main risk factors. Use RF to select the top two and top four attributes from each dataset. Split data for training and testing (i.e., 70% for training and 30% for testing) and k = 10 cross-validation. Similary, calculate the best parameters for RandomizedSearchCV for n_estimators, max_attributes, and max_depth. Most of the algorithms have these parameters in common, except for KNN. The parameter ramdom_state was set to 42 in all the assessments;
  • Run ML classifiers, bagging ensemble, and boosting ensemble: Apply the nine ML classifiers to observe members with liver illnesses from healthy people. Tune bagging and boosting parameters such as n_estimators and max_samples on the train and test split and cross-validation techniques;
  • Apply evaluation metrics. Analyze MLA classification performance with respect to five criteria: accuracy, precision, recall, f1-score, and area under the curve (ROC-AUC);
  • Process performance results. Assemble and compare execution values from the nine MLAs with the bagging and boosting ensembles and record such outcomes for further analysis. At that point, select the best-performing MLA or ensemble.

4. Results and Discussion

In this section, we discuss the results of the ML performance analyses for five performance evaluation metrics or criteria (see step 5 of the methodology). We performed the performance evaluations of the classifiers first using the train-test split method (70–30%) and then using k-fold cross-validation (k = 10). We recorded five performance measures during the evaluations: accuracy, precision, recall, f1 score, and roc_auc.

4.1. Attribute Selection in Datasets

We analyzed the performance of RF to identify the top four attributes in the datasets.
(a)
The BUPA Liver Disorders Dataset. We applied RF with the six numerical attributes of the dataset to identify and select the six most important ones. Figure 2a depicts the ranking of these attributes from the most important to the least important;
(b)
HCC Survival Dataset. The 49 attributes were ranked using RF on the HCC Survival dataset. Figure 2b depicts a graph of said ranking. As in the previous case, the top six attributes were used in the classifier performance;
(c)
ILPD. The 10 attributes were ranked using RF on the ILPD dataset. Figure 2c depicts a graph ranking the first six attributes, of which the top six were used in the analysis;
(d)
CPD. The 19 attributes were ranked using RF in this dataset. Figure 2d graphically shows the ranking of said attributes, of which the top six were used in the analysis.
Figure 2. Ranking of attributes selected by RF for predictions on (a) 6/6 attributes on the BUPA dataset, (b) 6/49 attributes on the HCC Survival Dataset, (c) 6/10 attributes on ILPD, and (d) 6/19 attributes on CPD.
Figure 2. Ranking of attributes selected by RF for predictions on (a) 6/6 attributes on the BUPA dataset, (b) 6/49 attributes on the HCC Survival Dataset, (c) 6/10 attributes on ILPD, and (d) 6/19 attributes on CPD.
Mathematics 11 03026 g002

4.2. Results

We analyzed the performance of the nine ML classifiers (i.e., AdaBoost, DT, ETT, GB, KNN, LGBM, LR, RG, and SVC) on the top four attributes of each dataset using the train-test data split technique (70–30%), cross-validation and no ensembles, bagging ensemble, and boosting ensemble. We tuned bagging and boosting parameters such as n_estimators and max_samples with the train and test split and cross-validation techniques. The analysis results are discussed below.

4.2.1. Classifier Performance on the BUPA Dataset

We tested the performance of the nine ML classifiers (AdaBoost, DT, ET, GB, KNN, LGBM, LR, RF, and SVC) on the top four attributes of the BUPA dataset. These attributes comprised gammagt (score = 20.84), sgpt (score = 20.48), sgot (score = 16.48), and alkphos (score = 16.42). Table 6, Table 7 and Table 8 summarize the results of our analysis.
As can be observed from Table 6, with non-ensemble learning, we calculated algorithm performance using default values. With train and test split, LR exhibited the best results in terms of accuracy (73.08%). When using cross-validation, SVC exhibited the highest accuracy (71.60%). Conversely, the lowest-performing classifiers with respect to accuracy included DT and LGBM (65.38%) in train and test split and DT (58.87%) in cross-validation. Regarding precision, KNN proved to be the best-performing classifier in train and test split, while LGBM in cross-validation, with precision scores of 74.51% and 73.53%, respectively. The lowest-performing classifiers in terms of precision were AdaBoost (69.23%) with train and test split and DT (64.53%) with cross-validation. In conclusion, on the BUPA dataset with train and test split, LR exhibited good performance in terms of accuracy, f1-score, and roc_auc, whereas KNN outperformed in precision, and SVC outperformed in terms of recall. As for cross-validation, SVC exhibited the best results in accuracy, recall, f1-score, and roc_auc, whereas LGBM exhibited the best performance in precision.
With the bagging ensemble classifier, we used number of trees = 50 and max_samples = 0.5. As can be observed from Table 7, ETT achieved the highest accuracy (74.04%) with train and test split, whereas GB exhibited the highest accuracy with cross-validation (71.61%). Conversely, the lowest-performing classifiers with respect to accuracy included LGBM (65.38%) in train and test split and KNN (67.50%) in cross-validation. As for precision, ET proved to be the best-performing classifier (74.63%) when using train and test split and DT when using cross-validation (73.99%). The lowest-performing classifiers in terms of precision were LGBM (65.18%) in train and test split and KNN (67.95%) in cross-validation. In conclusion, on the BUPA dataset, when using the train and test split, ETT exhibited good performance in terms of accuracy, precision, f1-score, and roc_auc, whereas SVC exhibited the best results in recall. Conversely, when using cross-validation, GB achieved the best results in accuracy, and roc_auc and DT outperformed the other classifiers in terms of precision, and SVC exhibited the best performance in recall and f1-score.
With the boosting ensemble classifier, we used the best arguments for the BUPA dataset, including n_estimators = 200, max_depth = None, and max_features = sqrt. Our results revealed that when using train and test split, AdaBoost and GB exhibited the best accuracy results (66.35%), whereas GB outperformed when using cross-validation (68.14%). Conversely, the lowest-performing classifiers with respect to accuracy included LGBM (65.38%) with train and test split and AdaBoost (62.87%) with cross-validation. As regards precision, AdaBoost proved to be the best-performing classifier with train and test split and GB with cross-validation (scores of 71.19% and 71.37%, respectively). On the other hand, the lowest-performing classifiers in terms of precision were GB (68.66%) when using train and test split and AdaBoost (67.14%) when using cross-validation (see Figure 3). In conclusion, on the BUPA dataset using a boosting ensemble AdaBoost exhibited good performance with train and test split in terms of accuracy, precision, and roc_auc, whereas GB proved to be the best-performing classifier in recall and f1-score. On the other hand, with cross-validation, GB outperformed in accuracy, precision, recall, and f1-score, whereas LGBM exhibited the best performance in terms of roc_auc.

4.2.2. Classifier Performance on the CPD

We tested the performance of the nine ML classifiers on the top four attributes of the CPD. The selected attributes comprised prothrombin (score = 13.25), albumin (score = 12.19), platelets (score = 11.22), and age (score = 9.05). Table 9, Table 10 and Table 11 summarize the results of the analysis.
As can be observed from Table 9, with non-ensemble learning, LGBM achieved the best results in terms of accuracy (69.05%) with the train and test split technique, whereas GB exhibited the highest accuracy (72.74%) with cross-validation. Conversely, DT was the lowest-performing classifier in accuracy with both train and test split and cross-validation, with scores of 57.14% and 61.98%, respectively. Regarding precision, AdaBoost proved to be the best-performing classifier in train and test split (57.14%), while SVC showed the best results in cross-validation (70.17%). The lowest-performing classifier in terms of precision was DT with both train and test split and cross-validation, with scores of 41.07% and 45.64%, respectively. In conclusion, on the CPD with train and test split, LGBM exhibited good performance in accuracy, f1-score, and roc_auc, whereas AdaBoost showed the best results in precision. Also, DT exhibited the best results in terms of recall. On the other hand, with cross-validation, GB outperformed the other classifiers in accuracy and f1-score. SVC exhibited the best performance in precision, DT showed the highest score in recall, and LR outperformed the other classifiers in terms of roc_auc.
As summarized in Table 10, with the bagging ensemble classifier, we used number of trees = 50 and max_samples = 0.5. With the train and test split strategy, LGBM achieved the highest accuracy (67.46%), whereas AdaBoost exhibited the highest accuracy (73.21%) with cross-validation. Conversely, the lowest-performing classifiers with regards to accuracy were KNN (61.90%) in train and test split and SVC (68.39%) in cross-validation. Regarding precision, LGBM proved to be the best-performing classifier (54.05%) in train and test split, and AdaBoost exhibited the highest score during cross-validation (68.28%). The lowest-performing classifiers in precision included KNN (42.86%) in train and test split and DT (58.06%) in cross-validation. In conclusion, on the CPD with train and test split, LGBM exhibited good performance in accuracy, precision, recall, f1-score, and roc_auc. Conversely, with cross-validation, AdaBoost showed the best results in accuracy, precision, recall, and f1-score, while LR outperformed the other classifiers in roc_auc.
As observed in Table 11, we analyzed the performance of the boosting ensemble using the best arguments, i.e., n_estimators = 500, max_depth = 3, and max_features = auto. With the train and test split technique, AdaBoost achieved the highest accuracy (66.67%), whereas LGBM displayed the highest accuracy score (67.72%) with the cross-validation technique. On the other hand, the lowest-performing classifiers with regard to accuracy included GB (63.49%) with train and test split and AdaBoost (66.51%) with cross-validation. As for precision, AdaBoost proved to be the best-performing classifier in train and test split (52.38%), while LGBM achieved the highest score with cross-validation (53.89%). On the other hand, GB (47.62%) was the lowest-performing classifier in terms of precision with train and test split, and AdaBoost showed the lowest score (50.60%) with cross-validation (see Figure 4). In conclusion, in the boosting ensemble applied on the CPD, AdaBoost exhibited good performance with train and test split across the five metrics. However, with cross-validation, LGBM outperformed the other classifiers in accuracy, precision, recall, and f1-score, whereas GB exhibited the best performance in roc_auc.

4.2.3. Classifier Performance on the HCC Survival Dataset

We tested the performance of the nine ML classifiers on the top four attributes of the HCC Survival dataset. The selected attributes comprised AFP (score = 11.17), hemoglobin (score = 7.11), ALP (score = 6.24), and albumin (score = 6.22). Table 12, Table 13 and Table 14 summarize the results of the analysis.
As summarized in Table 12, with non-ensemble learning, DT and GB achieved the highest accuracy (75.81%) with train and test split, while GB exhibited the highest accuracy (74.60%) with cross-validation. On the other hand, the lowest-performing classifier in terms of accuracy was SVC in both train and test split and cross-validation, with scores of 45.16% and 53.95%, respectively. As for precision, GB and DT proved to be the best-performing classifiers (75.81%) with the train and test split strategy, whereas KNN exhibited the best result in cross-validation (76.00%). SVC in train test split and LR in cross-validation were the lowest-performing classifiers in precision, with values of 43.33% and 41.01%, respectively. In conclusion, on the HCC Survival dataset with train and test split, GB and DT exhibited good performance in accuracy, precision, f1-score, and roc_auc, whereas SVC outperformed the other classifiers in recall with a value of 100%. When using cross-validation, GB outperformed the other classifiers in accuracy, f1-score, and roc_auc. Additionally, KNN exhibited the best performance in precision, and SVC displayed the best score in recall.
With the bagging ensemble classifier, we used number of trees = 50 and max_samples = 0.5. When using the train and test split strategy, AdaBoost, GB, and RF achieved the highest accuracy score (74.19%), whereas RF exhibited the highest accuracy score (77.45%) with cross-validation. On the other hand, SVC proved to be the lowest-performing classifier in accuracy, with values of 45.16% and 49.60% in train and test split and cross-validation, respectively. As regards precision, GB and RF exhibited the best score with train and test split (69.23%), while GB outperformed the other classifiers when using the cross-validation technique (78.45%). The lowest-performing classifier in precision was SVC, with scores of 43.33% in train and test split and 45.26% in cross-validation. In conclusion, on the HCC Survival dataset with train and test split, AdaBoost exhibited good performance in accuracy, f1-score, and roc_auc. GB performed best in accuracy and precision and SVC in recall. Conversely, with cross-validation, AdaBoost displayed the highest recall and f1-scores, RF outperformed in accuracy, and GB exhibited the highest scores in precision and roc_auc.
We analyzed the performance of the ML classifiers (see Table 14) with the boosting ensemble using best arguments n_estimators = 100, max_depth = 3, and max_features = auto. With the train and test split strategy, GB displayed the best accuracy score (75.81%), whereas GB exhibited the highest accuracy (74.60%) with cross-validation. Conversely, the lowest-performing classifiers in accuracy included AdaBoost Classifier (66.13%) with train and test split and LGBM (71.07%) with cross-validation. As regards precision, GB proved to be the best-performing classifier in both train and test split and cross-validation with scores of 68.97% and 74.20%, respectively. The lowest-performing classifiers in precision were AdaBoost (57.58%) with train and test split and LGBM (71.52%) with cross-validation (see Figure 5). In conclusion, on the HCC Survival dataset using a boosting ensemble, GB exhibited the highest performance scores across the five metrics and with the two validation strategies: train and test split and cross-validation.

4.2.4. Classifier Performance on ILPD Dataset

We tested the performance of the nine ML classifiers (AdaBoost, DT, ETT, GB, KNN, LGBM, LR, RG, and SVC) on the top four attributes of the ILPD dataset. The selected attributes comprised Alkphos (score = 14.83), Sgot (score = 13.18), Sgpt (score = 12.86), and age (score = 12.83). Table 15, Table 16 and Table 17 summarize the results of the analysis.
As summarized in Table 15, with non-ensemble learning, LR displayed the best results in accuracy with both train and test split and cross-validation (74.29% and 71.51%, respectively). However, DT proved to be the lowest-performing classifier in accuracy with both train and test split and cross-validation (66.29% and 62.95%, respectively). As far as precision is concerned, LGBM proved to be the best-performing classifier, with scores of 79.85% in train and test split and 77.69% in cross-validation. The lowest-performing classifier in precision was SVC with both train and test split and cross-validation (72.579% and 71.35%, respectively). In conclusion, on the ILPD with train and test split, LR exhibited good performance in accuracy and f1-score, LGBM outperformed the other classifiers in precision and roc_auc, and SVC exhibited the highest score in recall. With cross-validation, LR exhibited the highest score in accuracy and roc_auc, LGBM in precision, and SVC in f1-score and recall.
As summarized in Table 16, with the bagging ensemble classifier, we used number of trees = 100 and max_samples = 0.4. With the train and test split strategy, AdaBoost exhibited the best result in accuracy (76.00%), while LR exhibited the highest accuracy (72.03%) with cross-validation. Conversely, the lowest-performing classifiers in accuracy were DT (70.86%) with train and test split and LGBM (69.12%) with cross-validation. Regarding precision, AdaBoost proved to be the best-performing classifier with both strategies, train and test split and cross-validation, with scores of 78.15% and 76.32%, respectively. In precision, the lowest-performing classifiers were SVC (72.57%) with train and test split and 71.35% with cross-validation. In conclusion, on the ILPD, AdaBoost exhibited good performance in accuracy, precision, and roc_auc with train and test split. LR outperformed the other classifiers in f1-score and SVC in terms of recall. On the other hand, when using cross-validation, LR outperformed in accuracy, AdaBoost in precision, SVC in f1-score and recall, and ETT in roc_auc.
As observed in Table 17, we analyzed the performance of the classifiers with the boosting ensemble using best arguments for ILPD, i.e., n_estimators = 100, max_depth = 7, and max_features = auto. With both train and test split and cross-validation, LGBM exhibited the best result accuracy, with values of 73.14 and 68.77%, respectively. Conversely, GB was the lowest-performing classifier with respect to accuracy, with values of 68.00% during train and test split and 67.57% during cross-validation. As for precision, LGBM proved to be the best-performing classifier with the train and test split technique (79.85%) and GB with cross-validation (75.89%). The lowest-performing classifiers in precision included GB (77.10%) with train and test split and AdaBoost (75.69%) with cross-validation (see Figure 6). In conclusion, on the ILPD using a boosting ensemble, LGBM exhibited the best performance across the five metrics with train and test split. On the other hand, with cross-validation, LGBM outperformed in accuracy, recall, f1-score, and roc_auc, and GB exhibited the best results in precision.

4.3. Most Important Dataset Attributes

This research aims to identify the top four attributes or risk factors for the detection and prevention of liver disease by finding the best precision and accuracy results from the nine ML classifiers in different ensembles. We compared the results obtained from all the non-ensemble and ensemble (bagging and boosting) analyses and found out that the nine ML classifiers performed adequately on all the datasets with the two performance analysis strategies: train and test split and cross-validation (see Figure 7). Similarly, we noted that the bagging ensemble improved the accuracy of ET and AdaBoost on the BUPA and ILPD datasets with train and test split. However, in three of the four datasets (BUPA, CPD, and HCC Survival Dataset), the algorithms performed much better with cross-validation.
Concerning the precision metrics, Figure 8 shows that the classifiers adequately performed for all the top four attribute classifications. Additionally, we found that when working with the BUPA and the HCC Survival datasets, the classifiers performed better in precision when using the bagging ensemble, whereas, in the ILPD, the classifiers performed much better with the boosting ensemble. On the other hand, when dealing with the non-ensemble method on the datasets, we achieved better classifier performance results on the ILPD and the HCC Survival dataset. As for the best-evaluated technique, train-and-test split validation worked best on the BUPA dataset, the CPD, and the ILPD. On the other hand, on the HCC Survival dataset, some algorithms performed better when using train-and-test split with the boosting ensemble, whereas some others worked best when using both the bagging ensemble and the non-ensemble models.
We identified the main attributes for diagnosing liver disease in the four datasets due to the previous analysis (Table 18). On the BUPA dataset, such attributes include gammagt (gamma-glutamyl transpeptidase), sgpt (alanine aminotransferase), sgot (aspartate aminotransferase), and alkphos (alkaline phosphatase). Regarding the best ensemble performance with train and test split, the bagging ensemble with ETT achieved the best results in accuracy (74.04%), precision (74.63%), f1-score (78.74%), and roc_auc with (72.35%). On the other hand, SVC yielded the best recall performance (86.67%). When using k-fold cross-validation, the bagging ensemble with GB exhibited the best performance in accuracy (71.61%) and roc_auc (75.84%), while DT outperformed the other classifiers in precision (73.99%). SVC with no ensemble displayed the best results in recall (84.87%) and f1-score (76.38%).
In the HCC Survival dataset, the main attributes identified included AFP (alpha-fetoprotein), hemoglobin (hemoglobin), ALP (alkaline phosphatase), and albumin (albumin). With train and test split, DT without ensemble achieved the best performance in accuracy (75.81%), precision (68.97%), f1-score (72.33%), and roc_auc (75.96%), whereas SVC exhibited the best results in recall (100.00%). Regarding precision, RF and GB with the bagging ensemble yielded the best performance (69.23%). On the other hand, when using k-fold cross-validation with the bagging ensemble, RF exhibited the highest score in accuracy (77.45%), GB in precision (78.45%), roc-auc (84.98%), and AdaBoost in f1-score (78.59%). SVC exhibited the best performance with no ensemble in recall (99.23%).
In the ILPD dataset, the main attributes identified included Alkphos (alkaline phosphatase), Sgot (aspartate aminotransferase), Sgpt (alanine aminotransferase), and age (age of the patient). Regarding the best ensemble performance with train and test split, the bagging ensemble with AdaBoost achieved the best accuracy (76.00%), while LR displayed the best score in the f1-score (85.03%). With no ensemble, LGBM showed the best performance in precision (79.85%), SVC in recall (100.00%), and LGBM in roc_auc (64.00%). On the other hand, when using k-fold cross-validation with the bagging ensemble, LR showed the best results in accuracy (72.03%) and ETT in roc_auc (73.62%). With no ensemble, LGBM exhibited the best performance in precision (77.69%), while SVC outperformed the other classifiers in recall (100.00%) and f1-score (83.08%)
In the CPD, the main attributes identified included prothrombin (prothrombin time in seconds [s]), albumin (albumin in [gm/dl]), platelets (platelets per cubic [ml/1000]), and age (Age of the patient). The non-ensemble method achieved the best accuracy (69.05%) and precision (57.14%) with AdaBoost, whereas, in recall, DT outperformed the other classifiers (52.27%). Similarly, LGBM exhibited the best score in f1-score (51.85%) and roc_auc (64.11%). When using k-fold cross-validation, the bagging ensemble performed best with AdaBoost in accuracy (73.21%) and f1-score (55.73%) and with LR in roc_auc (75.99%). With no ensemble, SVC showed the best performance in precision (70.17%) and DT in recall (50.57%).
From our previous discussion, we concluded that the nine ML classifiers achieved adequate performance in the classification of the top four dataset attributes and can be successfully used for liver disease prediction.
The implemented algorithms validate the main features considered relevant for the diagnosis of liver diseases according to the literature, as shown in Table 18. Of the variables identified in the BUPA, HCC Survival, ILPD, CPD, alanine aminotransferase is strongly related to metabolic syndrome in NAFLD [57], alkaline phosphatase is an important serum analyte that can be elevated in liver disease [58], and aspartate aminotransferase is a significant marker for alcoholic liver disease [6]. Testing of alanine aminotransferase [59] and aspartate aminotransferase is important in children with symptoms of possible liver disease, such as jaundice, dark urine, nausea, vomiting, or belly pain. Alkaline phosphatase in older female patients with isolated elevated alkaline phosphatase and risk factors for NAFLD should be evaluated for evidence of significant steatohepatitis [60]. Two other important risk factors are prothrombin [61] (identified in the CPD) and albumin [62] (identified in the HCC Survival dataset and the CPD). Prothrombin levels in liver disease are important indicators of liver function and pathology [63], and albumin binding function is a novel biomarker for early liver damage in NAFLD [64]. Both prothrombin and albumin must be primarily monitored when suspecting the presence of NAFLD. Platelets are another critical risk factor for liver disease [62,65]. Platelet counts can be performed to monitor or diagnose liver diseases or to look for the cause of too much bleeding or clotting. As for patient age, it was identified as a top risk factor for liver disease in the CPD and ILPD, which is consistent with the fact that cirrhosis can start at an early age [62]. In the BUPA dataset, gamma-glutamyl transpeptidase surfaced as a major risk factor for liver disease. Serum gamma GT activity is a valuable diagnostic tool for liver disease in children [66]. A GGT test is often used to diagnose liver disease and determine whether liver damage is due to liver disease or bone disease. The test is also used to check for blocked bile ducts and detect or monitor alcohol use disorder. When monitoring results are higher than normal, it may be a sign of liver damage caused by hepatitis, cirrhosis, alcohol use disorder, pancreatitis, diabetes, congestive heart failure, or a side effect of a medication. Finally, Alpha-fetoprotein (AFP) is a biomarker that can be used in the diagnosis and monitoring of liver diseases, particularly hepatocellular carcinoma (HCC) [67]. Accordingly, the algorithms implemented in the different ensembles identified the markers considered the main risk factors for liver disease in the datasets evaluated.

5. Conclusions and Future Directions

In this research, we aimed to compare the performance of nine machine learning algorithms (MLAs) across four datasets in predicting liver disease based on the top four attribute classifications using different ensembles with train-test split strategy and k-fold cross-validation methods. Our results indicated that alanine aminotransferase, aspartate aminotransferase, alkaline phosphatase, and albumin were identified as significant risk factors for liver diseases such as hepatitis, cirrhosis, alcohol use disorder, pancreatitis, diabetes, and mainly the aim of our study NAFLD. Additionally, gamma-glutamyl transpeptidase, hemoglobin, age, prothrombin, alpha-fetoprotein, and platelets contributed significantly toward detection. Our main findings revealed that the analyzed MLAs exhibited the best performance in the BUPA dataset and the CPD across the five performance metrics with non-ensemble learning. However, in the bagging ensemble, only ETT and LR exhibited high accuracy. The studied algorithms were classified appropriately to predict if a person were to have non-alcoholic fatty liver disease and exhibited good accuracy and precision across the four datasets. As for which classifier exhibited the highest accuracy with train and test split, AdaBoost outperformed in the ILPD using a bagging ensemble. As for cross-validation, the HCC Survival dataset obtained the best performance with RF using a bagging ensemble. The main contribution of this research is to validate the top risk factors for NAFLD: alanine aminotransferase, alkaline phosphatase, aspartate aminotransferase, alpha-fetoprotein, and gamma-glutamyl transpeptidase. Having proper medical follow-up on these attributes can contribute to the early diagnosis and treatment of non-alcoholic fatty liver disease. As for future proposals, research into medical databases for other common ailments, such as colon and breast cancer, is suggested. In addition, the risk factors detected in this research can be prioritized in mobile applications aimed at diagnosing and monitoring liver diseases. Finally, creating a database with the main attributes of liver diseases (risk factors) from various sources, such as clinical datasets, portable devices, mobile applications, and medical records, would be interesting. This goal could be achieved using big data methodologies combined with machine learning, which will play a crucial role in improving our standard of living.

Author Contributions

Conceptualization, G.A.-H.; methodology, L.R.G.-N.; software L.R.G.-N. and J.L.S.-C.; validation, L.R.G.-N.; formal analysis, G.P.-A.; investigation, J.L.S.-C., L.R.G.-N. and G.A.-H.; data curation, L.R.G.-N.; visualization, G.P.-A.; supervision, G.A.-H.; project administration, J.L.S.-C.; funding acquisition, J.L.S.-C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Council for Scientific Research and Technological Development in Veracruz (COVEICYDET) through the project Prevention and Early Detection of Cardiovascular Diseases (Arrhythmias and Tachycardias) using Machine Learning Techniques, Big Data, and the Internet of Things, grant number 12 1806.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Publicly available datasets were analyzed in this study. These data can be found here: Liver Disorders Data Set (https://archive.ics.uci.edu/ml/datasets/liver+disorder) (accessed on 22 May 2022); HCC Survival Dataset (https://archive.ics.uci.edu/ml/datasets/HCC+Surviva) (accessed on 22 May 2022); ILPD Dataset (https://archive.ics.uci.edu/ml/datasets/ILPD+%28Indian+Liver+Patient+Dataset%29) (accessed on 22 May 2022); Cirrhosis Prediction Dataset (https://www.kaggle.com/datasets/fedesoriano/cirrhosis-prediction-dataset) (accessed on 22 May 2022).

Acknowledgments

L.G.N. and G.P.A. thank CONACYT for supporting their postdoctoral stays at Tecnológico Nacional de México. J.L.S.C. thanks CONACYT for the research position granted under the Cátedras-CONACYT program.

Conflicts of Interest

The authors declare no potential conflict of interest with respect to the publication of this research.

References

  1. INEGI. INEGI Instituto Nacional de Estadística, Geografía e Informática. Características de las Defunciones Registradas en México Durante Enero a Agosto de 2020. INEGI. 28 June 2022. Available online: https://www.inegi.org.mx/contenidos/saladeprensa/boletines/2021/EstSociodemo/DefuncionesRegistradas2020_Pnles.pdf (accessed on 27 June 2023).
  2. Lee, H.W.; Sung, J.J.Y.; Ahn, S.H. Artificial intelligence in liver disease. J. Gastroenterol. Hepatol. 2021, 36, 539–542. [Google Scholar] [CrossRef] [PubMed]
  3. Goldman, O.; Ben-Assuli, O.; Rogowski, O.; Zeltser, D.; Shapira, I.; Berliner, S.; Zelber-Sagi, S.; Shenhar-Tsarfaty, S. Non-alcoholic Fatty Liver and Liver Fibrosis Predictive Analytics: Risk Prediction and Machine Learning Techniques for Improved Preventive Medicine. J. Med. Syst. 2021, 45, 22. [Google Scholar] [CrossRef] [PubMed]
  4. Kwak, M.S.; Kim, D. Non-alcoholic fatty liver disease and lifestyle modifications, focusing on physical activity. Korean J. Intern. Med. 2018, 33, 64–74. [Google Scholar] [CrossRef] [Green Version]
  5. Ahmed, M.H. Biochemical Markers the Road Map for the Diagnosis of Nonalcoholic Fatty Liver Disease. Am. J. Clin. Pathol. 2007, 127, 20–22. [Google Scholar] [CrossRef]
  6. Aravind, G.N.; Abhilash, K.; Syed, U.F. A study of alanine aminotransferase—Aspartate aminotransferase as a marker of advanced alcoholic liver disease. Int. J. Adv. Med. 2020, 7, 551–553. [Google Scholar] [CrossRef] [Green Version]
  7. Pancreas, J.J.; Das, R.N.; Mukherjee, S.; Sharma, I. Alkaline Phosphatase Determinants of Liver Patients. 2018. Available online: http://pancreas.imedpub.com/ (accessed on 27 June 2023).
  8. Lin, E.; Lin, C.H.; Lane, H.Y. Applying a bagging ensemble machine learning approach to predict functional outcome of schizophrenia with clinical symptoms and cognitive functions. Sci. Rep. 2021, 11, 6922. [Google Scholar] [CrossRef] [PubMed]
  9. Ponnaganti, N.D.; Anitha, R. A Novel Ensemble Bagging Classification Method for Breast Cancer Classification Using Machine Learning Techniques. Trait. Signal 2022, 39, 229–237. [Google Scholar] [CrossRef]
  10. Chicco, D.; Jurman, G. An ensemble learning approach for enhanced classification of patients with hepatitis and cirrhosis. IEEE Access 2021, 9, 24485–24498. [Google Scholar] [CrossRef]
  11. Anisha, C.D.; Saranya, K.G. Early diagnosis of stroke disorder using homogenous logistic regression ensemble classifier. Int. J. Nonlinear Anal. Appl. 2021, 12, 1649–1654. [Google Scholar] [CrossRef]
  12. Devi, M.S.; Swathi, P.; Upadhyay, S.S.; Sah, N.K.; Budhia, A.; Srivastava, S.; Rohella, M. Feature Predominance Ensemble Inquisition towards Liver Disease Prediction using Machine Learning. In Proceedings of the International Conference on Innovative Computing & Communication (ICICC), Delhi, India, 20–21 February 2021. [Google Scholar] [CrossRef]
  13. Lin, E.; Lin, C.H.; Lane, H.Y. A bagging ensemble machine learning framework to predict overall cognitive function of schizophrenia patients with cognitive domains and tests. Asian J. Psychiatr. 2022, 69, 103008. [Google Scholar] [CrossRef]
  14. Ejiofor, C.I.; Ochei, L.C. Application of Heterogenous Bagging Ensemble Model for predicting Breast Cancer. J. Comput. Sci. Its Appl. 2021, 28. [Google Scholar] [CrossRef]
  15. Rahman, F.; Mahmood, M.A. A Dynamic Approach to Identify the Most Significant Biomarkers for Heart Disease Risk Prediction utilizing Machine Learning Techniques. Available online: https://www.researchgate.net/publication/357458668 (accessed on 28 April 2023).
  16. Thomgkam, J.; Sukmak, V.; Klangnok, P. Application of Machine Learning Techniques to Predict Breast Cancer Survival. In Lecture Notes in Computer Science, Proceedings of the 14th Multi-disciplinary International Conference on Artificial Intelligence (MIWAI 2021), Online, 2–3 July 2021; Springer: Cham, Switzerland, 2021; Volume 12832, pp. 141–151. [Google Scholar] [CrossRef]
  17. Yadav, S.; Singh, M.K. Hybrid Machine Learning Classifier and Ensemble Techniques to Detect Parkinson’s Disease Patients. SN Comput. Sci. 2021, 2, 189. [Google Scholar] [CrossRef]
  18. Buyrukoglu, S. Improvement of Machine Learning Models Performances based on Ensemble Learning for the detection of Alzheimer Disease. In Proceedings of the 6th International Conference on Computer Science and Engineering (UBMK), Ankara, Turkey, 15–17 September 2021; pp. 102–106. [Google Scholar] [CrossRef]
  19. Singh, A.; Mehta, J.C.; Anand, D.; Nath, P.; Pandey, B.; Khamparia, A. An intelligent hybrid approach for hepatitis disease diagnosis: Combining enhanced k-means clustering and improved ensemble learning. Expert Syst 2021, 38, e12526. [Google Scholar] [CrossRef]
  20. Sarvestany, S.S.; Kwong, J.C.; Azhie, A.; Dong, V.; Cerocchi, O.; Ali, A.F.; Karnam, R.S.; Kuriry, H.; Shengir, M.; Candido, E.; et al. Development and validation of an ensemble machine learning framework for detection of all-cause advanced hepatic fibrosis: A retrospective cohort study. Lancet Digit Health 2022, 4, e188–e199. [Google Scholar] [CrossRef] [PubMed]
  21. Dutta, K.; Chandra, S.; Gourisaria, M.K. Early-Stage Detection of Liver Disease Through Machine Learning Algorithms. Lect. Notes Netw. Syst. 2022, 318, 155–166. [Google Scholar] [CrossRef]
  22. Verma, A.; Mehta, S. A comparative study of ensemble learning methods for classification in bioinformatics. In Proceedings of the 7th International Conference on Cloud Computing, Data Science & Engineering—Confluence, Noida, India, 12–13 January 2017; pp. 155–158. [Google Scholar] [CrossRef]
  23. Meng, L.; Treem, W.; Heap, G.; Chen, J. Predicting Clinical Outcomes of Alpha-1 Antitrypsin Deciency-Associated Liver Disease Using a Stacking Ensemble Machine Learning Model Based on UK Biobank Data. 2022; preprint. [Google Scholar] [CrossRef]
  24. Al Telaq, B.H.; Hewahi, N. Prediction of Liver Disease using Machine Learning Models with PCA. In Proceedings of the 2021 International Conference on Data Analytics for Business and Industry (ICDABI), Sakheer, Bahrain, 25–26 October 2021; pp. 250–254. [Google Scholar] [CrossRef]
  25. Gupta, S.; Gupta, M.K. Computational Prediction of Cervical Cancer Diagnosis Using Ensemble-Based Classification Algorithm. Comput. J. 2021, 65, 1527–1539. [Google Scholar] [CrossRef]
  26. Pouriyeh, S.; Vahid, S.; Sannino, G.; De Pietro, G.; Arabnia, H.; Gutierrez, J. A comprehensive investigation and comparison of Machine Learning Techniques in the domain of heart disease. In Proceedings of the 2017 IEEE Symposium on Computers and Communications (ISCC), Heraklion, Greece, 3–6 July 2017; pp. 204–207. [Google Scholar] [CrossRef]
  27. Kabir, M.F.; Ludwig, S.A. Enhancing the Performance of Classification Using Super Learning. Data-Enabled Discov. Appl. 2019, 3, 5. [Google Scholar] [CrossRef]
  28. Doğaner, A.; Çolak, C.; Küçükdurmaz, F.; Ölmez, C. Prediction of Renal Cell Carcinoma Based on Ensemble Learning Methods. Middle Black Sea J. Health Sci. 2021, 7, 104–114. [Google Scholar] [CrossRef]
  29. Hakim, M.A.; Jahan, N.; Zerin, Z.A.; Farha, A.B. Performance Evaluation and Comparison of Ensemble Based Bagging and Boosting Machine Learning Methods for Automated Early Prediction of Myocardial Infarction. In Proceedings of the 2021 12th International Conference on Computing Communication and Networking Technologies (ICCCNT), Kharagpur, India, 6–8 July 2021. [Google Scholar] [CrossRef]
  30. Yadav, D.C.; Pal, S. An Experimental Study of Diversity of Diabetes Disease Features by Bagging and Boosting Ensemble Method with Rule Based Machine Learning Classifier Algorithms. SN Comput. Sci. 2021, 2, 50. [Google Scholar] [CrossRef]
  31. Gao, X.Y.; Ali, A.A.; Hassan, H.S.; Anwar, E.M. Improving the Accuracy for Analyzing Heart Diseases Prediction Based on the Ensemble Method. Complexity 2021, 2021, 6663455. [Google Scholar] [CrossRef]
  32. Taser, P.Y. Application of Bagging and Boosting Approaches Using Decision Tree-Based Algorithms in Diabetes Risk Prediction. Proceedings 2021, 74, 6. [Google Scholar] [CrossRef]
  33. Murthy, H.S.N.; Manjunatha, M.N. Early Prognosis of Coronary Heart Disease using Ensemble Classifiers: A Comparative Analysis. Volatiles Essent. Oils 2021, 8, 2136–2142. [Google Scholar]
  34. Fraiwan, L.; Hassanin, O. Computer-aided identification of degenerative neuromuscular diseases based on gait dynamics and ensemble decision tree classifiers. PLoS ONE 2021, 16, e0252380. [Google Scholar] [CrossRef]
  35. Dhilsath, F.M.; Samuel, S.J. Hyperparameter Tuning of Ensemble Classifiers Using Grid Search and Random Search for Prediction of Heart Disease. Comput. Intell. Healthc. Inform. 2021, 139–158. [Google Scholar] [CrossRef]
  36. Khanam, F.; Mondal, M.R.H. Ensemble Machine Learning Algorithms for the Diagnosis of Cervical Cancer. In Proceedings of the 2021 International Conference on Science and Contemporary Technologies, ICSCT, Dhaka, Bangladesh, 5–7 August 2021. [Google Scholar] [CrossRef]
  37. Bang, C.S.; Ahn, J.Y.; Kim, J.H.; Kim, Y.I.; Choi, I.J.; Shin, W.G. Establishing Machine Learning Models to Predict Curative Resection in Early Gastric Cancer with Undifferentiated Histology: Development and Usability Study. J. Med. Internet Res. 2021, 23, e25053. [Google Scholar] [CrossRef]
  38. UCI Machine Learning Repository: Liver Disorders Data Set. Available online: https://archive.ics.uci.edu/ml/datasets/liver+disorders (accessed on 22 May 2023).
  39. Santos, M.S.; Abreu, P.H.; García-Laencina, P.J.; Simão, A.; Carvalho, A. A new cluster-based oversampling method for improving survival prediction of hepatocellular carcinoma patients. J. Biomed. Inf. 2015, 58, 49–59. [Google Scholar] [CrossRef] [Green Version]
  40. UCI Machine Learning Repository: HCC Survival Data Set. Available online: https://archive.ics.uci.edu/ml/datasets/HCC+Survival# (accessed on 22 May 2023).
  41. UCI Machine Learning Repository: ILPD (Indian Liver Patient Dataset) Data Set. Available online: https://archive.ics.uci.edu/ml/datasets/ILPD+%28Indian+Liver+Patient+Dataset%29 (accessed on 22 May 2023).
  42. Cirrhosis Prediction Dataset. Kaggle. Available online: https://www.kaggle.com/datasets/fedesoriano/cirrhosis-prediction-dataset (accessed on 22 May 2023).
  43. Iyer, R.; Hosmer, D.W.; Lemeshow, S. Applied Logistic Regression. J. R. Stat. Soc. Ser. D 1991, 40, 458. [Google Scholar] [CrossRef]
  44. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef] [Green Version]
  45. Sharma, J.; Giri, C.; Granmo, O.-C.; Goodwin, M. Multi-layer intrusion detection system with ExtraTrees feature selection, extreme learning machine ensemble, and softmax aggregation. EURASIP J. Inf. Secur. 2019, 2019, 15. [Google Scholar] [CrossRef] [Green Version]
  46. Lecun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
  47. Sorich, M.J.; Miners, J.O.; McKinnon, R.A.; Winkler, D.A.; Burden, F.R.; Smith, P.A. Comparison of Linear and Nonlinear Classification Algorithms for the Prediction of Drug and Chemical Metabolism by Human UDP-Glucuronosyltransferase Isoforms. J. Chem. Inf. Comput. Sci. 2003, 43, 2019–2024. [Google Scholar] [CrossRef] [PubMed]
  48. Ramana, B.V.; Babu, M.S.P.; Venkateswarlu, N.B. A Critical Study of Selected Classification Algorithms for Liver Disease Diagnosis. Int. J. Database Manag. Syst. 2011, 3, 101–114. [Google Scholar] [CrossRef]
  49. Biau, G.; Cadre, B.; Rouvière, L. Accelerated gradient boosting. Mach. Learn. 2019, 108, 971–992. [Google Scholar] [CrossRef] [Green Version]
  50. Ke, G.; Meng, Q.; Finley, T.; Wang, T.; Chen, W.; Ma, W.; Ye, Q.; Liu, T.Y. LightGBM: A Highly Efficient Gradient Boosting Decision Tree. 2017. Available online: https://github.com/Microsoft/LightGBM (accessed on 31 May 2021).
  51. Zhu, J.; Zou, H.; Rosset, S.; Hastie, T. Multi-class AdaBoost. Stat. Its Interface 2009, 2, 349–360. [Google Scholar]
  52. Dietterich, T.G. An Experimental Comparison of Three Methods for Constructing Ensembles of Decision Trees: Bagging, Boosting, and Randomization. Mach. Learn. 2000, 40, 139–157. [Google Scholar] [CrossRef]
  53. Zhang, W.; Zeng, F.; Wu, X.; Zhang, X.; Jiang, R. A comparative study of ensemble learning approaches in the classification of breast cancer metastasis. In Proceedings of the 2009 International Joint Conference on Bioinformatics, Systems Biology and Intelligent Computing, IJCBS, Shanghai, China, 3–5 August 2009; pp. 242–245. [Google Scholar] [CrossRef]
  54. Guarneros-Nolasco, L.R.; Cruz-Ramos, N.A.; Alor-Hernández, G.; Rodríguez-Mazahua, L.; Sánchez-Cervantes, J.L. Identifying the main risk factors for cardiovascular diseases prediction using machine learning algorithms. Mathematics 2021, 9, 2537. [Google Scholar] [CrossRef]
  55. Van Rossum, G.; Drake, F.L. Python 3 Reference Manual; CreateSpace: Scotts Valley, CA, USA, 2009. [Google Scholar]
  56. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Scikit-learn: Machine Learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  57. Chen, Z.; Chen, L.; Dai, H.; Chen, J.; Fang, L. Relationship between alanine aminotransferase levels and metabolic syndrome in nonalcoholic fatty liver disease. J. Zhejiang Univ. Sci. B 2008, 9, 616–622. [Google Scholar] [CrossRef]
  58. Grytczuk, A.; Gruszewska, E.; Panasiuk, A.; Cylwik, B.; Chrostek, L. Serum Profile of Lactate Dehydrogenase (LDH) and Alkaline Phosphatase (ALP) in Alcoholic Liver Diseases. 2020; preprint. [Google Scholar] [CrossRef]
  59. Arsik, I.; Frediani, J.K.; Frezza, D.; Chen, W.; Ayer, T.; Keskinocak, P.; Jin, R.; Konomi, J.V.; Barlow, S.E.; Xanthakos, S.A.; et al. Alanine Aminotransferase as a Monitoring Biomarker in Children with Nonalcoholic Fatty Liver Disease: A Secondary Analysis Using TONIC Trial Data. Children 2018, 5, 64. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  60. Pantsari, M.W.; Harrison, S.A. Nonalcoholic fatty liver disease presenting with an isolated elevated alkaline phosphatase. J. Clin. Gastroenterol. 2006, 40, 633–635. [Google Scholar] [CrossRef] [PubMed]
  61. Tripodi, A.; Caldwell, S.H.; Hoffman, M.; Trotter, J.F.; Sanyal, A.J. Review article: The prothrombin time test as a measure of bleeding risk and prognosis in liver disease. Aliment Pharmacol. Ther. 2007, 26, 141–148. [Google Scholar] [CrossRef] [PubMed]
  62. Angulo, P.; Hui, J.M.; Marchesini, G.; Bugianesi, E.; George, J.; Farrell, G.C.; Enders, F.; Saksena, S.; Burt, A.D.; Bida, J.P.; et al. The NAFLD fibrosis score: A noninvasive system that identifies liver fibrosis in patients with NAFLD. Hepatology 2007, 45, 846–854. [Google Scholar] [CrossRef] [PubMed]
  63. Stancu, G.; Iliescu, E.L. The Influence of Liver Transplant on Serum Cholinesterase Levels: A Case Report. Cureus 2023, 15, e33761. [Google Scholar] [CrossRef]
  64. Sun, L.; Wang, Q.; Liu, M.; Xu, G.; Yin, H.; Wang, D.; Xie, F.; Jin, B.; Jin, Y.; Yang, H.; et al. Albumin binding function is a novel biomarker for early liver damage and disease progression in non-alcoholic fatty liver disease. Endocrine 2020, 69, 294–302. [Google Scholar] [CrossRef]
  65. Enomoto, H.; Bando, Y.; Nakamura, H.; Nishiguchi, S.; Koga, M. Liver fibrosis markers of nonalcoholic steatohepatitis. World J. Gastroenterol. 2015, 21, 7427–7435. [Google Scholar] [CrossRef]
  66. Maggiore, G.; Bernard, O.; Hadchouel, M.; Lemonnier, A.; Alagille, D. Diagnostic value of serum gamma-glutamyl transpeptidase activity in liver diseases in children. J. Pediatr. Gastroenterol. Nutr. 1991, 12, 21–26. [Google Scholar] [CrossRef]
  67. Luo, X.; Cui, H.; Cai, L.; Zhu, W.; Yang, W.-C.; Patrick, M.; Zhu, S.; Huang, J.; Yao, X.; Yao, Y.; et al. Selection of a Clinical Lead TCR Targeting Alpha-Fetoprotein-Positive Liver Cancer Based on a Balance of Risk and Benefit. Front. Immunol. 2020, 11, 623. [Google Scholar] [CrossRef]
Figure 1. Methodology for evaluating liver disease datasets.
Figure 1. Methodology for evaluating liver disease datasets.
Mathematics 11 03026 g001
Figure 3. ML classifier performance in accuracy (a) and precision (b): train and test split vs. K-Fold cross-validation—BUPA dataset.
Figure 3. ML classifier performance in accuracy (a) and precision (b): train and test split vs. K-Fold cross-validation—BUPA dataset.
Mathematics 11 03026 g003
Figure 4. ML classifier performance in accuracy (a) and precision (b): train test split vs. K-fold cross-validation—CPD.
Figure 4. ML classifier performance in accuracy (a) and precision (b): train test split vs. K-fold cross-validation—CPD.
Mathematics 11 03026 g004
Figure 5. ML classifier performance in accuracy (a) and precision (b): train test split vs. K-fold cross-validation—HCC Survival dataset.
Figure 5. ML classifier performance in accuracy (a) and precision (b): train test split vs. K-fold cross-validation—HCC Survival dataset.
Mathematics 11 03026 g005
Figure 6. ML classifier performance in accuracy (a) and precision (b): train test split vs. K-fold cross-validation—ILPD.
Figure 6. ML classifier performance in accuracy (a) and precision (b): train test split vs. K-fold cross-validation—ILPD.
Mathematics 11 03026 g006
Figure 7. ML classifier accuracy performance top four attributes of liver disease datasets: train and test split vs. K-fold cross-validation.
Figure 7. ML classifier accuracy performance top four attributes of liver disease datasets: train and test split vs. K-fold cross-validation.
Mathematics 11 03026 g007
Figure 8. ML classifier precision performance on top four attributes of liver disease datasets: train and test split vs. K-fold cross-validation.
Figure 8. ML classifier precision performance on top four attributes of liver disease datasets: train and test split vs. K-fold cross-validation.
Mathematics 11 03026 g008
Table 1. Characteristics of liver disease datasets.
Table 1. Characteristics of liver disease datasets.
DatasetNumber of
Attributes
Number of ClassesNumber of
Records
Prediction/Diagnosis
BUPA Liver
Disorders
61345Prediction/Diagnosis
HCC Survival491165Prediction
ILPD91313Diagnosis
CPD191424Prediction
Table 2. Description of the Attributes of the BUPA Liver Disorders Dataset.
Table 2. Description of the Attributes of the BUPA Liver Disorders Dataset.
Attribute NameAttribute Description
McvMean corpuscular volume
AlkphosAlkaline phosphatase
SgptAlanine aminotransferase
SgotAspartate aminotransferase
GammagtGamma-glutamyl transpeptidase
DrinksNumber of half-pint equivalents of alcoholic beverages drunk per day
SelectorField created by BUPA researchers to split the data into trains/test sets
Table 3. Description of the Attributes of the HCC Survival Dataset.
Table 3. Description of the Attributes of the HCC Survival Dataset.
Attribute NameAttribute Description
GenderGender of the patient
SymptomsSymptoms
AlcoholAlcohol
HBsAgHepatitis B Surface Antigen
HBeAgHepatitis B e Antigen
HBcAbHepatitis B Core Antibody
HCVAbHepatitis C Virus Antibody
CirrhosisCirrhosis
Endemic countriesEndemic countries
SmokingSmoking
DiabetesDiabetes
ObesityObesity
HemochromatosisHemochromatosis
AHTArterial Hypertension
CRIChronic Renal Insufficiency
HIVHuman Immunodeficiency Virus
NASHNonalcoholic Steatohepatitis
Esophageal varicesEsophageal varices
SplenomegalySplenomegaly
Portal hypertensionPortal hypertension
Portal vein thrombosisPortal vein thrombosis
Liver metastasisLiver metastasis
Radiological hallmarkRadiological hallmark
Age at diagnosisAge at diagnosis
Grams/dayGrams of Alcohol per day
Packs/yearPacks of cigarettes per day
Performance statusPerformance status
EncephalopathyEncephalopathy
AscitesAscites degree
INRInternational Normalized Ratio
AFPAlpha-Fetoprotein (ng/mL)
HemoglobinHemoglobin (g/gL)
MCVMean Corpuscular Volume (fl)
LeukocytesLeukocytes (G/L)
PlateletsPlatelets (G/L)
AlbuminAlbumin (mg/dL)
Total BilTotal bilirubin (mg/dL)
ALTAlanine Transaminase (U/L)
ASTAspartate Transaminase (U/L)
GGTGamma Glutamyl Transferase (U/L)
ALPAlkaline phosphatase (U/L)
TPTotal proteins (g/dL)
CreatinineCreatinine (mg/dL)
Number of nodulesNumber of nodules
Major dimensionMajor dimension of nodule
Dir. BilDirect bilirubin (mg/dL)
IronIron (mcg/dL)
SatOxygen saturation (%)
FerritinFerritin (ng/mL)
Table 4. Description of the Attributes of the ILPD.
Table 4. Description of the Attributes of the ILPD.
Attribute NameAttribute Description
AgeAge of the patient
GenderGender of the patient
TBTotal Bilirubin
DBDirect Bilirubin
AlkphosAlkaline Phosphatase
SgptAlanine Aminotransferase
SgotAspartate Aminotransferase
TPTotal Proteins
ALBAlbumin
A/GAlbumin Ratio and Globulin Ratio
SelectorField used to split the data into two sets (labeled by the experts)
Table 5. Description of the Attributes of the CPD.
Table 5. Description of the Attributes of the CPD.
Attribute NameAttribute Description
IDUnique Identifier
N_DaysNumber of days between registration and the earlier of death, transplantation,
or study analysis time in July 1986
StatusStatus of the patient: C (censored), CL (censored due to liver tx), or D (death)
DrugType of drug D-penicillamine or placebo
AgePatient age in days
SexM(male) or F (female)
AscitesPresence of ascites: N (No) or Y (Yes)
HepatomegalyPresence of hepatomegaly: N (No) or Y (Yes)
SpidersPresence of spiders: N (No) or Y (Yes)
EdemaPresence of edema: N (no edema and no diuretic therapy for edema),
S (edema present without diuretics, or edema resolved by diuretics), or
Y (edema despite diuretic therapy)
BilirubinSerum bilirubin in [mg/dL]
CholesterolSerum cholesterol in [mg/dL]
AlbuminAlbumin in [gm/dL]
CopperUrine copper in [ug/day]
Alk_PhosAlkaline phosphatase in [U/L]
SGOTSGOT in [U/mL]
TriglyceridesTriglycerides in [mg/dL]
PlateletsPlatelets per cubic [mL/1000]
ProthrombinProthrombin time in seconds [s]
StageHistologic stage of disease (1, 2, 3, or 4)
Table 6. Non-ensemble learning performance analysis of classifiers on top four attributes—BUPA dataset.
Table 6. Non-ensemble learning performance analysis of classifiers on top four attributes—BUPA dataset.
EnsembleTechniquePredictive ModelPerformance Evaluation Metrics
% Accuracy% Precision% Recall% f1-Score% roc_auc
Non-ensemble
learning
Train and
test split
AdaBoost66.3569.2375.0072.0064.77
DT65.3870.0070.0070.0064.55
ETT72.1273.8580.0076.8070.68
GB68.2770.1578.3374.0266.44
KNN66.3574.5163.3368.4766.89
LGBM65.3870.0070.0070.0064.55
LR73.0874.2481.6777.7871.52
RF70.1972.3178.3375.2068.71
SVC70.1969.8685.0076.6967.50
Cross-validationAdaBoost66.9569.7875.9772.2470.40
DT58.8764.5365.8864.0957.44
ETT68.0970.2577.7073.4871.96
GB66.6969.3276.2071.9373.02
KNN65.1970.4567.5068.4467.42
LGBM70.4573.5377.9475.1173.00
LR68.6670.1881.4274.5971.59
RF69.6171.5079.6674.9271.41
SVC71.6071.2186.4977.6274.58
The best-performing results for the critical metrics calculated by each algorithm are shown in bold.
Table 7. Bagging classifier performance analysis using nine base learners on the BUPA dataset (top four attributes).
Table 7. Bagging classifier performance analysis using nine base learners on the BUPA dataset (top four attributes).
EnsembleTechniqueBase EstimatorPerformance Evaluation Metrics
% Accuracy% Precision% Recall% f1-Score% roc_auc
Bagging
ensemble
Train and test splitAdaBoost68.2769.5780.0074.4266.14
DT71.1572.7380.0076.1969.55
ETT74.0474.6383.3378.7472.35
GB67.3169.7076.6773.0265.61
KNN70.1970.4283.3376.3467.80
LGBM65.3868.1875.0071.4363.64
LR73.0874.2481.6777.7871.52
RF68.2769.0181.6774.8165.83
SVC71.1570.2786.6777.6168.33
Cross-validationAdaBoost68.3970.3376.4670.9569.62
DT69.0173.9976.6774.1272.84
ETT71.2970.6783.0975.9173.88
GB71.6171.6980.9374.2675.84
KNN67.5067.9581.6974.1170.22
LGBM69.8673.1779.8374.7475.28
LR69.8170.2981.0774.5271.73
RF69.2871.8381.7176.0374.51
SVC69.5569.1384.8776.3875.18
The best-performing results for the critical metrics calculated by each algorithm are shown in bold.
Table 8. Boosting ensemble learning performance analysis of classifiers on the BUPA dataset (top four attributes).
Table 8. Boosting ensemble learning performance analysis of classifiers on the BUPA dataset (top four attributes).
EnsembleTechniquePredictive ModelPerformance Evaluation Metrics
% Accuracy% Precision% Recall% f1-Score% roc_auc
Boosting ensembleTrain and test splitAdaBoost66.3571.1970.0070.5965.68
GB66.3568.6676.6772.4464.47
LGBM65.3870.0070.0070.0064.55
Cross-validationAdaBoost62.8767.1469.8567.9968.76
GB68.1471.3775.3572.6069.05
LGBM66.9770.5474.9972.1571.33
The best-performing results for the critical metrics calculated by each algorithm are shown in bold.
Table 9. Non-ensemble learning performance analysis of classifiers on the CPD (top four attributes).
Table 9. Non-ensemble learning performance analysis of classifiers on the CPD (top four attributes).
EnsembleTechniquePredictive ModelPerformance Evaluation Metrics
% Accuracy% Precision% Recall% f1-Score% roc_auc
Non-ensemble
learning
Train and
test split
AdaBoost69.0557.1445.4550.6363.58
DT57.1441.0752.2746.0056.01
ETT65.0850.0040.9145.0059.48
GB66.6752.7843.1847.5061.23
KNN62.7044.0025.0031.8853.96
LGBM69.0556.7647.7351.8564.11
LR65.8751.4340.9145.5760.09
RF66.6752.7843.1847.5061.23
SVC64.2947.3720.4528.5754.13
Cross-validationAdaBoost71.7960.0448.5453.0668.53
DT61.9845.6450.5747.1759.29
ETT68.2055.0144.4848.5772.01
GB72.7462.8449.9054.8371.64
KNN67.6954.4228.7137.1062.58
LGBM69.8658.3749.4452.3668.32
LR72.2566.0041.7550.0475.84
RF70.1059.5146.6551.6072.01
SVC69.3670.1722.4332.0465.46
The best-performing results for the critical metrics calculated by each algorithm are shown in bold.
Table 10. Bagging classifier performance using nine base learners on the CPD (top four attributes).
Table 10. Bagging classifier performance using nine base learners on the CPD (top four attributes).
EnsembleTechniqueBase EstimatorPerformance Evaluation Metrics
% Accuracy% Precision% Recall% f1-Score% roc_auc
Bagging ensembleTrain and test splitAdaBoost66.6752.7843.1847.5061.23
DT66.6752.9440.9146.1560.70
ETT65.8751.6136.3642.6759.04
GB65.8751.5238.6444.1659.56
KNN61.9042.8627.2733.3353.88
LGBM67.4654.0545.4549.3862.36
LR65.8751.4340.9145.5760.09
RF66.6752.9440.9146.1560.70
SVC65.0850.0020.4529.0354.74
Cross-validationAdaBoost73.2168.2850.3955.7372.68
DT69.1558.0644.3451.5572.80
ETT71.0565.2343.3353.1274.97
GB71.5466.5649.3855.5174.62
KNN69.3759.0131.1341.8265.58
LGBM70.8165.6749.6254.1072.20
LR72.0167.2341.7549.0675.99
RF72.0266.3743.0954.1875.15
SVC68.3960.3316.2429.1565.09
The best-performing results for the critical metrics calculated by each algorithm are shown in bold.
Table 11. Boosting ensemble learning performance analysis of classifiers on the CPD (top four attributes).
Table 11. Boosting ensemble learning performance analysis of classifiers on the CPD (top four attributes).
EnsembleTechniquePredictive ModelPerformance Evaluation Metrics
% Accuracy% Precision% Recall% f1-Score% roc_auc
Boosting ensembleTrain and test splitAdaBoost66.6752.3850.0051.1662.80
GB63.4947.6245.4546.5159.31
LGBM65.0850.0047.7348.8461.06
Cross-validationAdaBoost66.5150.6044.2246.6963.68
GB67.5052.3545.3248.1068.25
LGBM67.7253.8948.6550.0067.33
The best-performing results for the critical metrics calculated by each algorithm are shown in bold.
Table 12. Non-ensemble learning performance analysis of classifiers on the HCC Survival dataset (top four attributes).
Table 12. Non-ensemble learning performance analysis of classifiers on the HCC Survival dataset (top four attributes).
EnsembleTechniquePredictive ModelPerformance Evaluation Metrics
% Accuracy% Precision% Recall% f1-Score% roc_auc
Non-ensemble
learning
Train and
test split
AdaBoost67.7459.3873.0865.5268.48
DT75.8168.9776.9272.7375.96
ETT69.3562.0769.2365.4569.34
GB75.8168.9776.9272.7375.96
KNN67.7460.7165.3862.9667.41
LGBM70.9764.2969.2366.6770.73
LR62.9054.2973.0862.3064.32
RF70.9765.3865.3865.3870.19
SVC45.1643.33100.0060.4752.78
Cross-validationAdaBoost72.5573.4170.5671.0678.57
DT66.7166.8163.8964.4766.81
ETT74.5572.7474.8773.0982.39
GB74.6074.2073.1473.1783.65
KNN73.1276.0067.8170.1275.54
LGBM69.6470.4169.5768.7679.87
LR57.3841.0152.9045.5174.11
RF73.9874.2771.9172.4982.53
SVC53.9552.5199.2367.7380.02
The best-performing results for the critical metrics calculated by each algorithm are shown in bold.
Table 13. Bagging classifier performance analysis using nine base learners on top four attributes—HCC Survival dataset.
Table 13. Bagging classifier performance analysis using nine base learners on top four attributes—HCC Survival dataset.
EnsembleTechniqueBase EstimatorPerformance Evaluation Metrics
% Accuracy% Precision% Recall% f1-Score% roc_auc
Bagging
ensemble
Train and test splitAdaBoost74.1966.6776.9271.4374.57
DT69.3565.2257.6961.2267.74
ETT69.3562.9665.3864.1568.80
GB74.1969.2369.2369.2373.50
KNN70.9765.3865.3865.3870.19
LGBM69.3562.0769.2365.4569.34
LR62.9054.8465.3859.6563.25
RF74.1969.2369.2369.2373.50
SVC45.1643.33100.0060.4752.78
Cross-validationAdaBoost75.9877.7179.6078.5982.99
DT73.5074.8672.9976.2482.59
ETT76.5075.9175.5574.4183.74
GB76.9878.4577.6776.4884.98
KNN73.5775.2872.1771.9575.82
LGBM74.5575.5373.4271.0882.81
LR73.1471.8674.9172.0280.06
RF77.4576.2776.7576.4284.56
SVC49.6045.2669.2351.9979.92
The best-performing results for the critical metrics calculated by each algorithm are shown in bold.
Table 14. Boosting ensemble learning performance analysis of classifiers on the HCC Survival dataset (top four attributes).
Table 14. Boosting ensemble learning performance analysis of classifiers on the HCC Survival dataset (top four attributes).
EnsembleTechniquePredictive ModelPerformance Evaluation Metrics
% Accuracy% Precision% Recall% f1-Score% roc_auc
Boosting
ensemble
Train and test splitAdaBoost66.1357.5873.0864.4167.09
GB75.8168.9776.9272.7375.96
LGBM69.3562.9665.3864.1568.80
Cross-validationAdaBoost72.0772.2471.5471.0778.95
GB74.6074.2073.1473.1783.65
LGBM71.0771.5271.7470.6480.95
The best-performing results for the critical metrics calculated by each algorithm are shown in bold.
Table 15. Non-ensemble learning performance analysis of classifiers on top four attributes—ILPD.
Table 15. Non-ensemble learning performance analysis of classifiers on top four attributes—ILPD.
EnsembleTechniquePredictive ModelPerformance Evaluation Metrics
% Accuracy% Precision% Recall% f1-Score% roc_auc
Non-ensemble
learning
Train and
test split
AdaBoost73.1479.4185.0482.1363.35
DT66.2978.8173.2375.9260.57
ETT69.7176.4384.2580.1557.75
GB68.5776.0982.6879.2556.96
KNN67.4378.6975.5977.1160.71
LGBM73.1479.8584.2581.9964.00
LR74.2974.4098.4384.7554.42
RF70.2977.3783.4680.3059.44
SVC72.5772.57100.0084.1150.00
Cross-validationAdaBoost69.4676.0683.3479.3569.51
DT62.9574.0973.8673.7553.92
ETT68.9576.0282.6178.8571.34
GB69.4676.0184.1179.5471.56
KNN63.2975.0473.0173.8366.49
LGBM70.6677.6982.9079.8971.36
LR71.5173.8893.5582.2272.52
RF68.4475.8381.8178.4871.39
SVC71.3571.35100.0083.0861.78
The best-performing results for the critical metrics calculated by each algorithm are shown in bold.
Table 16. Bagging classifier performance using nine base learners on the ILPD (top four attributes).
Table 16. Bagging classifier performance using nine base learners on the ILPD (top four attributes).
EnsembleTechniqueBase EstimatorPerformance Evaluation Metrics
% Accuracy% Precision% Recall% f1-Score% roc_auc
Bagging ensembleTrain and test splitAdaBoost76.0078.1592.9184.8962.08
DT70.8676.7685.8381.0458.54
ETT71.4375.1690.5582.1455.69
GB71.4375.8488.9881.8856.99
KNN72.0074.6892.9182.8154.79
LGBM71.4376.9286.6181.4858.93
LR74.8674.8598.4385.0355.46
RF70.8675.3388.9881.5955.95
SVC72.5772.57100.0084.1150.00
Cross-validationAdaBoost70.6576.3289.0983.0171.23
DT70.3275.9384.9779.8572.07
ETT70.1574.5688.6281.1573.62
GB71.6975.6987.9381.7473.31
KNN69.9773.8287.9980.2468.17
LGBM69.1274.8687.0079.6772.74
LR72.0374.4993.1082.3272.62
RF70.6575.2288.1480.9772.72
SVC71.3571.35100.0083.0869.71
The best-performing results for the critical metrics calculated by each algorithm are shown in bold.
Table 17. Boosting ensemble learning performance analysis of classifiers on the ILPD (top four attributes).
Table 17. Boosting ensemble learning performance analysis of classifiers on the ILPD (top four attributes).
EnsembleTechniquePredictive ModelPerformance Evaluation Metrics
% Accuracy% Precision% Recall% f1-Score% roc_auc
Boosting
ensemble
Train and test splitAdaBoost72.0079.5582.6881.0863.21
GB68.0077.1079.5378.2958.51
LGBM73.1479.8584.2581.9964.00
Cross-validationAdaBoost68.5975.6982.2878.6468.07
GB67.5775.8979.9477.6468.97
LGBM68.7775.7982.5378.7571.52
The best-performing results for the critical metrics calculated by each algorithm are shown in bold.
Table 18. Main risk factors identified in the four datasets for the diagnosis and prediction of liver disease.
Table 18. Main risk factors identified in the four datasets for the diagnosis and prediction of liver disease.
DataBest Rated FeatureDescription
BUPAGammagtGamma-glutamyl transpeptidase
SgptAlanine aminotransferase
SgotAspartate aminotransferase
alkphosAlkaline phosphatase
HCC SURVIVAL DATASETAFPAlpha-Fetoprotein (ng/mL)
HemoglobinHemoglobin (g/dL)
ALPAlkaline phosphatase (U/L)
AlbuminAlbumin (mg/dL)
ILPDAlkphosAlkaline Phosphatase
SgotAspartate Aminotransferase
SgptAlanine Aminotransferase
AgeAge of the patient
CPDProthrombinprothrombin time in seconds [s]
Albuminalbumin in [gm/dL]
Plateletsplatelets per cubic [mL/1000]
AgeAge of the patient
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Guarneros-Nolasco, L.R.; Alor-Hernández, G.; Prieto-Avalos, G.; Sánchez-Cervantes, J.L. Early Identification of Risk Factors in Non-Alcoholic Fatty Liver Disease (NAFLD) Using Machine Learning. Mathematics 2023, 11, 3026. https://doi.org/10.3390/math11133026

AMA Style

Guarneros-Nolasco LR, Alor-Hernández G, Prieto-Avalos G, Sánchez-Cervantes JL. Early Identification of Risk Factors in Non-Alcoholic Fatty Liver Disease (NAFLD) Using Machine Learning. Mathematics. 2023; 11(13):3026. https://doi.org/10.3390/math11133026

Chicago/Turabian Style

Guarneros-Nolasco, Luis Rolando, Giner Alor-Hernández, Guillermo Prieto-Avalos, and José Luis Sánchez-Cervantes. 2023. "Early Identification of Risk Factors in Non-Alcoholic Fatty Liver Disease (NAFLD) Using Machine Learning" Mathematics 11, no. 13: 3026. https://doi.org/10.3390/math11133026

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop