Next Article in Journal
The Effect of Participation Motivations on Interpersonal Relationships and Learning Achievement of Female College Students in Sports Club: Moderating Role of Club Involvement
Next Article in Special Issue
Factors Affecting Onset and Persistence of Metabolic Syndrome in Korean Breast Cancer Survivors: A Prospective Study
Previous Article in Journal
The Shifting Risk of Homelessness among Persons with a Disability: Insights from a National Policy Intervention
Previous Article in Special Issue
Effect of a Graduated Walking Program on the Severity of Obstructive Sleep Apnea Syndrome. A Randomized Clinical Trial
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

XGBoost-Based Framework for Smoking-Induced Noncommunicable Disease Prediction

by
Khishigsuren Davagdorj
1,†,
Van Huy Pham
2,†,
Nipon Theera-Umpon
3,4 and
Keun Ho Ryu
2,4,*
1
Database and Bioinformatics Laboratory, College of Electrical and Computer Engineering, Chungbuk National University, Cheongju 28644, Korea
2
Faculty of Information Technology, Ton Duc Thang University, Ho Chi Minh 700000, Vietnam
3
Department of Electrical Engineering, Faculty of Engineering, Chiang Mai University, Chiang Mai 50200, Thailand
4
Biomedical Engineering Institute, Chiang Mai University, Chiang Mai 50200, Thailand
*
Author to whom correspondence should be addressed.
These authors contributed equally to the research.
Int. J. Environ. Res. Public Health 2020, 17(18), 6513; https://doi.org/10.3390/ijerph17186513
Submission received: 22 July 2020 / Revised: 28 August 2020 / Accepted: 5 September 2020 / Published: 7 September 2020
(This article belongs to the Collection Health Behaviors, Risk Factors, NCDs and Health Promotion)

Abstract

:
Smoking-induced noncommunicable diseases (SiNCDs) have become a significant threat to public health and cause of death globally. In the last decade, numerous studies have been proposed using artificial intelligence techniques to predict the risk of developing SiNCDs. However, determining the most significant features and developing interpretable models are rather challenging in such systems. In this study, we propose an efficient extreme gradient boosting (XGBoost) based framework incorporated with the hybrid feature selection (HFS) method for SiNCDs prediction among the general population in South Korea and the United States. Initially, HFS is performed in three stages: (I) significant features are selected by t-test and chi-square test; (II) multicollinearity analysis serves to obtain dissimilar features; (III) final selection of best representative features is done based on least absolute shrinkage and selection operator (LASSO). Then, selected features are fed into the XGBoost predictive model. The experimental results show that our proposed model outperforms several existing baseline models. In addition, the proposed model also provides important features in order to enhance the interpretability of the SiNCDs prediction model. Consequently, the XGBoost based framework is expected to contribute for early diagnosis and prevention of the SiNCDs in public health concerns.

1. Introduction

Noncommunicable diseases (NCDs) have emerged as a major public health problem in the world. About 40 million people die from NCDs each year, equivalent to 70% of all deaths globally. The major risk factors of developing NCDs consist of tobacco use, physical inactivity, alcohol use, and unhealthy diets [1]. Approximately 80% of all heart disease, stroke, and diabetes would be prevented if these major risk factors were eliminated. Tobacco use is negatively associated with all of the United Nation (UN)’s Sustainable Development Goals (SDGs). In particular, smoking cessation plays an extensive role in a global effort to achieve the SDGs target to reduce deaths from NCDs by one-third by 2030 [2,3]. Studies by public health experts found that smokers were more likely to become infected with outbreaks of Middle East respiratory syndrome coronavirus (MERS-CoV) and corona virus disease (COVID) 19 compared to non-smokers [4,5].
The World Health Organization (WHO) introduced the MPOWER package, corresponding to assist in country-level implementation of interventions to increase smoking cessation, as ratified by the WHO Framework Convention on Tobacco Control (WHO-FCTC). The MPOWER package is part of the WHO action plan for the prevention and control of noncommunicable diseases. South Korea has ratified the WHO-FCTC and, since 2005, has launched anti-smoking clinics at community health centers across the country. However, the prevalence of NCDs continues to remain high, accounting for 81% of all deaths and seven of the top causes of death in South Korea [6]. In recent years, the United States government has aimed to expand its role in addressing the challenges of SiNCDs. NCDs account for 89% of all deaths in the United States, which far exceeds the cases of infectious diseases as considered causes of death [7]. Current smokers who suffer from NCDs need lifelong treatment. One of the main causes of this adverse scenario is the fact that smoking-induced noncommunicable diseases (SiNCDs) is mostly diagnosed in late stages.
Nowadays, predictive models are frequently employed in early diagnosis and forecasting of smoking-associated illnesses and diseases [8,9,10,11,12]. Early detection and efficient treatments are solutions for reducing death rates caused by chronic diseases. In one study [13], the authors studied the association between environmental factors and the development of Crohn’s disease among Japanese. Their results suggest that passive smoking history is associated with the development of Crohn’s disease. Another study [14] evaluated the association between Parkinson’s disease and rural living, farming, pesticide use, and cigarette smoking. The weight of the evidence and meta-analysis showed that there is a causal relationship between the risk of Parkinson’s disease and cigarette smoking, which has been consistently discovered in related literature. In contrast, rural living, well-water consumption, farming and the use of pesticides, herbicides, insecticides, and fungicides were less consistent within Parkinson’s disease. Furthermore, one study [15] focused on the role of feature risk pathways in smoking-induced lung cancer using patient data from the Gene Expression Omnibus database. They optimized the feature sets using the anomaly score and the recursive feature elimination (RFE) method. Then, the support vector machine (SVM) based prediction model was used. Their study concluded that smoking is the main cause of lung cancer; moreover, stress and self-protection mechanisms in a living organism can be identified as complex factors. In another study [16], authors developed an automatic classifier to increase the accuracy of the forced oscillation technique for early diagnosis of smoking-induced respiratory changes. They utilized several machine learning techniques, such as logistic linear classifiers, k nearest neighbor (KNN), neural networks (NN), and SVM. As their result, KNN and SVM classifiers resulted in a further increase in diagnostic accuracy.
Artificial intelligence (AI) utilization of automated diagnosis processes can highlight enhanced rules in the decision support system regarding patient health care. However, there are some difficulties in the selection of representative features and suitable classifier. Numerical studies have proposed selecting features using information gain, gain ratio, and correlation coefficients. However, these techniques do not consider the interactions among the features, and are not suitable for direct application to ensemble generation [17]. Moreover, it is evidently seen that ensemble-based classifiers improve the performance better than that of any single classifier [18]. Otherwise, a sequential ensemble learns to generate a model, and one tries to reduce the bias of the combined estimator and reach close to actual predictions. Thus, new models are learned from the mistakes of previous models by boosting techniques. Tree boosting provides high performance clinical predictive modeling; furthermore, it allows representation of feature importance and ranking [19].
Therefore, we propose efficient extreme gradient boosting (XGBoost) based framework incorporated with the hybrid feature selection (HFS) method for SiNCDs prediction, using real-world National Health and Nutrition Examination Survey (NHANES) datasets of South Korea and the United States. Firstly, HFS is performed in three stages: (I) significant features are selected based on statistical hypothesis tests, such as t-test and chi-square; (II) multicollinearity analysis serves to obtain dissimilar features; followed by (III) final selection of best representative features is done based on least absolute shrinkage and selection operator (LASSO). Then, selected features are fed into the XGBoost predictive model. Finally, our proposed model provides feature importance score based on XGBoost. Therefore, the proposed model is compared and contrasted against several existing baseline models, such as logistic regression (LR), random forest (RF), KNN, multilayer perceptron (MLP), NN, support vector machine recursive feature elimination (SVM-RFE), and RF feature importance (RFFI) feature selection methods. The accuracy, sensitivity, specificity, precision, f-scores, and area under the curve (AUC) analysis are employed to evaluate model performances. The main contributions of this study are as follows:
  • Proposing an efficient extreme gradient boosting (XGBoost) based framework incorporated with the hybrid feature selection (HFS) method for smoking–induced noncommunicable diseases (SiNCDs) prediction.
  • Applying the XGBoost based framework to real-world NHANES datasets of South Korea and the United States. Our empirical comparison analysis shows that the proposed model outperformed existing baseline models.
  • Findings are expected to contribute toward achieving good health and wellbeing (ultimate targets of SDGs of UN).
The remainder of this paper is logically structured as follows: Section 2 introduces the proposed framework—elaborating on its two main components: the three-stage HFS method and a brief introduction of the XGBoost algorithm. Therefore, it includes the experimental setup, regularizing hyperparameters. Section 3 provides discussed datasets, baseline models, and overall experimental results. Finally, the study is concluded in Section 4.

2. Materials Methods

2.1. Proposed Framework

In this paper, we propose an efficient extreme gradient boosting (XGBoost) based framework incorporated with the hybrid feature selection (HFS) method for smoking-induced noncommunicable diseases prediction. From its illustration in Figure 1, the proposed framework comprises of three main components: first, three-stage HFS, and then application of XGBoost to build the model. Finally, it provides the feature importance scores.

2.1.1. Three-Stage HFS

The feature selection to reduce the data dimensionality and keep only the important features is performed in three stages.
  • Step 1: Statistical Hypothesis Test (t-test and p-value)
    The first stage excludes redundant and irrelevant features in order to reduce the complexity for training model. For this purpose, it assesses chi-square test for categorical features and t-test for continuous features to accept or reject the alternative hypothesis. After the test, if deemed significant, features are stored for the first stage filtering. Otherwise, such features are excluded.
  • Step 2: Multicollinearity Analysis
    The key assumption behind the multicollinearity analysis [20] is that it indicates the correlation between independent features. The value of variance inflation factor is used to verify multicollinearity in regression analysis. In essence, this step takes the complete set of features and loops through all of them applying the appropriate test.
  • Step 3: Least Absolute Shrinkage and Selection Operator (LASSO)
    LASSO [21] has been extensively used in both fields of statistics and machine learning. Several studies [22,23] proposed the LASSO method for estimating the causal effect to identify their outcomes. A rational decision is taken to execute the LASSO in order to select a group of features simultaneously for a given task. During the feature selection process, LASSO penalizes the coefficients of the regression features, regularizing some of them to zero. On the contrary, features that still have a non-zero coefficient after the regularizing process remain to be part of the training model. This stage allows us to prevent the predictive of the causal inference problem.

2.1.2. XGBoost Classifier

XGBoost is an efficient and scalable machine learning classifier, which was popularized by Chen and Guestrin in 2016 [24]. Gradient boosting decision tree is the original model of XGBoost, which combines multiple decision trees in boosting way. In general, each new tree is created to reduce the residual of the previous model by the gradient boosting. Residual is designated by the differences between the actual and predicted values. Until the number of decision trees specify threshold, the model has been trained. XGBoost has following the same principle of gradient boosting; it uses the number of boosts, learning rate, subsampling ratio, and maximum tree depth to control overfitting and enhance the better performance. More importantly, XGBoost optimizes the objective of function, size of the tree, and magnitude of the weights, which are controlled by standard regularization parameters. The XGBoost accomplishes superior performance with numerous hyperparameters in specific searching space as summarized in Table 1.
According to the hyperparameters, gamma γ ( 0 , + ) denotes minimum loss reduction, which requires to make a split for making the partition on a leaf node of the tree. Minimum child weight w m c ( 0 , + ) defines as minimum sum of instance weight, which means if the tree partition step results in a leaf node with the sum of instance weight less than w m c , then the tree will discard further partition. Early stop algorithm works for finding the optimal epoch number referring to given other hyperparameters. Finally, XGBoost also offered subsampling techniques and r c   ( 0 ,   1 ) column subsample ratio constructs in each tree. In the final step, grid search is used to regulate the hyperparameters in order to minimize the classification error.

2.2. Experimental Setup

2.2.1. Experimental Environment

In this study, all experiments were performed on computer with 3.20 GHz, Intel Core i5-8250U (Intel Corporation, Santa Clara, CA, USA), and 8 GB Random access memory (RAM) using a Microsoft Windows 10 operating system (Microsoft Corporation, Redmond, WA, USA). Scikit-learn, Statsmodels, Matplotlib, and other libraries [25,26,27,28] of Python were used to develop the proposed and comparative models, respectively.

2.2.2. Baseline Models

We compare the proposed model with the following baseline classifiers and feature selection methods:
Logistic regression (LR) [29] is a widely used statistical method for classification and regression task. LR is used when our target of interest has two possible dichotomy values that is limited to values between 0 and 1.
Random forest (RF) [30] is a parallel structured ensemble tree-based method that utilizes bagging to aggregate multiple decision tree classifiers. Each tree of the RF is trained on bootstrap samples of the training sets, using randomly selected features in the tree generation process; after that, each tree votes for the most popular class. In this study, we have configured the number of estimators in the random forest as 500, 750, 1000, 1250, 1500; moreover, quality of split-measured criteria was selected by “gini” for the Gini impurity and “entropy” for the information gain, respectively.
K-nearest neighbor (KNN) [31] is a supervised machine learning algorithm that can solve the classification task. In the classification phase, instances are classified to the class most frequently occurring amongst the neighbors, measured by the distance function. For the KNN classifier, hyperparameters of weights and number of neighbors were adjusted in this study. The weights set up to “uniform”, where all points in each neighborhood are weighted the same, or “distance” where closer points are more heavily weighted toward the decision. The setting of the neighbor numbers refers to how many neighboring points are to fall inside of one group. Furthermore, we have turned the value of the k number between 3 and 12.
Multilayer perceptron (MLP) and neural network (NN) [32,33]: MLP is the most typical type of neural network application using back propagation for training. Neural networks are inspired by composed of nodes. MLPs consist of at least three layers, such as input, hidden, and output. Nodes in neighboring layers are interconnected, but nodes in the same layer are not. Each connection between neurons is multiplied by the corresponding weight during training. Finally, the output of hidden nodes is estimated by applying an activation function and output layer makes decisions. For the MLP models, we use the one and three hidden layers with five nodes. NN models consist of turning the two to ten hidden layers using two to five nodes, respectively. These models are optimized by Adam. Moreover, we set the learning rate is 0.001 with “sigmoid” activation function.
Support vector machine recursive feature elimination (SVM-RFE) [34] estimate the weights of the features according to the support vectors, then eliminate the necessary features until the specified number of features is reached.
Random forest based feature selection (RFFS) [35] has been found to provide feature importance scores that are successfully utilized in data mining. On the other hand, RF classifier estimates the importance of each features, then naturally ranks them.

3. Experimental Results and Discussion

The comparison of our proposed model with the baseline models is presented in this section. The flowchart of the experimental design is depicted in Figure 2. This study is reported according to the Transparent Reporting of a multivariable prediction model for Individual Prognosis or Diagnosis (TRIPOD) statement [36], shown in Appendix A (Table A1).
Initially, we preprocess the smoking-induced noncommunicable diseases (SiNCDs) datasets to eliminate missing values and outliers. Next, we elect a subset of representative features using feature selection methods. Finally, classifiers are used to build predictive models. Comparison findings would be to reveal a suitable combination of the feature selection methods and classifiers in an efficient predictive model among each dataset.

3.1. Dataset

In this study, the National Health and Nutrition Examination Survey datasets of South Korea (KNHANES) and the United States (NHANES) were used to build the proposed model and other existing baseline models for SiNCDs prediction.
KNHANES data is conducted by the Korea Centers for Disease Control and Prevention (KCDC) (http://knhanes.cdc.go.kr) [37]. It consists of a health examination of various numbers of diseases, health interviews, and nutrition surveys of the Korean population. NHANES is designed to assess the health and nutrition status of the general population in the United States. This nationwide survey is a major program of the National Center for Health Statistics (NCHS) that is part of the Centers for Disease Control and Prevention (CDC) (https://www.cdc.gov/nchs/nhanes) [38]. Generally, this survey examines approximately 5000 people each year across the United States. NHANES consists of demographic, socioeconomic, dietary, and health-related questions. Furthermore, some important features were surveyed from a minor of the population in both of the KNHANES and NHANES, similarly.
We combined KNHANES datasets from 2013 through 2017, and NHANES datasets from 2013 through 2018, as shown in Figure 3 and Figure 4. Datasets contain a large number of missing values and outliers. It is well known that missing values reduce the statistical power and become a cause of the bias in the estimation of parameters. To prevent model complications, we excluded all of the missing values and outliers. The outliers were removed based on the interquartile range. Essentially 22,183 subjects of KNHANES and 19,292 subjects of NHANES were excluded due to missing value, outliers, and class targets, which were stored in given initial features. Additionally, subjects aged 20 years old were considered for our analysis for both KNHANES and NHANES datasets. This study was designed to include target and healthy control groups. Healthy control group was defined by subjects who had never smoked and had not been diagnosed with NCDs. On the other hand, target group was defined by subjects who had a history of one of the NCDs, for diabetes, prediabetes, asthma, heart failure, corona hearth disease, heart attack, stroke, hypertension, kidney failure, or angina, as well as had smoked at least 100 cigarettes in their life.

3.2. The Results of Hybrid Feature Selection (HFS)

At the first stage of HFS, t-test and chi-square statistical hypothesis tests are used to determine the null or alternative hypothesis in order to select significant features. The threshold of 0.01 indicates statistical significance. If the feature is of a numeric type, the significance is tested using t-test, otherwise, for the categorical features, chi-square is used. In the KNHANES dataset, the p-values of “self-management”, “daily activities”, and “economic activity status” features were estimated by 0.16, 0.58, and 0.50, respectively. While p-values of “Pulse regular or irregular?”, “Salt usage level”, and “Description of job/work situation” features were indicated by 0.19, 0.12, and 0.13, respectively. A large p-value (>0.01) indicates weak evidence against the null hypothesis; thus, we exclude these features due to criteria of the first stage of HFS.
Thereafter, we verify to check the collinearity between independent features using multicollinearity in regression analysis after eliminating the non-significant features in the second stage of HFS. Multicollinearity is one of the major concerns in causal inference. This second stage leads to prevent the causality issue that occurs when two or more features are highly correlated. It is challenging for a reliable estimation of the variable coefficients. It is suspected that multicollinearity will present if the variance inflation factor (VIF) lies between 5 and 10 in this study. If the VIF value is greater than those values, it investigates a high correlation among features that remain problematic.
In our analysis of the KNHANES dataset, we did not remove any features in terms of their low VIF values. On the contrary, we removed “annual household income” and “poor appetite or overeating” features that represented VIF values of 5.312 and 5.005 in the NHANES dataset. The detailed results of the bivariate and multicollinearity analysis of the KNHANES dataset in Table A2 and NHANES dataset in Table A3 are shown in Appendix B.
In the third stage of HFS, the least absolute shrinkage and selection operator (LASSO) helps to increase the prediction of the models by removing irrelevant features that are not related to target classes. The LASSO identified irrelevant features, such as “residence area”, “walk duration (hours)”, and “health checkup status” in KNHANES dataset, whereas, “ever told doctor had trouble sleeping?” and “number of healthcare counseling over the past year” in NHANES dataset. Thus, these irrelevant features were eliminated by assigning them a coefficient equal to zero. In terms of the three-stage HFS method, we selected sufficiently representative 26 of 32 features in KNHANES and 28 of 35 features in NHANES. Moreover, these are used as inputs to the predictive model.

3.3. The Results of the Comparative Analysis

To prove the efficient XGBoost based framework equipped with the HFS method for SiNCDs, it is compared with other current techniques, including LR, RF, KNN, MLP, NN, and XGBoost incorporated with HFS, SVM-RFE, and RFFS methods in terms of KNHANES and NHANES datasets.
In this study, the entire process of parameter estimation in most baseline models are utilized based on the research paper [39]. For evaluating the prediction models, we split the data into 80% for the training set and 20% for the evaluation set. In order to prevent overfitting, a 5-fold cross-validation procedure [40] is applied to the training set. In the procedure of 5-fold cross-validation, the dataset is randomly partitioned into five parts: 4 folds of the training set are used to train classification models, and the remaining 1 fold is used to validate the model. To evaluate the performance of the predictive models, the classification accuracy, sensitivity, specificity, precision, f-score, and area under the receiver operating characteristic curve (AUC) [41,42] were used.
Table 2 shows the performances of all predictive models in the KNHANES dataset and highest performance of evaluation metrics are marked in bold. For KNHANES dataset, XGBoost with HFS model achieved the highest accuracy of 0.8812 precision of 0.8737 and F-score of 0.8707. NN with RFFS models outperformed the best sensitivity of 0.8871 and specificity of 0.8902. Following by it, the second-best accuracy of 0.8758 and precision of 0.8691 were achieved by NN with HFS model, a specificity of 0.8496 is reached by RF with HFS, and sensitivity of 0.8782 and F-score of 0.8703 was reached by XGBoost with RFFS in the prediction of SiNCDs. As can be seen, KNN with SVM-RFE based predictive model performed slightly lower results compared with other predictive models in terms of the evaluation metrics.
As shown in Table 3, we have summarized results of the predictive model in the NHANES dataset and highest performance of evaluation metrics are marked in bold. For the NHANES dataset, the best model was distinguished by our proposed XGBoost with HFS in terms of the accuracy, sensitivity, specificity, precision, and F-score, which reached 0.9309, 0.8944, 0.9522, 0.8874, and 0.8909, respectively. Moreover, second-best performances were yielded by the XGBoost with RFFS, which achieved the accuracy of 0.9029, sensitivity of 0.8507, specificity of 0.9379, precision of 0.8264, and f-score of 0.8384. Otherwise, it can be seen that the XGBoost classifier exhibited the best capability of probability prediction results incorporating different feature selection methods, significantly.
On the contrary, SVM-RFE method achieved the lowest prediction performances compared with the other feature selection method; therefore, that SVM-RFE method is not suitable for SiNCDs predictive models. Besides, the RFFS method performed computable results with the proposed HFS method in our prediction task. It is well known that accuracy is the most appropriate metric for evaluating predictive models.
According to the accuracy score, Figure 5 and Figure 6 illustrate the boxplots of the prediction models in the KNHANES and NHANES datasets. In the figures, x-axis denotes the accuracy scores and y-axis presents the utilized predictive models on SiNCDs. As depicted in Figure 5, the proposed XGBoost based framework equipped with HFS presented the highest score in the KNHANES dataset. Thereafter, NN with HFS showed the second-highest score, otherwise, the HFS method was capable of predicting the target in this task. By contrast, KNN with SVM-RFE and RFFS with RF models reached the worst scores of 0.7342 and 0.7804, respectively. Furthermore, the proposed XGBoost with the HFS model also achieved the highest score in the NHANES dataset as represented in Figure 6. Interestingly, the worst scores of 0.7349 and 0.7903 were exhibited by the LR with SVM-RFE and LR with HFS, respectively. In the NHANES dataset, the LR baseline classifier determined the worst predictive model, but results slightly improved when this classifier was combined with RFFS.
Table 4 shows the AUC analysis results of the predictive models in the KNHANES and NHANES datasets; results were verified by statistical significance test. It is evidently seen that, in terms of AUC, predictive models were statistically significant. A careful look at the results of KNHANES dataset, the highest performance of 0.8887 (95% CI, 0.8659–0.9005) was scored by NN with RFFS, following by 0.8402 (95% CI, 0.8384–0.8635) was achieved by our proposed XGBoost with the HFS model. Whilst SVM-RFE with the KNN model performed the lowest performance of 0.7170 (95% CI, 0.7094–0.7390) in this analysis. For the NHANES dataset, the proposed model indicated the better performances of 0.9233 (95% CI, 0.9073–0.9345) than other benchmark baselines. Moreover, XGBoost with the RFFS model achieved the second-best performance of 0.8943 (95% CI, 0.8757–0.9013), followed by HFS with RF of 0.8647 (95% CI, 0.8564–0.8859), significantly. These results provided evidence that our ensemble models evaluated the better performances among other baseline models in the NHANES dataset.
Figure 7 illustrates the ROC curves for the SiNCDs predictive models on the three kinds of feature selection methods across KNHANES and NHANES datasets. In ROC curves analysis, we can demonstrate the separation and discrimination ability of the predictive models. The ROC curve was plotted with the measurements of true positive rate (sensitivity) along with the y-axis, and false positive (1-specificity) along with the x-axis. For the KNHANES dataset, ROC curves of NN and XGBoost classifiers represented high results across SVM-RFE based models. RFFS based NN and XGBoost models indicated notably higher results compared with RFFS based models, significantly. Moreover, the three-stage HFS method based classifiers showed fluently good results among other baselines. ROC curve analysis of the NHANES dataset, the RF, and XGBoost classifiers incorporated with SVM-RFE had higher results across other SVM-RFE based models. Moreover, it can be seen that the proposed HFS based XGBoost model emerged as being a good combination model, as it can reach notable significant results than the other HFS and RFFS based baseline models.
The results of the KNHANES dataset reported in Table 4 and Figure 6 indicate the enhanced performances of not only the XGBoost classifier, but also the RFFS based NN considered as a computable model during the comparison task.

3.4. Interpretability of Predictive Model

Prediction performance concerns the ability of the best predictive model to make correct decisions. Meanwhile, predictive model interpretability concerns the understanding of model decisions by humans. Interpretability methods can be categorized into three types: explain data, build an inherently interpretable model (in modeling), and allow to explain it after building the models [43]. In practice, there have been some needs for using machine learning models to ensure which factors are used to make key decisions with boosted trees [44]. Model inherent interpretability is important to get a reasoning behind the predictive models. However, model interpretability tends to be ignored in previous studies [15,16].
Our proposed XGBoost based framework incorporated with HFS provides important features in order to enhance the interpretability of SiNCDs prediction model across KNHANES and NHANES datasets as depicted in Figure 8 and Figure 9. To ensure model interpretability, features were sorted in descending order of their importance scores in XGBoost based model construction in each dataset. For the KNHANES dataset, “monthly drinking rate”, “depression diagnosis”, “lifetime drinking experience”, and “total cholesterol“ were maintained as the most useful features, with importance scores of 0.2933, 0.2551, 0.1940, and 0.1763 to predict SiNCDs among the Korean population, as shown in Figure 8. While with the NHANES dataset analysis, it is evidently shown that “doctor ever said you were overweight”, “the number of people who smoke inside this home”, “general health condition”, and “age” were determined as the most important scores with 0.2158, 0.1754, 0.1621, and 0.1492, respectively, across the general population in the United States, as presented in Figure 9.
Accordingly, alcohol drinking frequency and consumption mostly occurred as highly scored features in the Korean population. Moreover, studies [45,46] found a similar factor for NCDs across Thailand and Korea, although, SiNCDs of Korea and the United States were eminently caused by overweight, cholesterol levels, and obesity, similarly. This result is similar to that study [47], where high rates of overweight and obesity increased the burden of type 2 diabetes, coronary heart disease, and stroke in most countries in the Middle East. According to Kinra et al. [48], the risks for severe illness from NCDs increased with older adults in 1600 villages from 18 states in India. Moreover, they identified that lower socioeconomic status is associated with smoking, alcohol use, low intake of fruit (vegetables), and underweight, whereas, higher socioeconomic status is associated with greater exposure to obesity, dyslipidemia, diabetes in men, and hypertension in women. In particular, authors highlighted that the prevalence of cigarette smoking among men and obesity among women was significantly higher in rural India. Dan et al. [49] examined the main (and interaction) effects of age, gender, body mass index (BMI), and dietary intake among Korean hypertensive patients. Their analysis has found that BMI, energy intake, and cholesterol intake decreased in the older-aged group compared to the middle-aged group. In addition, their findings interpreted that both genders considered weight and dietary management for reducing the incidence of hypertension. In one study, Maimela et al. [50] determined the prevalence of risk factors of NCDs among rural communities in the Limpopo Province of South Africa. Their results defined that tobacco prevalence, alcohol consumption, and being overweight has consistently higher association with NCDs among adults.
Otherwise, most of the notable risk factors for NCDs in each dataset were represented as modifiable. It is well known that modifiable risk factors are behaviors and exposures that are highly associated with the risk of developing various diseases. In order to prevent and correct these modifiable risk factors, it required actions, such as smoking cessation, alcohol reduction, and exercise in public health. The highly scored features enhance rational decisions in smoking-related health concerns and should be collected in SiNCDs prediction data. In addition, this analysis is expected to compare the similarities and differences between populations of two different countries for SiNCDs prediction.

4. Conclusions

In this study, we proposed extreme gradient boosting (XGBoost) based framework incorporated with a hybrid feature selection (HFS) method for SiNCDs prediction, using real-world National Health and Nutrition Examination Survey (NHANES) datasets of South Korea and the United States. The proposed framework consisted of three main steps: first, a three-stage hybrid feature selection method to select important features, then built the XGBoost predictive model to accomplish the task of predicting SiNCDs, and, finally, the framework provided the XGBoost based feature importance scores to enhance the understanding of the reasoning behind the predictive models. The model under the proposed framework was compared against various existing baselines and it has shown superior performance in terms of accuracy measures across each dataset. We also determined the most representative features for SiNCDs in the general populations of South Korea and the United States. Although the study has successfully demonstrated that smoking induced serious health hazards, it has certain limitations in terms of interpretability of deep learning, known as the black-box. In the future, this study can be extended by addressing the problem of global and local interpretability of black-box models and causal effect in the scenario of predictive models.

Author Contributions

K.D. and V.H.P. contributed equally to this work. K.D. conceived to perform the experimental results and V.H.P. also conceived to perform the experimental results. K.D., V.H.P., N.T.-U., and K.H.R. discussed and improved the contents of the manuscript. N.T.-U. provided critical insight and discussion. K.H.R. supervised this work. All authors approved the final manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Science, ICT & Future Planning (No. 2020R1A2B5B02001717), and (No.2019K2A9A2A06020672).

Acknowledgments

The authors would like to thank reviewers for their suggestions to improve the manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Table A1. Transparent Reporting of a multivariable prediction model for Individual Prognosis or Diagnosis (TRIPOD) Checklist: Prediction Model Development and Validation.
Table A1. Transparent Reporting of a multivariable prediction model for Individual Prognosis or Diagnosis (TRIPOD) Checklist: Prediction Model Development and Validation.
Section/TopicItem Checklist ItemPage
Title and abstract
Title1D; VIdentify the study as developing and/or validating a multivariable prediction model, the target population, and the outcome to be predicted.1
Abstract2D; VProvide a summary of objectives, study design, setting, participants, sample size, predictors, outcome, statistical analysis, results, and conclusions.2
Introduction
Background and objectives3aD; VExplain the medical context (including whether diagnostic or prognostic) and rationale for developing or validating the multivariable prediction model, including references to existing models.2–3
3bD; VSpecify the objectives, including whether the study describes the development or validation of the model or both.3–5
Methods
Source of data4aD; VDescribe the study design or source of data (e.g., randomized trial, cohort, or registry data), separately for the development and validation datasets, if applicable.6–7
4bD; VSpecify the key study dates, including start of accrual; end of accrual; and, if applicable, end of follow-up.6–7
Participants5aD; VSpecify key elements of the study setting (e.g., primary care, secondary care, general population), including number and location of centers.6–7
5bD; VDescribe eligibility criteria for participants.6–7
Figure 3 and Figure 4
5cD; VGive details of treatments received, if relevant.n/a
Outcome6aD; VClearly define the outcome that is predicted by the prediction model, including how and when assessed.8–12
6bD; VReport any actions to blind assessment of the outcome to be predicted.9–12
Predictors7aD; VClearly define all predictors used in developing the multivariable prediction model, including how and when they were measured.7–8
7bD; VReport any actions to blind assessment of predictors for the outcome and other predictors.7–8
Sample size8D; VExplain how the study size was arrived at.7–8
Missing data9D; VDescribe how missing data were handled (e.g., complete-case analysis, single imputation, multiple imputation) with details of any imputation method.7
Statistical analysis methods10aDDescribe how predictors were handled in the analyses.7–8
10bDSpecify type of model, all model-building procedures (including any predictor selection), and method for internal validation.4–5, 8
Figure 1
10cVFor validation, describe how the predictions were calculated.8
10dD; VSpecify all measures used to assess model performance and, if relevant, to compare multiple models.8
Figure 2
10eVDescribe any model updating (e.g., recalibration) arising from the validation, if done.8
Risk groups11D; VProvide details on how risk groups were created, if done.7
Development vs. validation12VFor validation, identify any differences from the development data in setting, eligibility criteria, outcome, and predictors.8
Results
Participants13aD; VDescribe the flow of participants through the study, including the number of participants with and without the outcome and, if applicable, a summary of the follow-up time. A diagram may be helpful.Figure 3 and Figure 4
13bD; VDescribe the characteristics of the participants (basic demographics, clinical features, available predictors), including the number of participants with missing data for predictors and outcome.Figure 3 and Figure 4
13cVFor validation, show a comparison with the development data of the distribution of important variables (demographics, predictors, and outcome).Table A2 and Table A3
Model development14aDSpecify the number of participants and outcome events in each analysis.7–8
14bDIf done, report the unadjusted association between each candidate predictor and outcome.8–9
Model specification15aDPresent the full prediction model to allow predictions for individuals (i.e., all regression coefficients, and model intercept or baseline survival at a given time point).10
15bDExplain how to use the prediction model.8–12
Model performance16D; VReport performance measures (with CIs) for the prediction model.11
Table 4
Model-updating17VIf done, report the results from any model updating (i.e., model specification, model performance).12–13
Discussion
Limitations18D; VDiscuss any limitations of the study (such as non-representative sample, few events per predictor, missing data).15
Interpretation19aVFor validation, discuss the results with reference to performance in the development data, and any other validation data.13–14
19bD; VGive an overall interpretation of the results, considering objectives, limitations, results from similar studies, and other relevant evidence.14–15
Implications20D; VDiscuss the potential clinical use of the model and implications for future research.14–15
Other information
Supplementary information21D; VProvide information about the availability of supplementary resources, such as study protocol, web calculator, and datasets.16–18
Funding22D; VGive the source of funding and the role of the funders for the present study.15
Items relevant only to the development of a prediction model are denoted by D, items relating solely to a validation of a prediction model are denoted by V, and items relating to both are denoted D; V.

Appendix B

Table A2. Bivariate and Multicollinearity analysis for KNHANES dataset.
Table A2. Bivariate and Multicollinearity analysis for KNHANES dataset.
Featuresp-ValueMulticollinearity Coefficient
1Gender<0.011.081
2Age<0.011.523
3Household income<0.012.901
4Education<0.011.125
5Occupation<0.012.016
6Marital status<0.013.553
7Subjective health status<0.011.778
8Depression diagnosis<0.011.286
9Health checkup status<0.011.047
10Athletic ability<0.011.124
11Self-management0.16~
12Daily activities0.58~
13Pain/discomfort<0.012.229
14Anxious/Depressed<0.014.345
15EQ-5D index<0.012.473
16Economic activity status0.50~
17Weight control: exercise<0.013.329
18Lifetime drinking experience<0.011.171
19Start drinking age<0.011.003
20Frequency of drinking for 1 year<0.011.532
21Monthly drinking rate<0.013.152
22Stress level<0.013.033
23Indoor indirect smoking exposure<0.011.221
24The usual time spent sitting (day)<0.011.096
25Walk duration (hours)<0.011.114
26Family history of chronic disease<0.011.087
27Body mass index (kg/m2)<0.011.048
28Obesity prevalence<0.012.675
29Fasting blood sugar<0.012.536
30Total cholesterol<0.011.151
31Flexible exercise days per week<0.011.038
32Residence area<0.011.547
Table A3. Bivariate and Multicollinearity analysis for NHANES dataset.
Table A3. Bivariate and Multicollinearity analysis for NHANES dataset.
Featuresp-ValueMulticollinearity Coefficient
1Gender<0.012.005
2Age<0.011.008
3Body mass index (kg/m2)<0.013.005
4Pulse regular or irregular?0.19~
5Systolic: blood pressure<0.012.015
6Diastolic: blood pressure<0.011.875
7Education level<0.011.076
8Marital status<0.013.092
9Total number of people in the household0.32~
10Annual household income<0.015.312
11Health risk for diabetes (among family history)<0.013.533
12Taking insulin or not<0.011.453
13Number of healthcare counseling over past year<0.012.027
14Salt usage level0.12~
15Total sugars (gm)<0.011.298
16Alcohol (gm)<0.012.479
17Frequency of alcohol usage<0.012.204
18High cholesterol level<0.011.340
19General health condition<0.011.199
20#times receive healthcare over past year<0.011.249
21Received hepatitis A vaccine<0.012.012
22Family monthly poverty level category<0.011.004
23Doctor ever said you were overweight<0.011.012
24Doctor told you to exercise<0.011.004
25Feeling down, depressed, or hopeless<0.011.012
26Feeling tired or having little energy<0.011.004
27Poor appetite or overeating<0.015.005
28Trouble concentrating on things<0.012.292
29Description of job/work situation0.13~
30Ever told doctor had trouble sleeping?<0.011.012
31Number of people who live here smoke tobacco?<0.011.035
32Number of people who smoke inside this home?<0.011.424
33Last 7-d worked at job not at home?<0.012.404
34Last 7-d at job someone smoked indoors?<0.011.205
35Last 7-d in other indoor area?<0.012.108

References

  1. Forouzanfar, M.H.; Afshin, A.; Alexander, L.T.; Anderson, H.R.; Bhutta, Z.A.; Biryukov, S.; Cohen, A.J. Global, regional, and national comparative risk assessment of 79 behavioural, environmental and occupational, and metabolic risks or clusters of risks, 1990–2015: A systematic analysis for the Global Burden of Disease Study 2015. Lancet 2016, 388, 1659–1724. [Google Scholar] [CrossRef] [Green Version]
  2. Kathirvel, S. Sustainable development goals and noncommunicable diseases: Roadmap till 2030–A plenary session of world noncommunicable diseases congress 2017. Int. J. Noncommunicable Dis. 2018, 3, 3. [Google Scholar] [CrossRef]
  3. World Health Organization. Action plan for the prevention and control of noncommunicable diseases in the WHO European Region. In Proceedings of the Regional Committee for Europe 66th Session, Copenhagen, Denmark, 12–15 September 2016. [Google Scholar]
  4. Vardavas, C.I.; Nikitara, K. COVID-19 and smoking: A systematic review of the evidence. Tob. Induc. Dis. 2020, 18. [Google Scholar] [CrossRef] [PubMed]
  5. Berlin, I.; Thomas, D.; Le Faou, A.L.; Cornuz, J. COVID-19 and smoking. Nicotine Tob. Res. 2020. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Yoon, J.; Seo, H.; Oh, I.H.; Yoon, S.J. The non-communicable disease burden in Korea: Findings from the 2012 Korean Burden of Disease Study. J. Korean Med Sci. 2016, 31 (Suppl. 2), S158–S167. [Google Scholar] [CrossRef]
  7. Chen, S.; Kuhn, M.; Prettner, K.; Bloom, D.E. The macroeconomic burden of noncommunicable diseases in the United States: Estimates and projections. PLoS ONE 2018, 13, e0206702. [Google Scholar] [CrossRef]
  8. Hu, X.; Wang, Y.; Huang, J.; Zheng, R. Cigarette Affordability and Cigarette Consumption among Adult and Elderly Chinese Smokers: Evidence from A Longitudinal Study. Int. J. Environ. Res. Public Health 2019, 16, 4832. [Google Scholar] [CrossRef] [Green Version]
  9. Davagdorj, K.; Yu, S.H.; Kim, S.Y.; Huy, P.V.; Park, J.H.; Ryu, K.H. Prediction of 6 Months Smoking Cessation Program among Women in Korea. Int. J. Mach. Learn. Comput. 2019, 9, 83–90. [Google Scholar] [CrossRef] [Green Version]
  10. Ng, M.; Freeman, M.K.; Fleming, T.D.; Robinson, M.; Dwyer-Lindgren, L.; Thomson, B.; Murray, C.J. Smoking prevalence and cigarette consumption in 187 countries, 1980-2012. JAMA 2014, 311, 183–192. [Google Scholar] [CrossRef] [Green Version]
  11. Davagdorj, K.; Lee, J.S.; Park, K.H.; Ryu, K.H. A machine-learning approach for predicting success in smoking cessation intervention. In Proceedings of the 2019 IEEE 10th International Conference on Awareness Science and Technology (iCAST), Morioka, Japan, 23–25 October 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 1–6. [Google Scholar]
  12. Al-Obaide, M.A.; Ibrahim, B.A.; Al-Humaish, S.; Abdel-Salam, A.S.G. Genomic and bioinformatics approaches for analysis of genes associated with cancer risks following exposure to tobacco smoking. Front. Public Health 2018, 6, 84. [Google Scholar] [CrossRef] [Green Version]
  13. Kondo, K.; Ohfuji, S.; Watanabe, K.; Yamagami, H.; Fukushima, W.; Ito, K. Japanese Case-Control Study Group for Crohn’s disease. The association between environmental factors and the development of Crohn’s disease with focusing on passive smoking: A multicenter case-control study in Japan. PLoS ONE 2019, 14, e0216429. [Google Scholar] [CrossRef] [PubMed]
  14. Breckenridge, C.B.; Berry, C.; Chang, E.T.; Sielken Jr, R.L.; Mandel, J.S. Association between Parkinson’s disease and cigarette smoking, rural living, well-water consumption, farming and pesticide use: Systematic review and meta-analysis. PLoS ONE 2016, 11, e0151841. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Chen, R.; Lin, J. Identification of feature risk pathways of smoking-induced lung cancer based on SVM. PLoS ONE 2020, 15, e0233445. [Google Scholar] [CrossRef] [PubMed]
  16. Amaral, J.L.; Lopes, A.J.; Jansen, J.M.; Faria, A.C.; Melo, P.L. An improved method of early diagnosis of smoking-induced respiratory changes using machine learning algorithms. Comput. Methods Programs Biomed. 2013, 112, 441–454. [Google Scholar] [CrossRef] [Green Version]
  17. Piao, Y.; Piao, M.; Ryu, K.H. Multiclass cancer classification using a feature subset-based ensemble from microRNA expression profiles. Comput. Biol. Med. 2017, 80, 39–44. [Google Scholar] [CrossRef]
  18. Dietterich, T.G. An experimental comparison of three methods for constructing ensembles of decision trees: Bagging, boosting, and randomization. Mach. Learn. 2000, 40, 139–157. [Google Scholar] [CrossRef]
  19. Zihni, E.; Madai, V.I.; Livne, M.; Galinovic, I.; Khalil, A.A.; Fiebach, J.B.; Frey, D. Opening the black box of artificial intelligence for clinical decision support: A study predicting stroke outcome. PLoS ONE 2020, 15, e0231166. [Google Scholar] [CrossRef] [Green Version]
  20. Salmerón Gómez, R.; García Pérez, J.; López Martín, M.D.M.; García, C.G. Collinearity diagnostic applied in ridge estimation through the variance inflation factor. J. Appl. Stat. 2016, 43, 1831–1849. [Google Scholar] [CrossRef]
  21. Meier, L.; Van De Geer, S.; Bühlmann, P. The group lasso for logistic regression. J. R. Stat. Soc. Ser. B (Stat. Methodol.) 2008, 70, 53–71. [Google Scholar] [CrossRef] [Green Version]
  22. Belloni, A.; Chernozhukov, V.; Hansen, C. High-dimensional methods and inference on structural and treatment effects. J. Econ. Perspect. 2014, 28, 29–50. [Google Scholar] [CrossRef] [Green Version]
  23. Ghosh, D.; Zhu, Y.; Coffman, D.L. Penalized regression procedures for variable selection in the potential outcomes framework. Stat. Med. 2015, 34, 1645–1658. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  24. Chen, T.; Guestrin, C. Xgboost: A scalable tree boosting system. In Proceedings of the 22nd Acm Sigkdd International Conference on Knowledge Discovery and Data Mining, New York, CA, USA, 13–17 August 2016; pp. 785–794. [Google Scholar]
  25. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Vanderplas, J. Scikit-learn: Machine learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  26. Seabold, S.; Perktold, J. Statsmodels: Econometric and statistical modeling with python. In Proceedings of the 9th Python in Science Conference, Austin, TX, USA, 28–30 June 2010. [Google Scholar]
  27. Hunter, J.D. Matplotlib: A 2D graphics environment. Comput. Sci. Eng. 2007, 9, 90. [Google Scholar] [CrossRef]
  28. Géron, A. Hands-on Machine Learning with Scikit-Learn, Keras, and TensorFlow: Concepts, Tools, and Techniques to Build Intelligent Systems, 2nd ed.; Géron, A., Ed.; O’Reilly Media Inc.: Seastopol, CA, USA, 2019. [Google Scholar]
  29. Bagley, S.C.; White, H.; Golomb, B.A. Logistic regression in the medical literature: Standards for use and reporting, with particular attention to one medical domain. J. Clin. Epidemiol. 2001, 54, 979–985. [Google Scholar] [CrossRef]
  30. Liaw, A.; Wiener, M. Classification and regression by randomForest. R News 2002, 2, 18–22. [Google Scholar]
  31. Tan, P.N. Introduction to Data Mining, Pearson Education India; Indian Nursing Council: New Delhi, India, 2018.
  32. Lisboa, P.J. A review of evidence of health benefit from artificial neural networks in medical intervention. Neural Netw. 2002, 15, 11–39. [Google Scholar] [CrossRef]
  33. Mazurowski, M.A.; Habas, P.A.; Zurada, J.M.; Lo, J.Y.; Baker, J.A.; Tourassi, G.D. Training neural network classifiers for medical decision making: The effects of imbalanced datasets on classification performance. Neural Netw. 2008, 21, 427–436. [Google Scholar] [CrossRef] [Green Version]
  34. Lin, X.; Yang, F.; Zhou, L.; Yin, P.; Kong, H.; Xing, W.; Xu, G. A support vector machine-recursive feature elimination feature selection method based on artificial contrast variables and mutual information. J. Chromatogr. B 2012, 910, 149–155. [Google Scholar] [CrossRef]
  35. Qi, Y. Random forest for bioinformatics. In Ensemble Machine Learning; Springer: Boston, MA, USA, 2012; pp. 307–323. [Google Scholar]
  36. Collins, G.S.; Reitsma, J.B.; Altman, D.G.; Moons, K.G. Transparent Reporting of a Multivariable Prediction Model for Individual Prognosis or Diagnosis (TRIPOD) The TRIPOD Statement. Circulation 2015, 131, 211–219. [Google Scholar] [CrossRef] [Green Version]
  37. Korea Centers for Disease Control & Prevention. Available online: http://knhanes.cdc.go.kr (accessed on 7 September 2020).
  38. Centers for Disease Control and Prevention. Available online: https://www.cdc.gov/nchs/nhanes (accessed on 7 September 2020).
  39. Davagdorj, K.; Lee, J.S.; Pham, V.H.; Ryu, K.H. A Comparative Analysis of Machine Learning Methods for Class Imbalance in a Smoking Cessation Intervention. Appl. Sci. 2020, 10, 3307. [Google Scholar] [CrossRef]
  40. Fushiki, T. Estimation of prediction error by using K-fold cross-validation. Stat. Comput. 2011, 21, 137–146. [Google Scholar] [CrossRef]
  41. Goutte, C.; Gaussier, E. A probabilistic interpretation of precision, recall and F-score, with implication for evaluation. In Proceedings of the European Conference on Information Retrieval, Santiago de Compostela, Spain, 21–23 March 2005; Springer: Berlin/Heidelberg, Germany, 2005; pp. 345–359. [Google Scholar]
  42. Altman, D.G.; Bland, J.M. Diagnostic tests. 1: Sensitivity and specificity. BMJ Br. MedJ. 1994, 308, 1552. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  43. Carvalho, D.V.; Pereira, E.M.; Cardoso, J.S. Machine learning interpretability: A survey on methods and metrics. Electronics 2019, 8, 832. [Google Scholar] [CrossRef] [Green Version]
  44. Elshawi, R.; Al-Mallah, M.H.; Sakr, S. On the interpretability of machine learning-based model for predicting hypertension. BMC Med. Inform. Decis. Mak. 2019, 19, 146. [Google Scholar] [CrossRef] [Green Version]
  45. Wakabayashi, M.; McKetin, R.; Banwell, C.; Yiengprugsawan, V.; Kelly, M.; Seubsman, S.A. Thai Cohort Study Team. Alcohol consumption patterns in Thailand and their relationship with non-communicable disease. BMC Public Health 2015, 15, 1297. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  46. Kim, H.C.; Oh, S.M. Noncommunicable diseases: Current status of major modifiable risk factors in Korea. J. Prev. Med. Public Health 2013, 46, 165. [Google Scholar] [CrossRef]
  47. Kilpi, F.; Webber, L.; Musaigner, A.; Aitsi-Selmi, A.; Marsh, T.; Rtveladze, K.; Brown, M. Alarming predictions for obesity and non-communicable diseases in the Middle East. Public Health Nutr. 2014, 17, 1078–1086. [Google Scholar] [CrossRef] [Green Version]
  48. Kinra, S.; Bowen, L.J.; Lyngdoh, T.; Prabhakaran, D.; Reddy, K.S.; Ramakrishnan, L.; Smith, G.D. Sociodemographic patterning of non-communicable disease risk factors in rural India: A cross sectional study. BMJ 2010, 341, c4974. [Google Scholar] [CrossRef] [Green Version]
  49. Dan, H.; Kim, J.; Kim, O. Effects of gender and age on dietary intake and body mass index in hypertensive patients: Analysis of the korea national health and nutrition examination. Int. J. Environ. Res. Public Health 2020, 17, 4482. [Google Scholar] [CrossRef]
  50. Maimela, E.; Alberts, M.; Modjadji, S.E.; Choma, S.S.; Dikotope, S.A.; Ntuli, T.S.; Van Geertruyden, J.P. The prevalence and determinants of chronic non-communicable disease risk factors amongst adults in the Dikgale health demographic and surveillance system (HDSS) site, Limpopo Province of South Africa. PLoS ONE 2016, 11, e0147926. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Extreme gradient boosting (XGBoost) based framework for smoking-induced noncommunicable diseases prediction. LASSO: least absolute shrinkage and selection operator.
Figure 1. Extreme gradient boosting (XGBoost) based framework for smoking-induced noncommunicable diseases prediction. LASSO: least absolute shrinkage and selection operator.
Ijerph 17 06513 g001
Figure 2. Design of comparative experiments for smoking-induced noncommunicable diseases prediction. K(NHANES): Korean (national health and nutrition examination survey); SVM-RFE: support vector machine recursive feature elimination; RFFS: random forest feature selection; HFS: hybrid feature selection; LR: logistic regression; KNN: k-nearest neighbors; NN: neural network; RF: random forest; MLP: multilayer perceptron; XGBoost: extreme gradient boosting.
Figure 2. Design of comparative experiments for smoking-induced noncommunicable diseases prediction. K(NHANES): Korean (national health and nutrition examination survey); SVM-RFE: support vector machine recursive feature elimination; RFFS: random forest feature selection; HFS: hybrid feature selection; LR: logistic regression; KNN: k-nearest neighbors; NN: neural network; RF: random forest; MLP: multilayer perceptron; XGBoost: extreme gradient boosting.
Ijerph 17 06513 g002
Figure 3. Sample selection procedure of the Korea National Health and Nutrition Examination Survey (KNHANES) dataset.
Figure 3. Sample selection procedure of the Korea National Health and Nutrition Examination Survey (KNHANES) dataset.
Ijerph 17 06513 g003
Figure 4. Sample selection procedure of National Health and Nutrition Examination Survey (NHANES) dataset.
Figure 4. Sample selection procedure of National Health and Nutrition Examination Survey (NHANES) dataset.
Ijerph 17 06513 g004
Figure 5. Boxplot of the accuracy over prediction models in the KNHANES dataset.
Figure 5. Boxplot of the accuracy over prediction models in the KNHANES dataset.
Ijerph 17 06513 g005
Figure 6. Boxplot of the accuracy over prediction models in the NHANES dataset.
Figure 6. Boxplot of the accuracy over prediction models in the NHANES dataset.
Ijerph 17 06513 g006
Figure 7. The comparison of receiver operating characteristic (ROC) curves for smoking-induced noncommunicable diseases (SiNCDs) prediction in the KNHANES and NHANES datasets. (a) ROC of SVM-RFE based models in KNHANES; (b) ROC of SVM-RFE based models in NHANES; (c) ROC of RFFS based models in KNHANES; (d) ROC of RFFS based models in NHANES; (e) ROC of HFS based models in KNHANES; (f) ROC of HFS based models in NHANES.
Figure 7. The comparison of receiver operating characteristic (ROC) curves for smoking-induced noncommunicable diseases (SiNCDs) prediction in the KNHANES and NHANES datasets. (a) ROC of SVM-RFE based models in KNHANES; (b) ROC of SVM-RFE based models in NHANES; (c) ROC of RFFS based models in KNHANES; (d) ROC of RFFS based models in NHANES; (e) ROC of HFS based models in KNHANES; (f) ROC of HFS based models in NHANES.
Ijerph 17 06513 g007aIjerph 17 06513 g007b
Figure 8. Feature Importance of XGBoost based framework incorporated with hybrid feature selection in the KNHANES dataset.
Figure 8. Feature Importance of XGBoost based framework incorporated with hybrid feature selection in the KNHANES dataset.
Ijerph 17 06513 g008
Figure 9. Feature Importance of XGBoost based framework incorporated with hybrid feature selection in the NHANES dataset.
Figure 9. Feature Importance of XGBoost based framework incorporated with hybrid feature selection in the NHANES dataset.
Ijerph 17 06513 g009
Table 1. Searching space of XGBoost model.
Table 1. Searching space of XGBoost model.
ParametersSymbolSearch Space
Maximum tree depth D m a x 2, 4, 6, 8
Minimum child weight w m c 2, 3, 4, 5
Early stop rounde100
Learning rate τ 0.1
Number of boostN60
Maximum delta step δ 0.4, 0.6, 0.8, 1
Subsample ratio r s 0.9, 0.95, 1
Column subsample ratio r c 0.9, 0.95, 1
Gamma γ 0, 0.001
Table 2. Evaluation results of the prediction models in the Korea National Health and Nutrition Examination Survey dataset.
Table 2. Evaluation results of the prediction models in the Korea National Health and Nutrition Examination Survey dataset.
Feature SelectionClassifierAccuracySensitivitySpecificityPrecisionF-Score
SVM-RFELR0.79480.78180.75320.76760.7746
RF0.78900.79890.79840.81150.8052
KNN0.73420.69580.73810.79610.7426
MLP0.80700.79360.77910.80160.7976
NN0.81970.82740.82030.83870.8330
XGBoost0.80980.81080.83100.85330.8315
RFFSLR0.78040.73710.74220.80240.7684
RF0.82640.76990.73380.82360.7958
KNN0.80480.71280.76610.77530.7427
MLP0.79940.78080.73960.81150.7959
NN0.85070.88710.89020.85220.8693
XGBoost0.83110.87820.79840.86260.8703
HFSLR0.78340.79890.78130.79590.7974
RF0.83620.78050.84960.81150.7957
KNN0.80320.80180.71230.78720.7944
MLP0.84210.83050.75130.82570.8281
NN0.87580.85180.81580.86910.8604
XGBoost0.88120.86770.81260.87370.8707
SVM-RFE: support vector machine recursive feature elimination; RFFS: random forest feature selection; HFS: hybrid feature selection; LR: logistic regression; KNN: k-nearest neighbors; NN: neural network; RF: random forest; MLP: multilayer perceptron; XGBoost: extreme gradient boosting. Highest scores are marked in bold.
Table 3. Evaluation results of the prediction models in the National Health and Nutrition Examination Survey dataset.
Table 3. Evaluation results of the prediction models in the National Health and Nutrition Examination Survey dataset.
Feature SelectionClassifierAccuracySensitivitySpecificityPrecisionF-Score
SVM-RFELR0.73490.69690.88740.70860.7027
RF0.85220.79040.88050.81570.8029
KNN0.81180.74320.86080.81050.7754
MLP0.80020.71710.87590.68160.6989
NN0.83390.76590.83970.76090.7634
XGBoost0.82480.77070.85120.80660.7882
RFFSLR0.83560.71690.86850.69380.7052
RF0.87410.78630.90650.73560.7601
KNN0.84440.77160.86350.75940.7655
MLP0.82210.70430.89490.68420.6941
NN0.86390.76510.90030.75340.7592
XGBoost0.90290.85070.93790.82640.8384
HFSLR0.79030.77810.89900.77320.7756
RF0.89610.81570.91360.78570.8004
KNN0.83630.79280.89900.79810.7954
MLP0.79180.75860.90830.76350.7610
NN0.85530.81730.88080.79340.8052
XGBoost0.93090.89440.95220.88740.8909
Highest scores are marked in bold.
Table 4. Statistical significance test of the area under the curve results for predictive models in the KNHANES and NHANES datasets.
Table 4. Statistical significance test of the area under the curve results for predictive models in the KNHANES and NHANES datasets.
Feature SelectionClassifierKNHANES DatasetNHANES Dataset
AUCCI 95%p-ValueAUCCI 95%p-Value
SVM-RFELR0.76750.7474–0.7896<0.0010.79220.7731–0.8088<0.001
RF0.79870.7869–0.8118<0.0010.83550.8254–0.8668<0.001
KNN0.71700.7094–0.7390<0.0010.80200.7818–0.8210<0.001
MLP0.78640.7703–0.8001<0.0010.79650.7851–0.8180<0.001
NN0.82390.8017–0.8405<0.0010.80280.7981–0.8447<0.001
XGBoost0.82090.8097–0.8327<0.0010.81100.8041–0.8315<0.001
RFFSLR0.73970.7713–0.7971<0.0010.79270.7806–0.8197<0.001
RF0.75190.7683–0.8111<0.0010.84640.8359–0.8637<0.001
KNN0.73950.7570–0.8037<0.0010.81760.8070–0.8308<0.001
MLP0.76020.7721–0.8267<0.0010.79960.7872–0.8135<0.001
NN0.88870.8659–0.9005<0.0010.83270.8206–0.8492<0.001
XGBoost0.83830.8245–0.8567<0.0010.89430.8757–0.9013<0.001
HFSLR0.79010.7812–0.8253<0.0010.83860.8234–0.8539<0.001
RF0.81510.7947–0.8286<0.0010.86470.8564–0.8859<0.001
KNN0.75710.7401–0.7796<0.0010.84590.8284–0.8653<0.001
MLP0.79090.7846–0.8243<0.0010.83350.8195–0.8506<0.001
NN0.83380.8249–0.8494<0.0010.84910.8310–0.8588<0.001
XGBoost0.84020.8384–0.8635<0.0010.92330.9073–0.9345<0.001
Highest scores are marked in bold.

Share and Cite

MDPI and ACS Style

Davagdorj, K.; Pham, V.H.; Theera-Umpon, N.; Ryu, K.H. XGBoost-Based Framework for Smoking-Induced Noncommunicable Disease Prediction. Int. J. Environ. Res. Public Health 2020, 17, 6513. https://doi.org/10.3390/ijerph17186513

AMA Style

Davagdorj K, Pham VH, Theera-Umpon N, Ryu KH. XGBoost-Based Framework for Smoking-Induced Noncommunicable Disease Prediction. International Journal of Environmental Research and Public Health. 2020; 17(18):6513. https://doi.org/10.3390/ijerph17186513

Chicago/Turabian Style

Davagdorj, Khishigsuren, Van Huy Pham, Nipon Theera-Umpon, and Keun Ho Ryu. 2020. "XGBoost-Based Framework for Smoking-Induced Noncommunicable Disease Prediction" International Journal of Environmental Research and Public Health 17, no. 18: 6513. https://doi.org/10.3390/ijerph17186513

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop