Next Article in Journal
Delivering Extended Cellular Coverage and Capacity Using High-Altitude Platforms
Next Article in Special Issue
Automatic Classification of Hospital Settings through Artificial Intelligence
Previous Article in Journal
A Novel Anti-Risk Method for Portfolio Trading Using Deep Reinforcement Learning
Previous Article in Special Issue
Deep-Forest-Based Encrypted Malicious Traffic Detection
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Machine Learning Models for Early Prediction of Sepsis on Large Healthcare Datasets

by
Javier Enrique Camacho-Cogollo
1,†,
Isis Bonet
1,†,
Bladimir Gil
2 and
Ernesto Iadanza
3,*,†
1
Biomedical Engineering Department, EIA University, Calle 23 AA Sur Nro. 5-200, Km 2 + 200 Via Al Aeropuerto, Envigado 55428, Colombia
2
Clínica Las Américas, Diagonal 75B N. 2A-80/140, Medellin 50025, Colombia
3
Department of Medical Biotechnologies, University of Siena, Via Aldo Moro 2, 53100 Siena, Italy
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Electronics 2022, 11(9), 1507; https://doi.org/10.3390/electronics11091507
Submission received: 7 April 2022 / Revised: 30 April 2022 / Accepted: 2 May 2022 / Published: 7 May 2022

Abstract

:
Sepsis is a highly lethal syndrome with heterogeneous clinical manifestation that can be hard to identify and treat. Early diagnosis and appropriate treatment are critical to reduce mortality and promote survival in suspected cases and improve the outcomes. Several screening prediction systems have been proposed for evaluating the early detection of patient deterioration, but the efficacy is still limited at individual level. The increasing amount and the versatility of healthcare data suggest implementing machine learning techniques to develop models for predicting sepsis. This work presents an experimental study of some machine-learning-based models for sepsis prediction considering vital signs, laboratory test results, and demographics using Medical Information Mart for Intensive Care III (MIMIC-III) (v1.4), a publicly available dataset. The experimental results demonstrate an overall higher performance of machine learning models over the commonly used Sequential Organ Failure Assessment (SOFA) and Quick SOFA (qSOFA) scoring systems at the time of sepsis onset.

1. Introduction

The most recent definition of sepsis defines the syndrome as a “life-threatening organ dysfunction due to a dysregulated host response to infection” [1,2]. Sepsis is a highly complex and heterogeneous syndrome, which is influenced by two types of characteristics: those associated with the patient (e.g., immunologic status, age, comorbidities) [3,4], and those associated with the infection (e.g., site of infection, pathogen type, virulence) [1,5]. As a consequence, their relative contribution to organ damage may vary across patients due to the combinations of such characteristics: sepsis is a common pathway from infection to death. Kidneys, liver, lungs, heart, the central nervous system, and the hematologic system are among the most commonly affected [2].
Management is a major challenge for healthcare systems, affecting high- and low-income countries alike [6]. Sepsis continues to be a leading cause of death worldwide; in 2017, there were an estimated 48.9 million cases and 11 million sepsis-related deaths worldwide [7]. Globally, it is associated with a 10–20% in-hospital mortality rate [2,8,9,10]. Sepsis can also be very expensive to treat, the costs for sepsis management in U.S. hospitals rank highest among admissions for all disease states, amounting to over USD 20 billion in 2011, growing to over USD 23 billion in 2013, and cost more than USD 24 billion annually [9,11,12,13].
Various investigations have demonstrated that early diagnosis and appropriate antibiotic therapy have the potential to reduce mortality among the septic population [14,15,16]. Despite this, detecting sepsis at this stage of the disease is difficult. One of the main reasons is that sepsis is a heterogeneous syndrome whose course depends on different pathophysiological mechanisms, complexity in clinical context, and clinical phenotypes. For clinicians and researchers, the challenge is to objectively assess the true magnitude of organ failure for each patient [17]. There is also a lack of accurate diagnostic tools [18]. Different and complex scoring systems have been defined for bedside evaluation, allowing early detection of patient deterioration [2,19,20]. Several scores are currently used, such as: Acute Physiology and Chronic Health Evaluation (APACHE II) [21], Simplified Acute Physiology Score (SAPS II) [22], Sequential Organ Failure Assessment (SOFA) [23], and Quick SOFA (qSOFA) [2], which have been validated to assess illness severity from the weighted combination of clinical and laboratory measures. However, these scoring systems, indeed useful for predicting general deterioration or mortality in studies conducted in intensive care units (ICU), cannot identify sepsis with high sensitivity and specificity at an individual level [2,24]. In addition, they present significant errors at patient data in clinical use, showing methodological weaknesses [25].
The accelerating generation of vast amounts of healthcare data will completely change the nature of medical care. The clinical decision support systems that integrate both laboratory data and biomarkers have the potential to greatly improve patient outcomes [26,27,28]. Assessing the utility of these complex datasets for supporting clinical decisions is the aim of most current research projects [29,30]. Recent studies have deployed ensemble models using real-world events to find important predictive parameters to track down those patients who are at the highest risk of death [31].
In the last decade, a substantial number of papers have been published on clinical applications of machine learning for the early prediction of sepsis in ICU [32,33,34,35,36,37,38,39]. However, there are important disparities in the identification of sepsis cohorts due the use of various approaches and considerable heterogeneity in definitions [30,32,40]. Many studies use ICD coding [36,41,42], which is considered an unreliable tool with limitations in identifying septic patients [43,44,45]. Others are based on systemic inflammatory response syndrome (SIRS) [37,46] or use the latest sepsis-3 criteria [47,48,49]. The major concerns regarding the clinical applicability and the actual accuracy of these models are related with the definition of sepsis, the identification of organ dysfunction, and the selection of input features [32,50,51,52,53,54,55]. By strategically deploying machine learning (ML) models and carefully selecting underlying data, cohorts, and features, algorithm developers can mitigate these concerns [51,52].
Several works used vital signs measurements and laboratory test results in order to develop models to predict sepsis during ICU stay [56,57]. Some trained ML algorithms using physiological parameters only, as input features. For example, in [49] the authors used six vital signs—heart rate (HR), respiratory rate (RR), temperature (T), peripheral oxygen saturation (SpO2), systolic blood pressure (SBP), and diastolic blood pressure (DBP)—as the inputs in an XGBoost model. Recent studies have investigated the change in variability of vital sings with a set of the following four features: arterial pressure (AP), HR, RR and T. They developed a support vector machine (SVM) algorithm achieving a 0.88 AUC-ROC for predicting four hours before the sepsis onset [58]. Other models have added the use of laboratory test results, demographics, and comorbidities to improve the results [42,59]. In another study, a collection of six continuous minute-by-minute physiological data (HR, RR, T, S p O 2 , SBP, DBP) and the white blood cell (WBC) count were used in a random forest (RF) model with a 0.8 sensitivity within an hour before the sepsis onset [37]. In [60], the authors used 34 physiological variables, of which 5 were vitals and 29 were laboratory values in a multitask Gaussian process through a deep recurrent neural network model. Others have added the use of the six vitals, Ph, WBC, and the age to be analyzed as times series in the inSight model [41] for a 0.92 AUC-ROC.
Although there is no database with labeled sepsis cases, many studies have used the MIMIC II and MIMIC III databases. Using these databases, a wide range of prediction models have been applied for the early detection of sepsis. Some of the most prominent are artificial neural networks (ANN) [46], deep temporal convolution networks [61], deep learning (DL) models [48], RFs [37], and SVMs [58,62]. Although some of these recent efforts are promising, the lack of comparability of these studies does not allow a direct comparison [33,48]. In order to reduce this gap in a meaningful way and promote the cooperation between clinicians and researchers, the first contribution of this work is the provision of a reference database for researchers. This was developed with a detailed analysis of organs dysfunction related with sepsis patients on the MIMIC-III database.
Furthermore, this research explores the use of a proposed ML-based ensemble classification technique to develop a model for early sepsis prediction. The role of feature selection methods on the performance of the proposed ML model was investigated. To summarize, this work focuses on answering the following questions:
1.
What are the most relevant factors associated with the early prediction of ICU sepsis?
2.
Do ML techniques, especially ensemble models, outperform the current sepsis mortality scoring?
The contributions of this work are summarized as follows:
  • An applicable framework is proposed for identifying sepsis onset cases in ICU for adult patients. This structured approach yields a cohort selection well suited for identifying organs dysfunction and suspected infection, following the SEPSIS-3 definition.
  • An accurate and explainable sepsis prediction ensemble model is also proposed based on a comprehensive list of critical features from ICU patients. The model is based on a new ensemble algorithm using vital signs, laboratory test results, and demographics tabular variables as input. We discuss the development of different ML models, and the results are compared with the traditional scoring system.
  • The effects of various feature selection methods on the identification of the relevant features for 24, 12, and 6 h observation windows were analyzed, and the results were evaluated for 1 h prediction time.

2. Materials and Methods

This section details the selected dataset, preprocessing steps, and feature selection in the experiments, following the methodology described in Figure 1.

2.1. Data Source

We used the de-identified public database MIMIC-III v1.4, which contains patient data from the Beth Israel Deaconess Medical Center (BIDMC, Boston, MA, USA) [63]. The database uses two information systems, Philips CareVue Clinical and IMDsoft MetaVision ICU, that have very different data structures. MIMIC III contains detailed information on more than 60,000 stays corresponding to more than 40,000 patients. The data are associated with 53,423 distinct hospital admissions for adult patients [64].

2.2. Data Selection

The aim of data preprocessing is improving the quality of the collected dataset. The data preprocessing includes selection patients cohort, data balancing, removing outliers, and handling missing data. Figure 2 details the number of patients in each step and the exclusion steps performed.
The first step of dataset preprocessing was the selection of the patient cohort. For this selection, we took into account different patient characteristics, as shown in Figure 3 and described below.
For this study, we worked only with adult patients, considering as such all patients older than 14 years. For this, following the approach proposed by [63], we used the age logged at the time of ICU admission.
For consistency and because the important data for identifying suspected infection does not appear in the CareVue system or on admission for cardio surgery, we decided to select only data collected via Meta Vision [64]. MIMIC III contains missing and outliers values during the admission and ICU stay for many reasons, including medical equipment failure, systems errors, and others. In this work, those patients with features with more than 60% missing data were entirely removed. Following the medical specialist recommendation, all the outliers were removed from the dataset. Patients that had at least three records for vital signs were selected; other features with missing data were imputed using k-nearest neighbors imputation [65].
Since the purpose of this study is the prediction of sepsis, we also excluded patients who were admitted to the ICU with sepsis, following the sepsis-3 criteria [2]. This means that we chose to focus only on those patients who developed organs dysfunctions related to sepsis at any time during the ICU stay (Figure 3).

2.3. Sepsis-3 Cohort Performance

For developing a prediction sepsis model, defining the sepsis cohort and the onset is important because this is the reference time to take the prior data. The structured approach used here yielded a well-suited cohort selection, firstly identifying the suspected infection and secondly the organs dysfunction following sepsis-3 [2,40]. To create this particular population, the sepsis-3 MIMIC III query reported by [30,59] was edited and used in Postgres. The code of the SQL query to obtain the cohort is available on Github (https://github.com/Biocamacho/Sepsis-cohort, accessed on 2 April 2022).
Suspected infection is defined as the administration of antibiotics and sampling of body fluid culture in a specified period. Figure 4 shows two examples, one referring to each case. Case I illustrates an example of a patient having a body fluid culture sampling at a specified time prior to antibiotics administration. In that case, the time of suspected infection was when the culture sampling was ordered. The second case (Case II) shows a patient who was administered antibiotics at a time prior to the culture sampling. In that case, the time of suspected infection was when the antibiotics were administered to the patient.
We retrieved all the timestamps of antibiotic administration and body fluid sampling over the length of ICU stay. That is, if the culture was obtained before the antibiotic, then the medication was required to be administered within 72 h, whereas if the antibiotic medication was administered first, then the sampling was required within 24 h. In this study, for each timestamp, when the first suspected infection episode was identified, we extracted the 72 h window of data for each patient, before the episode.
An organ dysfunction can be identified as an acute change in total SOFA ≧ 2 points consequent to the infection [23]. Afterwards, we calculated the SOFA score at each hour for the suspected infection cohort in a 72 h window, considering a period of 48 h before and up to 24 h after the onset of infection, using several queries reported by [59]. Using the Postgres function detailed in Figure 5, we extracted all the variables used in SOFA score from laboratory test results, and the Glasgow score, the ventilation status, and the vasopressors from the charted events. Finally, we labeled those patients with a SOFA score in this window increased by two or more points as “sepsis onset”.
Table 1 shows the summary of characteristics of this population. The table shows how the dataset is divided in two groups: those who developed sepsis and those who did not. A total of 2377 patients met the inclusion criteria. They had a mean age of 62 years and there were no notable differences in comorbidities (diabetes, cancer, and Elixhauser index), mortality, and weight, but sepsis patients had a longer overall length of stay.

2.4. Feature Selection and Extraction

We processed the dataset to determine the candidates features to be used as input for the prediction models, excluding those features with a high missing rate, as we detail in Figure 6. In this work, the queries provided by [30,59] for the selected patient cohort were extended and refined.
In the present research, 145 features were scanned and based on the medical knowledge of an expert physician, and the 31 medically relevant features (MRF) for sepsis prediction were gathered. The complete set of patient features included was grouped into three categories: physiological data (e.g., heart rate, temperature, etc.), laboratory test results (e.g., white blood count, glucose, hematocrit, hemoglobin, creatinine, bicarbonate, PH, and arterial blood gases) and demographics/score (age, Elixhauser Index, weight, and Glasgow coma score). According to [57], we developed two engineered features to improved predictive performance: (a) shock index [66] and (b) the product of age and systolic blood pressure (min, mean, and max, labeled as FE1, FE2, and FE3, respectively). All the features included are listed in Table 2. Several articles and books have mentioned those particular variables being effective and therefore should be considered during sepsis diagnosis [2,18,67].
Following [56,57,58], we calculated statistical measures including the mean, maximum (max), minimum (min), and standard deviation (std) for laboratory and physiological data. We extracted 88 statistical features from the 31 MRF to represent the patients. As shown in Figure 6, we applied a feature selection process to select the most important features prior to classification.
Feature selection has proven to be a successful preprocessing step in ML applications, building a subset of the original features with the following specific advantages: (a) avoiding overfitting, (b) improving model performance, (c) increasing speed, (d) providing cost-effective models. Feature selection methods are also classified into filters, embedded methods, and wrappers, depending on the relationship with the learning method [68]. Wrappers and embedded techniques have a higher risk of overfitting and are very computationally intensive [69]. In this study, it was preferred resorting to filter models because they are computationally simple and fast because they are independent of any learning methods, focusing on the general characteristics of the data [70]. In filter models, features are first scored and ranked according to the relevance to the class label, and then are selected based on a threshold value [71]. For each feature, each of these feature selection models has an evaluation value. Features with evaluation values greater than the threshold are chosen. Several feature selection processes exist, such as the one used by [72]. In this study, the most common filter techniques were defined to perform setting the threshold to 0.002. The filter techniques used for feature selection are: information-gain-based methods [73], relief [74], and Gini. Table A1 details the features names used in every feature set.

2.5. Prediction Methodology

To support the exhaustive comparison experiments, three long observational periods were considered and defined as the look back (LB) sequences of values that were used to predict. For each subset extracted earlier, three-time frames (24, 12, or 6 h) for the LB data were the inputs for the models. Finally, three horizons (1, 2, or 3 h) were defined as the prediction time to evaluate and compare the performance of all the models developed using the test set. The illustration of prediction methodology is shown in Figure 7.

2.6. Classification Models

In the present research, an experimental comparison study of traditional and ensemble supervised models of ML has been developed to predict sepsis during ICU stay. The following section defines the conventional classification models used and a few more ensemble methods.

2.6.1. Support Vector Machine (SVM)

Proposed by [75], this model is based on kernel classification to map N dimensional input data sets into a higher dimensional feature space. SVMs try to find the optimal generalization separating hyperplanes solving quadratic programming, represented by Equation (1).
W t · X b = 0
SVMs depend on a small subset of training dataset called support vectors. In the input space, each training datum is considered as an N dimensional vector X and its label is + 1 or 1 .

2.6.2. K-Nearest Neighbor (KNN)

As a local classification method, KNN is based on the labels of the K-nearest patterns in data space [76]. This method assumes that all the instances correspond to points in the n-dimensional space. After defining a specific distance, a neighbor is nearest if it has the smallest distance in the n-dimensional feature space. The most commonly used distance is the Euclidean (Equation (2)).
d ( x i , x j ) r = 1 n a r ( x i ) a r ( x j ) 2
The distance between two instances, x i and x j , is defined by d ( x i , x j ) , where a r ( x ) denotes the value of the rth attribute of instance x.

2.6.3. Artificial Neural Network (ANN)

This model consists of three layers: an input layer, a hidden layer, and an output layer [76]. This layer has interconnected neurons with weights and specific activation functions. The layers transform the data and pass them to the next layer. ANN can perform the task by learning from its previous inputs. Given a network with a fixed set of units and interconnections, this network adjusts the weights using the back-propagation learning process.

2.6.4. Naïve Bayes (NVC)

This classifier applies tasks where each instance x is represented by a conjunction of attribute values and where the target function can take on any value from some finite set [76]. This model is described through the construction of a Bayesian probabilistic model that assigns a posterior class probability to an instance: [77]. The NVC uses these probabilities to assign an instance to a class, applying Bayes’ theorem, as shown in Equation (3).
P ( y j | x i ) = P ( y j ) P ( x i | y j ) P ( x i )

2.6.5. Random Forest (RF)

This classifier consists of a collection of classification or regression trees, which are simple models using binary splits on predictor variables to determine outcome predictions. The process is called feature sampling [78]. The trees are the base learners for RF, and the prediction is defined by classifying the data with each tree and collecting a majority rule on the whole forest. Through the bagging mechanism, each tree is built using a different bootstrap sample of the dataset.

2.6.6. AdaBoost

Adaptive boosting (AdaBoost) is defined by Freund and Schapire [79] as a boosting ensemble learning model that selects various classifier instances by preserving an adaptive weight distribution over the training examples. It is the sum of the weights of the misclassified instances divided by the total weight of all instances [80]. Following each cycle, the weight distribution is updated based on the prediction results from the training samples. The weight of correctly classified instances is reduced, while the weight of misclassified instances is increased. Thus, AdaBoost produces a set of “easy” instances that are correctly classified with low weight and a set of “hard” ones that are misclassified with high weight. In the following iteration, a classifier is built for the reweighed data, which consequently focuses on classifying the hard instances correctly. This process continues for various cycles until, finally, AdaBoost linearly combines all the component classifiers into a single final hypothesis. Greater weights are given to component classifiers with a lower training error.

2.6.7. Stacking

Stacking is an ensemble learning technique that combines in parallel a set of different classifiers by training a meta-classifier to output a strong prediction based on the various models predictions [81]. Figure 8 shows the stacking model, which implements the following steps: first, all the base models compute on the original dataset; second, the predictions made by these classifiers are considered as a new input data to the final model; eventually, a learning process occurs using this new input to obtain a final prediction.

2.6.8. XGBoost

Extreme gradient boosting (XGBoost) [82] is another ensemble of classifiers, considered as an optimized gradient tree boosting model, based on algorithmic optimizations and hyperparameters to perform the learning process and control the overfitting. The most important factor behind the XGBoost model is the tree learning algorithm for handling sparse data. XGboost use an objective function to optimize the loss function, considering the results from the previous level, and adds regularization to perform the results.

2.7. Performance Metrics and Model Evaluation Procedures

To evaluate all the predictions models train–test split procedure and k-fold cross-validation were used as model evaluation procedures. The dataset was randomly divided into training set (75%) and test set (25%).
Figure 9 shows the structure of train and test data sets. In this study, a 10-fold cross-validation method was used for selecting the best model for training [76]. Cross-validation is defined as a method for the evaluation of the performance of a predictive model. Statistical analysis will generalize to an independent dataset. K-fold cross-validation is a method for dividing the original sample randomly into K sub-samples. The model is then tested using a single sub-sample as validation data, while the remaining K 1 sub-samples are used as training data. These process is repeated K times, with one of the K sub-samples representing as the validation data.
Class imbalance is a particular problem in medical datasets. In MIMIC-III, a majority patient had no sepsis (94.5%) and a minority of patients were sepsis diagnosed (5.5%) during ICU Stay. Two techniques are the most commonly used for data sampling: oversampling and undersampling. The latter, according to the research, has shown to be the superior option over oversampling [83,84]. In this research, the random undersampling technique was used to improve the data balance by removing patients from the prevalent class (no sepsis). Therefore, the undersampling process cut the sample to 2377 patients (77.4% no sepsis; 22.5% sepsis).
For each model included in this study, different metrics were obtained to evaluate the performance according to [85]: area under receiver operating characteristic (AUC-ROC), accuracy, F1 score, recall, specificity, and precision as the evaluation metrics to report results. A binary classification allows to create a confusion matrix, as shown in Table 3, based on the following four categorizations of predictions:
  • True positives (TP): instances correctly labeled as positive.
  • False positives (FP): negative instances incorrectly labeled as positive.
  • True negatives (TN): instances correctly labeled as negative.
  • False negatives (FN): positive instances incorrectly labeled as negative.
Accuracy measures indicate that the overall proportion of labels were correctly identified.
A c c u r a c y = T P + T F T P + T F + F P + F N
F 1 score is defined as the harmonic mean of R e c a l l and P r e c i s i o n that sets their trade-off.
F 1 = 2 P r e c i s i o n R e c a l l P r e c i s i o n + R e c a l l = 2 T P 2 T P + F P + F N
R e c a l l indicates the proportion of correctly predicted sepsis cases from the sepsis set.
R e c a l l = T P T P + F N
Likewise, specificity represents the proportion of correctly predicted no-sepsis cases among the no-sepsis group. The AUC-ROC value considers the sensitivity against the specificity at various threshold settings. Lastly, P r e c i s i o n gives us the proportion of cases which were identified as having sepsis who actually had sepsis.
P r e c i s i o n = T P T P + F P
The performance of each classifier was evaluated with a number of comparisons including qSOFA and SOFA Score.

2.8. Selection of Classification Models

Several comparison experiments were performed using the following conventional classification models: SVM, KNN, RF, ANN, Naïve Bayes, and the combination with ensemble models: boosting (Adaboost and XGBoost) and stacking schemes.
An ensemble selection approach was used to achieve improved performance. The models were implemented in Orange version 3.31 [86] and Phyton using Google Colab infrastructure [87].
In the present study, three different prediction times (pt = 1/2/3 h) were considered. Therefore, nine different data collections were produced considering the different LB defined. For all the developed models, the performances were evaluated and compared against the traditional medical scores: the qSOFA and the SOFA Score.

3. Results and Discussion

In the present study, extensive experiments were performed to compare the results of conventional classification and ensemble models using the MIMIC III dataset including vital signs, laboratory test results, and demographics. To obtain a cohort, a balance strategy was conducted, which includes the 2377 patients (537 cases and 1840 controls) from undersampling process. In this study, the objective was to develop a model for early sepsis prediction in the ICU. This work presented an experimental investigation of the combination of heterogeneous learning algorithms to develop a more accurate ensemble model. Following an ensemble selection approach, a total of six ensemble models were finally selected to achieve improved performance. Other ensemble models were tested, but did not produce accurate results compared to the six previous models. As can be seen in Table 4, E1 to E6 consist of stacking ensemble models with logistic regression as the aggregation algorithm. To obtain the most optimized results for all the models developed the hyperparameters were tuned, Table 5 shows details. As a result, the stacking model is expected to outperform all models.

Results of Feature Selection

In the present study, the performance of five conventional classifiers were evaluated—KNN, random forest, ANN, Naïve Bayes, SVM, and ensemble models—by tuning the hyperparameters and testing the following three feature selection techniques to obtain the most optimized results: information gain, GINI index, and relief. The feature selection allows the reduction in dimension of data and improves the performance of classification and ensemble models using the best subset of features. Features were selected according to their importance using the 10-fold cross validation. For example, it is important to consider the minimum value for Glasgow score. For laboratory test results, the minimum and maximum values were considered. For example, high hemoglobin can be a sign of blood thickness, which may lead to strokes. Low values of hemoglobin may also be a risk indication of kidney disease. Table A1 shows the relevant feature subset filtered using feature selection techniques.
In this experiment, the role of feature selection techniques was explored to enhance the performance of the ML algorithms. The results have been reported in Table 6, presenting AUC-ROC, accuracy, recall, and precision, using pt = 1 h, lb = 24 h, and 10-fold cross validation.
The results exhibit high recall and AUC-ROC when predicting sepsis through ensemble models and implementing information gain feature selection. Interestingly, information gain demonstrated better results for conventional and ensemble classification models, compared with the others feature selection techniques using 24 h dataset. For information gain feature selection we observed that XGBoost and random forest generate the best testing results. The XGBoost algorithm achieves values of AUC-ROC = 0.918, accuracy = 0.872, recall = 0.852, and precision = 0.868. The random forest reported an AUC-ROC = 0.905, accuracy = 0.866, recall = 0.852, and precision = 0.860. Conversely, ANN, SVM, and KNN models shows the worst results. The confusion matrix is used to identify proportions of instances between the predicted and actual class. The number of false positives and false negatives can also be seen in this matrix. Figure 10 shows the confusion matrices of XGBoost, RF, SVM, and ANN models, respectively.
For the GINI feature set, the overall testing results were worsened by 6–16 % in terms of AUC-ROC, accuracy, recall, and precision. The random forest, E1, and E5 models achieved the best testing results. In particular, E1 test results are AUC-ROC = 0.759, accuracy = 0.810, recall = 0.714, and precision = 0.797. Random forest showed the following performance: AUC-ROC = 0.752, accuracy = 0.812, recall = 0.802, and precision = 0.807. For feature set filtered by relief, the overall testing performance was improved compared to GINI sets. XGBoost and E1 offer the best performance with AUC-ROC = 0.909–0.904, accuracy = 0.876–0.874, recall = 0.869–0.852, and precision = 0.872–0.870, respectively. Figure 11 compares the AUC-ROC curves for these ML algorithms based on information gain sets.
Feature selection techniques improve the results and stability of the ML algorithms and reduce the noise. From these previous experiments, we observed that the feature set selected using information gain achieved the best results.
The test set was used to validate the ML models and to explore their generalization capabilities. For this, the three sliding windows of 6, 12, and 24 h were used.
A statistical comparison is applied to evaluate the performance of all candidates models using 1, 2, and 3 h of prediction time for the three LB windows, as shown in Table 7, Table 8 and Table 9, respectively. The results confirm the effectiveness in achieving higher accuracy of ensemble learning techniques in comparison with single classifiers. In particular, the stacking ensemble method was applied to achieve a more accurate sepsis prediction for time prediction = 1 h and look back = 24 h (E5 AUC-ROC = 0.907, E6 AUC-ROC = 0.910) compared with other ensemble models (E1, E2, E3, and E4). One interesting observation is that SVM reveals a high recall and the poorest for the other metrics. Likewise, RF exhibits high performance in accuracy and AUC-ROC. Detailed numerical results in Table 7, Table 8 and Table 9, show the XGBoost model provides better results for various prediction windows and look back dataset. The XGBoost model obtains permanent best results, even for t = 3 achieving AUC-ROC = 0.911, 0.896, and 0.886 for various LBs. The performance decreased in all models when increasing the prediction time.
Information gain contains 75 features and we found that Glasgow, weight, temperature (min), and hematocrit (max) were the five most important ones for prediction sepsis using XGBoost algorithm, as shown in Figure 12. For interpreting the effects and relative contributions of selected features and clinical parameters on sepsis prediction, an explainer Shapley additive explanation (SHAP) SHAP was implemented [88]. A tree SHAP algorithm was used to explain the output of XGBoost models. Tree SHAP is a fast and exact method to estimate SHAP values for tree models and ensembles of trees [89]. The bee swarm plot, Figure 13, shows how the value of each feature impacts model output. The color on the right side represents the feature value; red represents a higher value; and blue represents a lower value. As expected, the Glasgow scale, temperature, glucose, S p O 2 , and heart rate were identified as the most significant contributors. In addition, the blood pressure and laboratory test results (PH and bilirubin) also contribute to the sepsis prediction.
The findings confirm the factor behind the success of the XGBoost model revealed [82], for instance the tree learning technique for handling sparse data and the scalability through algorithmic optimizations.
The ensemble models E1, E5, E6 and XGBoost exhibit higher results compared to others models in all computed scenarios. They were selected for comparison with two standard severity scores—SOFA and qSOFA. Figure 14 shows the results using one hour for time prediction and 24 h for the LB features. The selected ensemble showed higher performance in terms of accuracy, recall, F1, and AUC-ROC. This may explain the strong results during the ensemble model E1. In particular, the XGBoost model formed better than all other models, achieving a superior performance.

4. Conclusions

Early diagnosis and an appropriate antibiotic therapy of sepsis is crucial but sometimes challenging. Several patient-health-score-based prediction systems have been employed for evaluating the early detection of patient deterioration. However, these scoring systems are useful for predicting general deterioration or mortality, but cannot identify sepsis in patients with high sensitivity and specificity at individual level. The increasing availability and the versatility of healthcare data suggest to implement ML techniques to develop models for predicting sepsis.
In this study, several ML-based models were applied for sepsis prediction using vital signs, laboratory test results, and demographics variability. The results demonstrate overall higher performance of ML models over the commonly used SOFA and qSOFA scoring systems at the time of sepsis onset.
The results reveal high metrics when using ensemble models and feature selection strategies for predicting sepsis. Particularly, this study demonstrated better results for information gain compared with conventional feature selection techniques. The addition of some variables from laboratory test results as input variables overall increases the model performance.
This work provides a consistent set of comparison of ML models for sepsis early prediction on the large healthcare datasets, following the sepsis-3 definition and selecting features in a meaningful way. The results are a strong motivator for implementing these ensemble models for early sepsis prediction in ICU cases.
In this study, the 12 ML models were computed with various feature windows for 1, 2, and 3 h of sepsis prediction. The ensemble models combining the variability of physiological data and laboratory test results, as well as demographics values, outperformed all the conventional algorithms. They also showed a large margin of improvement over traditional scoring systems at the time of sepsis onset. The best results were obtained using the XGBoost model implementing the information gain feature selection technique, achieving 0.911 AUC-ROC for 24 h of look back data in a t = 1 h prediction time.

Author Contributions

Conceptualization, E.I., I.B., J.E.C.-C. and B.G.; methodology, E.I., I.B., J.E.C.-C. and B.G.; software, I.B. and J.E.C.-C.; validation, J.E.C.-C., B.G. and I.B.; formal analysis, J.E.C.-C., E.I. and I.B.; investigation, J.E.C.-C. and I.B.; resources, B.G.; data curation, J.E.C.-C., B.G. and I.B.; writing—original draft preparation, J.E.C.-C.; writing—review and editing, E.I. and I.B.; visualization, J.E.C.-C.; supervision, I.B. and E.I.; project administration, I.B. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
SOFASequential Organ Failure Assessment
AUC-ROCarea under the receiver operating characteristic
MIMICMedical Information Mart for Intensive Care
PTTpartial thromboplastin time
BUNblood urea nitrogen

Appendix A

The following Table A1 shows details related with the feature subset filtered using feature selection techniques
Table A1. Feature subset filtered using feature selection techniques.
Table A1. Feature subset filtered using feature selection techniques.
Feature Selection TechniqueFeatures Names
Information GainGlasgow (min), FE3, Systolic Blood Pressure (min), FE 2, Oliguria, Systolic Blood Pressure (mean), Respiratory Rate (min), SpO2 (max), FE 1, pCO2 (min), Mean Blood Pressure (min), glucose, glucose (min), Mean Blood Pressure (mean), Temperature (min), Bicarbonate (min), Age, Bicarbonate, Glucose (max), Temperature (mean), lactate, Elixhauser hospital, Diastolic Blood Pressure (min), Respiratory Rate (mean), ph (max), lactate (min), Heart rate (min), Lactate (std), lactate (max), SpO2 (mean), Temperature (max), Bicarbonate ((std)), Creatinine (max), Creatinine (min), Bilirubin (max), Systolic Blood Pressure (max), White Blood Count (min), Respiratory Rate (max), Creatinine, Aniongap (min), Shock Index (mean), pCO2 (max), bicarbonate (max), bilirubin (min), bilirubin, Shock Index (max), White Blood Count (max), PTT (max), INR (max), anion gap, Anion Gap (std), INR, Chloride (max), Diastolic blood pressure (mean), INR (min), PTT (min), Weight, anion gap (max), Chloride (min), PT (max), Platelet, BUN (max), Hematocrit (max), BUN (min), Platelet (min), PH (min), BUN, PT (min), Shock Index (min), Diastolic blood pressure (max), Hemoglobin (max), Platelet (max), Potassium (max)
GINI IndexFE3, oliguria, Systolic Blood Pressure (min), FE2, Systolic Blood Pressure (mean), Respiratory Rate (min), SpO2 (max), FE1, Mean Blood Pressure (min), pco2 (min), glucose, glucose (min), Mean Blood Pressure (mean), Temperature (min), Age, bicarbonate (min), glucose (max), bicarbonate, lactate, Elixhauser hospital, Temperature (mean),Diastolic blood pressure (min), lactate (min), respiratory rate (mean), lactate (std), ph (max), lactate (max), Heart rate (min), SpO2 (mean), Temperature (max), bicarbonate (std), creatinine (max), creatinine (min), White Blood Count (min), creatinine, Systolic Blood Pressure (max), Respiratory rate (max), Anion gap (min), bilirubin (max), Shock Index (mean), bicarbonate (max), White Blood Count (max), Shock Index (max), pco2 (max), PTT (max), INR (max)
RiliefBilirubin (min), Bilirubin, Bilirubin (max), lactate (max), lactate, pco2 (max), ph (min), ph (max), pco2 (min), lactate (min), SpO2 (min), Weight, Diastolic Blood Pressure (max), Temperature (max), INR (max), Mean Blood Pressure (max), Heart rate (max), Temperature (mean), PT (max), SpO2 (mean), Shock Index (max), Systolic Blood pressure (max), Diastolic Blood Pressure (mean), Mean blood Pressure (mean), Shock Index (mean), Respiratory rate (mean), Respiratory rate (max), INR (min), Respiratory rate (min), Mean blood Pressure (min), Temperature (min), Systolic Blood Pressure (mean), SpO2 (max), Shock Index (min), PT (min), INR, PTT (min), Platelet (std), Bicarbonate (std), Heart rate (min), FE3, Diastolic blood pressure (min), Anion gap, FE2, Heart rate (mean), Systolic blood pressure (min), White Blood count (min), FE1, Anion gap (min), Creatinine (min), White Blood Count, Potassium (max), Creatinine, Bicarbonate, Anion gap (max), Creatinine (max), White Blood Count (max), Bicarbonate (min), Bicarbonate (max), Anion gap (std), glucose (max), BUN, BUN (min), Glasgow (min), Potassium, BUN (max), Chloride (min), PTT (max), Diabetes, Glucose (min), Glucose, Chloride (max), Platelet (max)

References

  1. Arwyn-Jones, J.; Brent, A.J. Sepsis. Surgery 2019, 37, 1–8. [Google Scholar]
  2. Seymour, C.W.; Liu, V.X.; Iwashyna, T.J.; Brunkhorst, F.M.; Rea, T.D.; Scherag, A.; Rubenfeld, G.; Kahn, J.M.; Shankar-Hari, M.; Singer, M.; et al. Assessment of Clinical Criteria for Sepsis: For the Third International Consensus Definitions for Sepsis and Septic Shock (Sepsis-3). JAMA 2016, 315, 762–774. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Sinapidis, D.; Kosmas, V.; Vittoros, V.; Koutelidakis, I.M.; Pantazi, A.; Stefos, A.; Katsaros, K.E.; Akinosoglou, K.; Bristianou, M.; Toutouzas, K.; et al. Progression into sepsis: An individualized process varying by the interaction of comorbidities with the underlying infection. BMC Infect. Dis. 2018, 18, 242. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Rowe, T.A.; McKoy, J.M. Sepsis in older adults. Infect. Dis. Clin. 2017, 31, 731–742. [Google Scholar] [CrossRef]
  5. Klastrup, V.; Hvass, A.M.; Mackenhauer, J.; Fuursted, K.; Schønheyder, H.C.; Kirkegaard, H.; Network, C.S. Site of infection and mortality in patients with severe sepsis or septic shock. A cohort study of patients admitted to a Danish general intensive care unit. Infect. Dis. 2016, 48, 726–731. [Google Scholar] [CrossRef]
  6. Papali, A.; McCurdy, M.T.; Calvello, E.J.B. A “three delays” model for severe sepsis in resource-limited countries. J. Crit. Care 2015, 30, 861.e9–861.e14. [Google Scholar] [CrossRef]
  7. Rudd, K.E.; Johnson, S.C.; Agesa, K.M.; Shackelford, K.A.; Tsoi, D.; Kievlan, D.R.; Colombara, D.V.; Ikuta, K.S.; Kissoon, N.; Finfer, S.; et al. Global, regional, and national sepsis incidence and mortality, 1990–2017: Analysis for the Global Burden of Disease Study. Lancet 2020, 395, 200–211. [Google Scholar] [CrossRef] [Green Version]
  8. Stoller, J.; Halpin, L.; Weis, M.; Aplin, B.; Qu, W.; Georgescu, C.; Nazzal, M. Epidemiology of severe sepsis: 2008–2012. J. Crit. Care 2016, 31, 58–62. [Google Scholar] [CrossRef]
  9. Reinhart, K.; Daniels, R.; Kissoon, N.; Machado, F.R.; Schachter, R.D.; Finfer, S. Recognizing Sepsis as a Global Health Priority—A WHO Resolution. N. Engl. J. Med. 2017, 377, 414–417. [Google Scholar] [CrossRef]
  10. Freund, Y.; Lemachatti, N.; Krastinova, E.; Van Laer, M.; Claessens, Y.E.; Avondo, A.; Occelli, C.; Feral-Pierssens, A.L.; Truchot, J.; Ortega, M.; et al. Prognostic Accuracy of Sepsis-3 Criteria for In-Hospital Mortality Among Patients With Suspected Infection Presenting to the Emergency Department. JAMA 2017, 317, 301–308. [Google Scholar] [CrossRef]
  11. Paoli, C.J.; Reynolds, M.A.; Sinha, M.; Gitlin, M.; Crouser, E. Epidemiology and Costs of Sepsis in the United States-An Analysis Based on Timing of Diagnosis and Severity Level. Crit. Care Med. 2018, 46, 1889–1897. [Google Scholar] [CrossRef]
  12. Buchman, T.G.; Simpson, S.Q.; Sciarretta, K.L.; Finne, K.P.; Sowers, N.; Collier, M.; Chavan, S.; Oke, I.; Pennini, M.E.; Santhosh, A.; et al. Sepsis Among Medicare Beneficiaries: 1. The Burdens of Sepsis, 2012–2018. Crit. Care Med. 2020, 48, 276–288. [Google Scholar] [CrossRef] [PubMed]
  13. Arefian, H.; Heublein, S.; Scherag, A.; Brunkhorst, F.M.; Younis, M.Z.; Moerer, O.; Fischer, D.; Hartmann, M. Hospital-related cost of sepsis: A systematic review. J. Infect. 2017, 74, 107–117. [Google Scholar] [CrossRef] [PubMed]
  14. Wentowski, C.; Mewada, N.; Nielsen, N.D. Sepsis in 2018: A review. Anaesth. Intensive Care Med. 2019, 20, 6–13. [Google Scholar] [CrossRef]
  15. Levy, M.M.; Evans, L.E.; Rhodes, A. The Surviving Sepsis Campaign Bundle: 2018 update. Intensive Care Med. 2018, 44, 925–928. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. Gregorowicz, A.J.; Costello, P.G.; Gajdosik, D.A.; Purakal, J.; Pettit, N.N.; Bastow, S.; Ward, M.A. Effect of IV Push Antibiotic Administration on Antibiotic Therapy Delays in Sepsis. Crit. Care Med. 2020, 48, 1175–1179. [Google Scholar] [CrossRef]
  17. Caraballo, C.; Jaimes, F. Focus: Death: Organ dysfunction in sepsis: An ominous trajectory from infection to death. Yale J. Biol. Med. 2019, 92, 629. [Google Scholar]
  18. Seymour, C.W.; Kennedy, J.N.; Wang, S.; Chang, C.C.H.; Elliott, C.F.; Xu, Z.; Berry, S.; Clermont, G.; Cooper, G.; Gomez, H.; et al. Derivation, Validation, and Potential Treatment Implications of Novel Clinical Phenotypes for SepsisDerivation, Validation, and Potential Treatment Implications of Novel Clinical Phenotypes for SepsisDerivation, Validation, and Potential Treatment Implications of Novel Clinical Phenotypes for Sepsis. JAMA 2019, 321, 2003–2017. [Google Scholar] [CrossRef]
  19. Smith, M.E.B.; Chiovaro, J.C.; O’Neil, M.; Kansagara, D.; Quiñones, A.R.; Freeman, M.; Motu’apuaka, M.L.; Slatore, C.G. Early Warning System Scores for Clinical Deterioration in Hospitalized Patients: A Systematic Review. Ann. Am. Thorac. Soc. 2014, 11, 1454–1465. [Google Scholar] [CrossRef]
  20. Dremsizov, T.; Clermont, G.; Kellum, J.A.; Kalassian, K.G.; Fine, M.J.; Angus, D.C. Severe sepsis in community-acquired pneumonia: When does it happen, and do systemic inflammatory response syndrome criteria help predict course? Chest 2006, 129, 968–978. [Google Scholar] [CrossRef] [Green Version]
  21. Knaus, W.A.; Draper, E.A.; Wagner, D.P.; Zimmerman, J.E. APACHE II: A severity of disease classification system. Crit. Care Med. 1985, 13, 818–829. [Google Scholar] [CrossRef] [PubMed]
  22. Le Gall, J.R.; Lemeshow, S.; Saulnier, F. A New Simplified Acute Physiology Score (SAPS II) Based on a European/North American Multicenter Study. JAMA 1993, 270, 2957–2963. [Google Scholar] [CrossRef] [PubMed]
  23. Vincent, J.L.; Moreno, R.; Takala, J.; Willatts, S.; De Mendonça, A.; Bruining, H.; Reinhart, C.K.; Suter, P.M.; Thijs, L.G. The SOFA (Sepsis-related Organ Failure Assessment) score to describe organ dysfunction/failure. Intensive Care Med. 1996, 22, 707–710. [Google Scholar] [CrossRef]
  24. Parlato, M.; Philippart, F.; Rouquette, A.; Moucadel, V.; Puchois, V.; Blein, S.; Bedos, J.P.; Diehl, J.L.; Hamzaoui, O.; Annane, D.; et al. Circulating biomarkers may be unable to detect infection at the early phase of sepsis in ICU patients: The CAPTAIN prospective multicenter cohort study. Intensive Care Med. 2018, 44, 1061–1070. [Google Scholar] [CrossRef] [PubMed]
  25. Gerry, S.; Bonnici, T.; Birks, J.; Kirtley, S.; Virdee, P.S.; Watkinson, P.J.; Collins, G.S. Early warning scores for detecting deterioration in adult hospital patients: Systematic review and critical appraisal of methodology. BMJ 2020, 369, m1501. [Google Scholar] [CrossRef] [PubMed]
  26. Gunčar, G.; Kukar, M.; Notar, M.; Brvar, M.; Černelč, P.; Notar, M.; Notar, M. An application of machine learning to haematological diagnosis. Sci. Rep. 2018, 8, 411. [Google Scholar] [CrossRef] [PubMed]
  27. Rajkomar, A.; Dean, J.; Kohane, I. Machine Learning in Medicine. N. Engl. J. Med. 2019, 380, 1347–1358. [Google Scholar] [CrossRef]
  28. Gupta, A.; Liu, T.; Shepherd, S. Clinical decision support system to assess the risk of sepsis using tree augmented Bayesian networks and electronic medical record data. Health Inform. J. 2020, 26, 841–861. [Google Scholar] [CrossRef] [Green Version]
  29. Medic, G.; Kosaner Kliess, M.; Atallah, L.; Weichert, J.; Panda, S.; Postma, M.; El-Kerdi, A. Evidence-based Clinical Decision Support Systems for the prediction and detection of three disease states in critical care: A systematic literature review. F1000Res 2019, 8, 1728. [Google Scholar] [CrossRef]
  30. Johnson, A.E.W.; Aboab, J.; Raffa, J.D.; Pollard, T.J.; Deliberato, R.O.; Celi, L.A.; Stone, D.J. A Comparative Analysis of Sepsis Identification Methods in an Electronic Database. Crit. Care Med. 2018, 46, 494–499. [Google Scholar] [CrossRef]
  31. Chadaga, K.; Prabhu, S.; Umakanth, S.; Bhat, V.K.; Sampathila, N.; Chadaga, R.P.; Prakasha, K.K. COVID-19 Mortality Prediction among Patients Using Epidemiological Parameters: An Ensemble Machine Learning Approach. Eng. Sci. 2021, 16, 221–233. [Google Scholar] [CrossRef]
  32. Fleuren, L.M.; Klausch, T.L.T.; Zwager, C.L.; Schoonmade, L.J.; Guo, T.; Roggeveen, L.F.; Swart, E.L.; Girbes, A.R.J.; Thoral, P.; Ercole, A.; et al. Machine learning for the prediction of sepsis: A systematic review and meta-analysis of diagnostic test accuracy. Intensive Care Med. 2020, 46, 383–400. [Google Scholar] [CrossRef] [Green Version]
  33. Moor, M.; Rieck, B.; Horn, M.; Jutzeler, C.; Borgwardt, K. Early Prediction of Sepsis in the ICU using Machine Learning: A Systematic Review. Front. Med. 2020, 8, 348. [Google Scholar] [CrossRef] [PubMed]
  34. Ocampo-Quintero, N.; Vidal-Cortés, P.; del Río Carbajo, L.; Fdez-Riverola, F.; Reboiro-Jato, M.; Glez-Peña, D. Enhancing sepsis management through machine learning techniques: A review. Med. Intensiv. 2020, 46, 140–156. [Google Scholar] [CrossRef]
  35. Nemati, S.; Holder, A.; Razmi, F.; Stanley, M.D.; Clifford, G.D.; Buchman, T.G. An Interpretable Machine Learning Model for Accurate Prediction of Sepsis in the ICU. Crit. Care Med. 2018, 46, 547–553. [Google Scholar] [CrossRef] [PubMed]
  36. Kam, H.J.; Kim, H.Y. Learning representations for the early detection of sepsis with deep neural networks. Comput. Biol. Med. 2017, 89, 248–255. [Google Scholar] [CrossRef] [PubMed]
  37. van Wyk, F.; Khojandi, A.; Mohammed, A.; Begoli, E.; Davis, R.L.; Kamaleswaran, R. A minimal set of physiomarkers in continuous high frequency data streams predict adult sepsis onset earlier. Int. J. Med. Inf. 2019, 122, 55–62. [Google Scholar] [CrossRef] [PubMed]
  38. Islam, M.M.; Nasrin, T.; Walther, B.A.; Wu, C.C.; Yang, H.C.; Li, Y.C. Prediction of sepsis patients using machine learning approach: A meta-analysis. Comput. Methods Programs Biomed. 2019, 170, 1–9. [Google Scholar] [CrossRef]
  39. Nesaragi, N.; Patidar, S.; Thangaraj, V. A correlation matrix-based tensor decomposition method for early prediction of sepsis from clinical data. Biocybern. Biomed. Eng. 2021, 41, 1013–1024. [Google Scholar] [CrossRef]
  40. Vincent, J.L.; Opal, S.M.; Marshall, J.C.; Tracey, K.J. Sepsis definitions: Time for change. Lancet 2013, 381, 774–775. [Google Scholar] [CrossRef] [Green Version]
  41. Calvert, J.S.; Price, D.A.; Chettipally, U.K.; Barton, C.W.; Feldman, M.D.; Hoffman, J.L.; Jay, M.; Das, R. A computational approach to early sepsis detection. Comput. Biol. Med. 2016, 74, 69–73. [Google Scholar] [CrossRef] [Green Version]
  42. Mao, Q.; Jay, M.; Hoffman, J.L.; Calvert, J.; Barton, C.; Shimabukuro, D.; Shieh, L.; Chettipally, U.; Fletcher, G.; Kerem, Y.; et al. Multicentre validation of a sepsis prediction algorithm using only vital sign data in the emergency department, general ward and ICU. BMJ Open 2018, 8, e017833. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  43. Fleischmann-Struzek, C.; Thomas-Rüddel, D.O.; Schettler, A.; Schwarzkopf, D.; Stacke, A.; Seymour, C.W.; Haas, C.; Dennler, U.; Reinhart, K. Comparing the validity of different ICD coding abstraction strategies for sepsis case identification in German claims data. PLoS ONE 2018, 13, e0198847. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  44. Bouza, C.; Lopez-Cuadrado, T.; Amate-Blanco, J. Use of explicit ICD9-CM codes to identify adult severe sepsis: Impacts on epidemiological estimates. Crit. Care 2016, 20, 313. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  45. Shappell, C.N.; Klompas, M.; Rhee, C. Surveillance strategies for tracking sepsis incidence and outcomes. J. Infect. Dis. 2020, 222, S74–S83. [Google Scholar] [CrossRef]
  46. Scherpf, M.; Gräßer, F.; Malberg, H.; Zaunseder, S. Predicting sepsis with a recurrent neural network using the MIMIC III database. Comput. Biol. Med. 2019, 113, 103395. [Google Scholar] [CrossRef] [PubMed]
  47. Desautels, T.; Calvert, J.; Hoffman, J.; Jay, M.; Kerem, Y.; Shieh, L.; Shimabukuro, D.; Chettipally, U.; Feldman, M.D.; Barton, C.; et al. Prediction of Sepsis in the Intensive Care Unit With Minimal Electronic Health Record Data: A Machine Learning Approach. JMIR Med. Inf. 2016, 4, e28. [Google Scholar] [CrossRef]
  48. Shashikumar, S.P.; Josef, C.; Sharma, A.; Nemati, S. DeepAISE—An Interpretable and Recurrent Neural Survival Model for Early Prediction of Sepsis. Artif. Intell. Med. 2021, 113, 102036. [Google Scholar] [CrossRef]
  49. Barton, C.; Chettipally, U.; Zhou, Y.; Jiang, Z.; Lynn-Palevsky, A.; Le, S.; Calvert, J.; Das, R. Evaluation of a machine learning algorithm for up to 48-hour advance prediction of sepsis using six vital signs. Comput. Biol. Med. 2019, 109, 79–84. [Google Scholar] [CrossRef]
  50. Schinkel, M.; Paranjape, K.; Nannan Panday, R.S.; Skyttberg, N.; Nanayakkara, P.W.B. Clinical applications of artificial intelligence in sepsis: A narrative review. Comput. Biol. Med. 2019, 115, 103488. [Google Scholar] [CrossRef]
  51. Moons, K.G.; Wolff, R.F.; Riley, R.D.; Whiting, P.F.; Westwood, M.; Collins, G.S.; Reitsma, J.B.; Kleijnen, J.; Mallett, S. PROBAST: A Tool to Assess Risk of Bias and Applicability of Prediction Model Studies: Explanation and Elaboration. Ann. Intern. Med. 2019, 170, W1–W33. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  52. Beam, A.L.; Manrai, A.K.; Ghassemi, M. Challenges to the Reproducibility of Machine Learning Models in Health Care. JAMA 2020, 323, 305–306. [Google Scholar] [CrossRef] [PubMed]
  53. Guidi, J.L.; Clark, K.; Upton, M.T.; Faust, H.; Umscheid, C.A.; Lane-Fall, M.B.; Mikkelsen, M.E.; Schweickert, W.D.; Vanzandbergen, C.A.; Betesh, J.; et al. Clinician Perception of the Effectiveness of an Automated Early Warning and Response System for Sepsis in an Academic Medical Center. Ann. Am. Thorac. Soc. 2015, 12, 1514–1519. [Google Scholar] [CrossRef]
  54. Ginestra, J.C.; Giannini, H.M.; Schweickert, W.D.; Meadows, L.; Lynch, M.J.; Pavan, K.; Chivers, C.J.; Draugelis, M.; Donnelly, P.J.; Fuchs, B.D.; et al. Clinician Perception of a Machine Learning-Based Early Warning System Designed to Predict Severe Sepsis and Septic Shock. Crit. Care Med. 2019, 47, 1477–1484. [Google Scholar] [CrossRef] [PubMed]
  55. Topiwala, R.; Patel, K.; Twigg, J.; Rhule, J.; Meisenberg, B. Retrospective Observational Study of the Clinical Performance Characteristics of a Machine Learning Approach to Early Sepsis Identification. Crit. Care Explor. 2019, 1, e0046. [Google Scholar] [CrossRef]
  56. Guillén, J.; Liu, J.; Furr, M.; Wang, T.; Strong, S.; Moore, C.C.; Flower, A.; Barnes, L.E. Predictive models for severe sepsis in adult ICU patients. In Proceedings of the 2015 Systems and Information Engineering Design Symposium, Charlottesville, VA, USA, 24 April 2015; pp. 182–187. [Google Scholar] [CrossRef]
  57. Delahanty, R.J.; Alvarez, J.; Flynn, L.M.; Sherwin, R.L.; Jones, S.S. Development and evaluation of a machine learning model for the early identification of patients at risk for sepsis. Ann. Emerg. Med. 2019, 73, 334–344. [Google Scholar] [CrossRef] [PubMed]
  58. Bloch, E.; Rotem, T.; Cohen, J.; Singer, P.; Aperstein, Y. Machine learning models for analysis of vital signs dynamics: A case for sepsis onset prediction. J. Healthc. Eng. 2019, 2019, 5930379. [Google Scholar] [CrossRef] [Green Version]
  59. Moor, M.; Horn, M.; Rieck, B.; Roqueiro, D.; Borgwardt, K. Early recognition of sepsis with Gaussian process temporal convolutional networks and dynamic time warping. In Proceedings of the Machine Learning for Healthcare Conference, Ann Arbor, MI, USA, 8–10 August 2019; pp. 2–26. [Google Scholar]
  60. Futoma, J.; Hariharan, S.; Heller, K. Learning to detect sepsis with a multitask Gaussian process RNN classifier. In Proceedings of the International Conference on Machine Learning, Sydney, Australia, 6–11 August 2017; pp. 1174–1182. [Google Scholar]
  61. Kok, C.; Jahmunah, V.; Oh, S.L.; Zhou, X.; Gururajan, R.; Tao, X.; Cheong, K.H.; Gururajan, R.; Molinari, F.; Acharya, U.R. Automated prediction of sepsis using temporal convolutional network. Comput. Biol. Med. 2020, 127, 103957. [Google Scholar] [CrossRef]
  62. Shashikumar, S.P.; Stanley, M.D.; Sadiq, I.; Li, Q.; Holder, A.; Clifford, G.D.; Nemati, S. Early sepsis detection in critical care patients using multiscale blood pressure and heart rate dynamics. J. Electrocardiol. 2017, 50, 739–743. [Google Scholar] [CrossRef]
  63. Johnson, A.E.; Stone, D.J.; Celi, L.A.; Pollard, T.J. The MIMIC Code Repository: Enabling reproducibility in critical care research. J. Am. Med. Inf. Assoc. 2018, 25, 32–39. [Google Scholar] [CrossRef] [Green Version]
  64. Johnson, A.E.W.; Pollard, T.J.; Shen, L.; Lehman, L.w.H.; Feng, M.; Ghassemi, M.; Moody, B.; Szolovits, P.; Anthony Celi, L.; Mark, R.G. MIMIC-III, a freely accessible critical care database. Sci. Data 2016, 3, 160035. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  65. Zhang, S. Nearest neighbor selection for iteratively kNN imputation. J. Syst. Softw. 2012, 85, 2541–2552. [Google Scholar] [CrossRef]
  66. Tseng, J.; Nugent, K. Utility of the shock index in patients with sepsis. Am. J. Med Sci. 2015, 349, 531–535. [Google Scholar] [CrossRef] [PubMed]
  67. Gyawali, B.; Ramakrishna, K.; Dhamoon, A.S. Sepsis: The evolution in definition, pathophysiology, and management. SAGE Open Med. 2019, 7, 2050312119835043. [Google Scholar] [CrossRef] [Green Version]
  68. Guyon, I.; Gunn, S.; Nikravesh, M.; Zadeh, L.A. Feature Extraction: Foundations and Applications; Springer: Berlin/Heidelberg, Germany, 2008; Volume 207. [Google Scholar]
  69. Saeys, Y.; Inza, I.; Larranaga, P. A review of feature selection techniques in bioinformatics. Bioinformatics 2007, 23, 2507–2517. [Google Scholar] [CrossRef] [Green Version]
  70. Remeseiro, B.; Bolon-Canedo, V. A review of feature selection methods in medical applications. Comput. Biol. Med. 2019, 112, 103375. [Google Scholar] [CrossRef]
  71. Frank, E.; Hall, M.A. Data Mining: Practical Machine Learning Tools and Techniques; Morgan Kaufmann: Burlington, MA, USA, 2011. [Google Scholar]
  72. Hameed, B.Z.; Shah, M.; Naik, N.; Singh Khanuja, H.; Paul, R.; Somani, B.K. Application of Artificial Intelligence-based classifiers to predict the outcome measures and stone-free status following percutaneous nephrolithotomy for staghorn calculi: Cross-validation of data and estimation of accuracy. J. Endourol. 2021, 35, 1307–1313. [Google Scholar] [CrossRef]
  73. Peng, H.; Long, F.; Ding, C. Feature selection based on mutual information criteria of max-dependency, max-relevance, and min-redundancy. IEEE Trans. Pattern Anal. Mach. Intell. 2005, 27, 1226–1238. [Google Scholar] [CrossRef]
  74. Robnik-Šikonja, M.; Kononenko, I. Theoretical and empirical analysis of ReliefF and RReliefF. Mach. Learn. 2003, 53, 23–69. [Google Scholar] [CrossRef] [Green Version]
  75. Vapnik, V. The Nature of Statistical Learning Theory; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2013. [Google Scholar]
  76. Mitchell, T.M.; Learning, M. Mcgraw-hill science. Engineering/Math 1997, 1, 27. [Google Scholar]
  77. Berrar, D. Bayes’ Theorem and Naive Bayes Classifier. In Encyclopedia of Bioinformatics and Computational Biology; Ranganathan, S., Gribskov, M., Nakai, K., Schönbach, C., Eds.; Academic Press: Oxford, UK, 2019; pp. 403–412. [Google Scholar] [CrossRef]
  78. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef] [Green Version]
  79. Freund, Y.; Schapire, R.E. A decision-theoretic generalization of on-line learning and an application to boosting. J. Comput. Syst. Sci. 1997, 55, 119–139. [Google Scholar] [CrossRef] [Green Version]
  80. Witten, I.H.; Frank, E.; Hall, M.A. Chapter 8 - Ensemble Learning. In Data Mining: Practical Machine Learning Tools and Techniques, 3rd ed.; Witten, I.H., Frank, E., Hall, M.A., Eds.; The Morgan Kaufmann Series in Data Management Systems; Morgan Kaufmann: Boston, MA, USA, 2011; pp. 351–373. [Google Scholar] [CrossRef]
  81. Wolpert, D.H. Stacked generalization. Neural Netw. 1992, 5, 241–259. [Google Scholar] [CrossRef]
  82. Chen, T.; Guestrin, C. Xgboost: A scalable tree boosting system. In Proceedings of the 22nd Acm Sigkdd International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 785–794. [Google Scholar]
  83. Błaszczyński, J.; Stefanowski, J. Neighbourhood sampling in bagging for imbalanced data. Neurocomputing 2015, 150, 529–542. [Google Scholar] [CrossRef]
  84. Galar, M.; Fernandez, A.; Barrenechea, E.; Bustince, H.; Herrera, F. A review on ensembles for the class imbalance problem: Bagging-, boosting-, and hybrid-based approaches. IEEE Trans. Syst. Man, Cybern. Part C 2011, 42, 463–484. [Google Scholar] [CrossRef]
  85. Flach, P. Machine Learning: The Art and Science of Algorithms That Make Sense of Data; Cambridge University Press: Cambridge, UK, 2012. [Google Scholar]
  86. Demšar, J.; Curk, T.; Erjavec, A.; Gorup, Č.; Hočevar, T.; Milutinovič, M.; Možina, M.; Polajnar, M.; Toplak, M.; Starič, A.; et al. Orange: Data Mining Toolbox in Python. J. Mach. Learn. Res. 2013, 14, 2349–2353. [Google Scholar]
  87. Bisong, E. Google colaboratory. In Building Machine Learning and Deep Learning Models on Google Cloud Platform; Springer: Berlin/Heidelberg, Germany, 2019; pp. 59–64. [Google Scholar]
  88. Lundberg, S.M.; Erion, G.G.; Lee, S.I. Consistent individualized feature attribution for tree ensembles. arXiv 2018, arXiv:1802.03888. [Google Scholar]
  89. Lundberg, S.M.; Erion, G.; Chen, H.; DeGrave, A.; Prutkin, J.M.; Nair, B.; Katz, R.; Himmelfarb, J.; Bansal, N.; Lee, S.I. From local explanations to global understanding with explainable AI for trees. Nat. Mach. Intell. 2020, 2, 56–67. [Google Scholar] [CrossRef]
Figure 1. Architecture of the proposed methodology.
Figure 1. Architecture of the proposed methodology.
Electronics 11 01507 g001
Figure 2. Representation of the labeling method.
Figure 2. Representation of the labeling method.
Electronics 11 01507 g002
Figure 3. Inclusion and exclusion criteria for patients.
Figure 3. Inclusion and exclusion criteria for patients.
Electronics 11 01507 g003
Figure 4. Sepsis-3 suspected infection cohort.
Figure 4. Sepsis-3 suspected infection cohort.
Electronics 11 01507 g004
Figure 5. Sepsis-3 onset hourly computation.
Figure 5. Sepsis-3 onset hourly computation.
Electronics 11 01507 g005
Figure 6. Feature selection process.
Figure 6. Feature selection process.
Electronics 11 01507 g006
Figure 7. Prediction methodology.
Figure 7. Prediction methodology.
Electronics 11 01507 g007
Figure 8. Stacking learning model.
Figure 8. Stacking learning model.
Electronics 11 01507 g008
Figure 9. Detailed information about the training and test sets data samples.
Figure 9. Detailed information about the training and test sets data samples.
Electronics 11 01507 g009
Figure 10. Confusion matrices of four models.
Figure 10. Confusion matrices of four models.
Electronics 11 01507 g010
Figure 11. AUC-ROC curves comparison based on information gain sets.
Figure 11. AUC-ROC curves comparison based on information gain sets.
Electronics 11 01507 g011
Figure 12. Feature importance for sepsis prediction using XGBoost algorithm.
Figure 12. Feature importance for sepsis prediction using XGBoost algorithm.
Electronics 11 01507 g012
Figure 13. SHAP bee swarm plot for XGBoost model.
Figure 13. SHAP bee swarm plot for XGBoost model.
Electronics 11 01507 g013
Figure 14. Summary comparison of metrics results of four ensemble models vs. SOFA and qSOFA using t = 1 h and look back features = 24 h.
Figure 14. Summary comparison of metrics results of four ensemble models vs. SOFA and qSOFA using t = 1 h and look back features = 24 h.
Electronics 11 01507 g014
Table 1. Summary of characteristics of the populations included in the dataset.
Table 1. Summary of characteristics of the populations included in the dataset.
VariableSepsisNo Sepsis
n5371840
Diabetes (n,%)147 (27.3%)469 (86.2%)
Cancer (n,%)23 (4.3%)76 (4.1%)
Died (n,%)39 (7.41%)124 (6.6%)
Elixhauser comorbidity (Index)3.011.52
Age (years, mean)62.0762.5
Weight (kg)80.979.6
Days length of stay (mean)5.92.17
Table 2. Final feature subset.
Table 2. Final feature subset.
Physiological features
Heart rateSystolic blood pressureSpO 2
Diastolic blood pressureTemperature (Celsius)Mean blood pressure
Respiratory rate
Laboratory test results
White blood countPlateletLactate
PotassiumGlucoseChloride
HematocritCreatinineHemoglobin
The international normalized ratio (INR)Prothrombin timeBun
Thromboplastin timeBicarbonateBilirubin
PhCreatininePCO2
Anion gapOliguria
Demographics/Score
AgeElixhauser Comorbidity IndexWeight
Glasgow coma score
Feature engineered
Shock index
FE1: (Age × Min-Systolic Blood Pressure(min))FE2: (Age × Systolic Blood Pressure(mean))FE3: (Age × Systolic Blood Pressure(max))
Table 3. Confusion matrix.
Table 3. Confusion matrix.
Predicted
PositiveNegative
TruePositiveTPFN
NegativeFPTN
Table 4. Description of ensemble models developed.
Table 4. Description of ensemble models developed.
EnsembleLabelModels
E1RF + SVM
E2ANN + AdaBoost
E3RF + ANN
E4SVM + ANN
StackE5SVM + ADA + ANN + KNN + RF + Tree
E6ANN + KNN + RF + SVM + Tree
Table 5. Model hyperparameter used.
Table 5. Model hyperparameter used.
AlgorithmHyperparameter
KNNNeighbors = 7, Metric = Euclidean, Weight = Uniform
Random ForestNumber of Trees = 300, Number of attributes considered at each split = 10
Neural NetworkLayers = 4, Number of neurons in hidden Layers = (40, 20, 10, 5), Solver = Adams,
Activation = tah, Maximal number of iteration = 1000
AdaboostBase estimator = Tree, Number of estimator = 100, Learning rate = 1.0,
Fixed seed for random generator = 5, Classification algorithm = SAMME.R, Regression loss function = Linear
TreeMin number of instances in leaves = 4, Max tree depth = 100
Support Vector MachineCost = 1.60, Regression loss epsilon = 0.40, Kernel = linear, numerical tolerance = 0.0010
Table 6. Comparison of models results based on feature selection techniques using t = 1 h and look back 24 h.
Table 6. Comparison of models results based on feature selection techniques using t = 1 h and look back 24 h.
ModelFeature Selection TechniqueAUC-ROCAccuracyRecallPrecision
E1 0.9020.8700.7690.779
E2 0.8240.7250.6750.0.613
E3 0.9030.7350.6350.642
E4 0.8110.6050.5700.585
E5 0.8950.8630.8390.839
E6 0.8930.8730.8130.819
XGBoostInformation Gain > 0.0020.9180.8720.8520.868
Random Forest 0.9050.8660.8520.860
ANN 0.7530.8040.7340.795
Naïve Bayes 0.7800.7920.7190.779
KNN 0.7210.7960.7190.769
SVM 0.2310.4590.4190.664
E1 0.7540.8100.7140.797
E2 0.7240.5890.5730.561
E3 0.7570.8090.7880.782
E4 0.7380.5870.5630.573
E5 0.7590.8070.7030.739
E6 0.7580.8090.7400.644
XGBoostGINI > 0.0020.7550.7950.7870.770
Random Forest 0.7520.8120.8020.807
ANN 0.6430.7430.7340.726
Naïve Bayes 0.7260.7690.7190.758
KNN 0.6690.7850.7190.750
SVM 0.5170.4310.4190.668
E1 0.9040.8740.8520.870
E2 0.8130.8220.8130.790
E3 0.9060.8650.8130.803
E4 0.8020.8130.8010.8141
E5 0.8900.8680.8120.824
E6 0.9090.8370.8140.812
XGBoostReflief > 0.0020.9090.8560.8490.832
Random Forest 0.7520.8120.8020.807
ANN 0.7820.7900.7340.791
Naïve Bayes 0.7990.7900.690.763
KNN 0.7440.7940.7740.766
SVM 0.5430.4250.4190.677
Table 7. Detailed results of model evaluation using 24 h of look back data and 1, 2, and 3 h prediction time.
Table 7. Detailed results of model evaluation using 24 h of look back data and 1, 2, and 3 h prediction time.
Look Back = 24 h
t = 1 t = 2 t = 3
ModelAccuracyRecallAUC-ROCAccuracyRecallAUC-ROCAccuracyRecallAUC-ROC
E10.8590.6450.7730.8530.6180.7780.8510.6180.765
E20.8110.6050.7350.8110.6050.7340.8110.6050.735
E30.8030.5670.7730.8130.5210.7070.8140.5190.708
E40.8110.6050.6930.7950.5110.6920.7950.5140.694
E50.8610.5630.9070.8590.5590.9060.8620.5630.907
E60.8620.5700.9090.8620.5610.9100.8640.5740.910
XGBoost0.8640.6870.9190.8560.6680.9160.8470.6460.911
SVM0.4110.7120.5060.4110.7120.5060.4110.7120.506
AdaBoost0.8110.6000.7350.8110.6000.7350.8110.6000.735
RF0.9080.5320.9080.8580.5380.9080.8620.5400.909
ANN0.7950.5060.8170.7950.5060.8170.7950.5060.817
KNN0.7800.4450.7680.7800.4450.7680.8700.4450.768
Table 8. Detailed results of model evaluation using 12 h of look back data and 1, 2, and 3 h prediction time.
Table 8. Detailed results of model evaluation using 12 h of look back data and 1, 2, and 3 h prediction time.
t = 1 t = 2 t = 3
ModelAccuracyRecallAUC-ROCAccuracyRecallAUC-ROCAccuracyRecallAUC-ROC
E10.8480.6710.6600.8310.5670.6210.8210.5250.618
E20.8090.5790.7250.8010.5480.7090.7940.5320.674
E30.7960.5250.6960.7930.4910.6830.7970.5040.672
E40.7860.4800.6740.7860.4740.6740.7880.4870.668
E50.8530.5320.8890.8370.4900.8880.8290.4400.880
E60.8560.5330.9020.8860.4880.8860.8290.5400.880
XGBoost0.8500.6650.9120.8510.6430.9070.8380.6140.896
SVM0.3470.7140.4670.3990.7600.5190.3820.7600.521
AdaBoost0.8090.5570.7250.8010.5470.7090.7930.5310.699
RF0.8550.5270.9020.8420.4950.8870.8310.4480.883
ANN0.7790.5130.7850.7840.4830.8060.7840.5130.811
KNN0.7860.3910.7620.7880.3730.7760.7730.4030.769
Table 9. Detailed results of model evaluation using 6 h of look back data and 1, 2, and 3 h prediction time.
Table 9. Detailed results of model evaluation using 6 h of look back data and 1, 2, and 3 h prediction time.
t = 1 t = 2 t = 3
ModelAccuracyRecallAUC-ROCAccuracyRecallAUC-ROCAccuracyRecallAUC-ROC
E10.8480.6480.7720.8200.5270.7380.8110.4990.729
E20.7960.5730.7160.7890.5370.6990.7890.5370.695
E30.8000.4850.6480.7990.5040.6950.7910.4430.663
E40.7840.4720.6690.7840.5140.6870.7770.4190.646
E50.8530.4940.8830.8290.4390.8800.8570.3730.857
E60.8570.5520.8870.8860.4880.8860.8180.3520.856
XGBoost0.8610.6590.9000.8450.6210.8930.8400.6040.886
SVM0.3670.7920.5260.3820.7600.5130.3560.7830.488
AdaBoost0.7940.4530.7760.7930.5310.6990.7880.5310.695
RF0.8550.5260.8880.8270.4360.8820.8220.4500.865
ANN0.7930.4820.7890.7840.5130.8110.7740.4400.787
KNN0.7970.3700.7790.7730.4030.7690.7720.3570.764
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Camacho-Cogollo, J.E.; Bonet, I.; Gil, B.; Iadanza, E. Machine Learning Models for Early Prediction of Sepsis on Large Healthcare Datasets. Electronics 2022, 11, 1507. https://doi.org/10.3390/electronics11091507

AMA Style

Camacho-Cogollo JE, Bonet I, Gil B, Iadanza E. Machine Learning Models for Early Prediction of Sepsis on Large Healthcare Datasets. Electronics. 2022; 11(9):1507. https://doi.org/10.3390/electronics11091507

Chicago/Turabian Style

Camacho-Cogollo, Javier Enrique, Isis Bonet, Bladimir Gil, and Ernesto Iadanza. 2022. "Machine Learning Models for Early Prediction of Sepsis on Large Healthcare Datasets" Electronics 11, no. 9: 1507. https://doi.org/10.3390/electronics11091507

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop