Next Article in Journal
Comparing the Robustness of the Structural after Measurement (SAM) Approach to Structural Equation Modeling (SEM) against Local Model Misspecifications with Alternative Estimation Approaches
Previous Article in Journal
A Log-Det Heuristics for Covariance Matrix Estimation: The Analytic Setup
Previous Article in Special Issue
The Missing Indicator Approach for Accelerated Failure Time Model with Covariates Subject to Limits of Detection
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Semiparametric Survival Analysis of 30-Day Hospital Readmissions with Bayesian Additive Regression Kernel Model

1
Department of Statistics, University of Missouri, Columbia, MO 65211, USA
2
Liberty Mutual Group Inc., Boston, MA 02116, USA
3
Center for Surgery and Public Health, Brigham and Women’s Hospital, Harvard Medical School, Boston, MA 02120, USA
*
Author to whom correspondence should be addressed.
Stats 2022, 5(3), 617-630; https://doi.org/10.3390/stats5030038
Submission received: 8 June 2022 / Revised: 6 July 2022 / Accepted: 8 July 2022 / Published: 14 July 2022
(This article belongs to the Special Issue Survival Analysis: Models and Applications)

Abstract

:
In this paper, we introduce a kernel-based nonlinear Bayesian model for a right-censored survival outcome data set. Our kernel-based approach provides a flexible nonparametric modeling framework to explore nonlinear relationships between predictors with right-censored survival outcome data. Our proposed kernel-based model is shown to provide excellent predictive performance via several simulation studies and real-life examples. Unplanned hospital readmissions greatly impair patients’ quality of life and have imposed a significant economic burden on American society. In this paper, we focus our application on predicting 30-day readmissions of patients. Our survival Bayesian additive regression kernel model (survival BARK or sBARK) improves the timeliness of readmission preventive intervention through a data-driven approach.

1. Introduction

The Centers for Medicare and Medicaid Services (CMS) has defined hospital readmission as an admission to an acute care hospital within 30 days of a discharge from the index admission in the same or another acute care hospital [1]. Hospital readmissions have emerged as a global concern due to their high frequency and high associated cost. In the United States, 15.5% of Medicare beneficiaries were readmitted unplanned within 30 days after discharge during July 2014 to June 2015 [2]. It has been estimated that unplanned readmissions account for USD 17.4 billion in Medicare expenditures annually [3]. Although the causes are complicated, unplanned hospital readmissions are frequently seen as related to the substandard quality of care received during index admissions [4,5,6]. The pressure to reduce cost and improve healthcare quality has triggered the implementation of the Hospital Readmissions Reduction Program (HRRP) [1] in the United States in 2012. In the fiscal year 2017, 79% of hospitals were penalized by the HRRP, with estimated penalties of USD 528 million [7].
In recent years, we are seeing a growing body of literature on the development of hospital readmission predictive models for both hospital profiling purposes and clinical decision support, with the ultimate goal of reducing readmissions [8]. The most widely used approach is to treat readmission prediction as a binary classification or a probability regression problem with statistical or machine-learning-based algorithms. However, 30-day hospital readmission data is naturally right-censored, making it inappropriate to directly apply traditional prediction algorithms. A patient’s probability of being readmitted varies with the time after discharge. For example, a patient may be readmitted on the 31st day after discharge. Because it is outside of the monitoring time window (30 days), it is considered as no readmission within 30 days. However, the time is very close to the threshold, making it hard to contrast the readmissions within the window (e.g., 29th day, 30th day, etc.). This is one reason why most readmission models cannot achieve good performance. Therefore, it is more appropriate to study patients’ probability of readmission as a function of time within the observation time window.
Survival analysis has been one of the most important branches of statistics to estimate the duration of time until the occurrence of an event of interest [9]. The key characteristic of the survival analysis problem is the presence of censored instances whose event outcomes cannot be observed at some time point or when the event outcome does not occur within the monitoring time window [10]. As a result, regular statistical and machine learning predictive algorithms cannot be directly applied to survival analysis problems. Based on the assumption about the distribution of time, statistical survival models can be classified as either parametric, nonparametric, or semiparametric. Parametric models assume survival time follows some particular distribution, such as normal, exponential, and Weibull distributions. One popular parametric survival model is the accelerated failure time (AFT) model [11]. The AFT model assumes failure time is accelerated by some constant. For example, it is known that dogs become older seven times faster than human beings, and the probability of one dog surviving t years is equivalent to that of one human surviving 7t years [12]. Nonparametric models assume no distributions for the survival time. The survival function can be empirically estimated by the Kaplan–Meier [13], Nelson–Aalen estimator [14], or Life-Table [15] methods. Semiparametric methods do not require the knowledge of the distribution of the time to event. The Cox regression model [16] is the most popular semiparametric method. Wang et al. (2017) [10] have provided a comprehensive survey of both the statistical and machine-learning-based survival analysis methods.
In recent years, many machine-learning-based survival analysis models have been developed to complement the traditional statistical survival models [10]. Ensemble trees have been one of the most interesting base machine learning algorithm families for survival analysis because of their generally good performance in handling nonlinear relationships. The random survival forest (RSF) is another widely cited random forest model that can work with right-censored data [17]. This work defined a new predictive outcome called ensemble mortality. Bayesian additive regression trees (BART) [18] are a nonparametric “sum-of-trees” model constructed based on a fully Bayesian approach. Sparapani et al. (2016) [19] created a nonparametric survival analysis model in a discrete-time fashion based on the BART model. In their work, the survival analysis problem has been handled by transforming into a regular binary classification problem. Bonato et al. (2011) [20] modified the BART model further based on three classical statistical survival models and created an ensemble-based proportional hazards regression model, an ensemble-based Weibull regression model, and an ensemble-based AFT model.
Kernel learning models are another popular candidate for survival analysis. Support vector machines (SVM) were extended to support survival analysis by regression and ranking approaches [21]. In the regression problem, the censored data is handled by modifying the loss function [22,23]. In addition to that, relevance vector machines (RVM) have been extended to support parametric survival analysis based on the Weibull AFT model [24]. The classical SVM algorithm can only make point predictions without reporting predictive distributions for future observations. As a result, the uncertainty of the predictions is not known. To address this issue, RVM was developed as a probabilistic alternative of SVM [25]. By estimating prior parameters with type II maximum likelihood, RVM can deliver posterior distributions with Bayesian inference. However, this is not a fully Bayesian approach and the inferences might be unreliable. To address this issue, hierarchical Bayes SVM and hierarchical Bayes RVM [26,27] have been developed based on reproducing kernel Hilbert space (RKHS) [28] methodology under a complete Bayesian framework [26,27]. By using a similar approach, Maity et al. (2011) [29] performed a fully Bayesian analysis of proportional hazards (PH) regression based on SVM and RVM by using RKHS. Where the functional form of the baseline hazard function [30] was specified by a nonparametric Bayesian approach. Bayesian additive regression kernels (BARK) is a fully Bayesian kernel learning method, which was developed by using the symmetric α-stable (SαS) Lévy random field to induce the prior distribution [26]. Compared with classical SVM and RVM, BARK does not restrict the number of kernel functions and supports simultaneous feature selection through soft shrinkage and hard shrinkage.
The electronic health record (EHR) is a comprehensive electronic medical history which longitudinally records patients’ demographics, vital signs, medications, laboratory data, procedures, diagnoses, clinical notes, and other information pertaining to hospital encounters [31]. Compared with traditional medical data sets accumulated by recruiting patients, EHR data can provide a larger number of subjects and conditions [32]. Survival analysis has seen many applications in the studies of EHR data because subjects may lose follow-up or die [32].
In this work, a survival kernel learning model (sBARK) was developed based on the BARK (Bayesian additive regression kernel) model to analyze the 30-day hospital readmissions extracted from EHR data. sBARK is a fully Bayesian survival model fitted by MCMC. The survival time is assumed to be based on the covariates and a baseline survival time. The inherent features of BARK, such as the support of feature selection and the unrestricted number of kernels, make sBARK more flexible than other survival kernel learning algorithms. The remaining part of this paper is organized as follows. Section 2 discusses the theoretical background and the fitting method of sBARK. The evaluations on the simulated data and a standard benchmark data are detailed in Section 3. Section 4 is dedicated to solving the 30-day hospital readmission prediction based on our sBARK model. In Section 5, we provide our detailed discussion along with the future possibilities and extensions.

2. Bayesian Additive Regression Kernel Model for Survival Outcome

In this study, we state the survival analysis problem as the prediction of the survival time for a subject i, represented by X i , y i , δ i ,   where   X i is the vector of predictor variables of size p and δ i   is the indicator of whether the instance is censored δ i = 0   or not δ i = 1 [10]. y i   is the observed time and can be represented as   y i =   T i   i f   δ i = 1   C i   i f   δ i = 0   , where   T i   is the survival time for an uncensored instance and C i   is the survival time for a censored instance. The goal is to estimate the time to event T i   for a new instance i given X i . The survival analysis problem can also be understood as the study of survival probability as a function of time. Figure 1 shows an example of right censoring.

2.1. A Quick Background of Bayesian Additive Regression Kernels (BARK)

In this section, we give a quick overview of the BARK method, which is the cornerstone of our kernel-based nonparametric survival analysis model. In the BARK algorithm [26] models, the observed covariates and responses are expressed as the sum of weighted kernel functions y i = j = 1 : J β j K x , χ j + ϵ i , where β j   is the kernel regression coefficient, χ j is the kernel location parameter, and ϵ i   ~   N 0 , ϕ 1 is the independent additive Gaussian noise. The number of kernel functions J is not fixed and needs to be learned. In addition to that, a Gaussian kernel function is adopted with and independent scale parameter λ l ,   l = 1 , , p for each predictor variable. The full BARK model can be written as follows:
y i = j   β j K x i , χ j + ε i
where:
-
{ x i } i = 1 n : observed covariate vectors in R p .
-
{ y i } i = 1 n : responses.
-
β j : regression coefficient.
-
χ j : kernal location parameter in X.
-
ϵ i N 0 , ϕ 1 : independent additive Gaussian noise.
The Gaussian kernel function was used in the original BARK construction [26] by assigning a scale parameter for each predictor variable. The Gaussian kernel function has the following form:
K x , χ = exp l = 1 p λ l x l χ l 2
where:
-
λ l : the scale parameter, which measures the contribution of the lth variable in the kernel function.
-
λ l = 0 : no contribution made by the lth variable through the kernel function.
-
λ l is large: the lth variable is important in the kernel function.

2.1.1. Use of Symmetric α-Stable (SαS) Lévy Random Field as Prior Distribution on the Mean Function f(x)

The unknown mean function f(x) can be represented as a stochastic integral of a kernel function with respect to a random signed measure.
f x = j = 1   β j K x , χ j = R × X K x , χ L d χ
where:
-
L d χ : the random signed measure.
Putting a prior distribution on the mean function f(x) is exactly the same as putting a prior on the above random signed measure. To be noted here, the linear mapping from an L2 (square integrable) function to random variables by a random measure is called a random field. One such specific random field is the symmetric α-stable (SαS) Lévy random field defined by the (SαS) Lévy measure [26] as:
v d β , d χ = γ c α β 1 α d β π χ d χ
where:
-
c α = α π Γ α s i n π α 2 .
-
π χ d χ :   the probability measure on X.
-
0 < α < 2 : the stable index.
-
γ > 0 : the intensity parameter.
The above SαS Lévy random measure was used to induce the joint prior distribution for the number of kernels J , regression coefficient β , and kernel parameter λ . Because the SαS Lévy random measure is not finite, two methods were used for approximation. The first method approximated it by truncation of β with respect to a given threshold ϵ > 0 and had a finite mass as:
v ϵ t + α , γ , ϵ = 2 γ π ϵ α Γ α s i n π α 2
The second method is based on continuous approximation and had a finite mass as:
ν ϵ c + α , γ , ϵ = γ α 1 α / 2 2 1 α ϵ α Γ α / 2 Γ 1 α / 2

2.1.2. Automatic Feature Selection in BARK

Suppose the prior distribution for the sum of all λ l s is Ga a λ , b λ   , which is indefinitely divisible. Here, a λ and b λ are the shape parameter and scale parameter of Gamma distribution, respectively. There are four types of feature selections that are automatically built into the BARK structure, as mentioned in [26]. They are as follows:
  • Equal weights: All kernel scale parameters are the same and follow the same Gamma distribution as:
    λ l = λ Ga α λ , p b λ
  • Different weights: All kernel scale parameters are different, and each follows a different Gamma distribution as:
    λ l = λ Ga α λ p , b λ
  • Equal weights with hard shrinkage and selection: Introduce a vector δ { 0 , 1 } p to indicate if a variable is selected or not. Each indicator follows a Bernoulli distribution. If the indicator is 1, the corresponding λ l = λ   ~   Ga a λ , b λ   .   Note that here all selected kernel scale parameters are the same as in Equal weights.
  • Different weights with hard shrinkage and selection: Exactly the same formulation as in 3, however, each selected kernel scale parameter has its own Gamma distribution λ l   ~   Ga a λ / d , b λ   ,   where d is the number of nonzero kernel scale parameters.

2.1.3. Likelihood and the Posterior Distribution

Under continuous response with normal error explained in the beginning of Section 2.1, our likelihood of the training set can be written as:
p y | ϕ , J , β , χ , λ = ϕ n / 2 2 π n / 2 exp ϕ 2 ( y K β ) 2
where:
-
K: the n × j kernel matrix, with K i , j = K x i , χ j .
Apart from the feature selection prior and the α-stable (SαS) Lévy random field prior, we also put a noninformative prior on ϕ as π   ϕ     1 / ϕ .
By Bayes rule, the full posterior distribution can be written as:
p ϕ , J , β , χ , λ | y = p y | ϕ , J , β , χ , λ π ϕ , J , β , χ , λ | y p y
For a new observation at x*, the posterior predictive distribution for y* can be written as:
p y * | x * , y = p y * | x * , y , ϕ , J , β , χ , λ p ϕ , J , β , χ , λ | y d ϕ d J d β d χ d λ
In the next section, we extend the above-mentioned BARK modeling architecture for a right-censored survival time data set.

2.2. Nonparametric Survival Analysis with BARK

The implementation of nonparametric survival BARK is based on the data modification method discussed in [19]. There is no assumption about the distribution of the time to event and we put no functional restriction on the proportional hazard structure. Here, we handle the survival analysis problem in a discrete-time fashion by converting into a binary classification problem [19], which can then be directly modelled by BARK and the help of latent variable data augmentation.

2.2.1. Data Transformation Method: Converting Right-Censored Data to Binary Classification

Let us consider the survival data with the following form: x i ,   t i ,   δ i ,   i = 1 ,   2 ,   ,   n . Where x i : covariates of subject i ; t i : the event/censoring time of subject i ; and δ i : the indicator of status of subject i : 1—event, 0—right-censoring.
For n subjects, there can be k distinct events and censoring times. Because two different subjects can have the same event or censoring times, k n . Order the k distinct event and censoring times as:
0 < t 1 < t 2 < < t k <
Then, for each subject i , use a new binary indicator y i j to indicate the status at each distinct time t j up to and including the subject’s original observation time t i . In other words, each subject i has a series of binary indicators y i j to indicate its status at all discrete time points earlier than (inclusive) the original observation time t i . That means y i j = 0 for all j before the last indicator as the original observation time t i does not match with the ordered distinct time t j . Note that the last indicator y i j in the series is the original status indicator δ i .
For example, suppose there are three subjects in a dataset as follows: x i ’s are omitted because they are not directly involved in the data transformation.
t 1 ,   δ 1 = 2.2 ,   1 ,       t 2 ,   δ 2 = 0.8 ,   1 ,       t 3 ,   δ 3 = 4.1 ,   0
Order the distinct event/censoring times and obtain:
t 1 = 0.8 ,   t 2 = 2.2 ,   t 3 = 4.1
For the subject 1, its observation time t 1 is the second place t 1 = t 2 = 2.2 in the ordered distinct event and censoring times. It has two binary indicators y 11 and y 12 to represent its status at time t 1 and t 2 : y 11 = 0 ,   y 12 = 1 . For the instance 2, its observation time t 2 is the first place t 2 = t 1 = 0.8 in the ordered distinct event and censoring times. It has one binary indicator y 21 to represent its status at time t 1 : y 21 = 1 . For the instance 3, its observation time t 3 is the third place t 3 = t 3 = 4.1 in the ordered distinct event and censoring times. It has three binary indicators y 31 , y 32 , and y 33 to represent its status at time t 1 ,   t 2 and t 3 : y 31 = 0 ,   y 32 = 0 ,   y 33 = 0 . Note that y 33 = 0 because it is right-censored at t 3 . In this way, the original dataset with three instances can be transformed into a dataset with six records. The time t j is used as a new covariate together with the original covariates x i . Figure 2 shows the full transformation result for this example.

2.2.2. Bayesian Additive Kernel-Based Survival Model for the Transformed Data (sBARK)

Based on our data transformation method discussed before, we transformed our right-censored survival outcome data to binary outcome data. Here, we formulate a BARK model to handle the binary data based on data augmentation by [33,34]. Our hierarchical BARK model with data augmentation is given as below:
y i j p i j   ~   Bernoulli p i j p i j f = Φ μ i j ,       μ i j = μ 0 + f t j ,   x i   f   ~   BARK z i j y i j , f   ~   N μ i j , 1 I , 0   if   y i j = 0 N μ i j , 1 I 0 ,       is   y i j = 1
where:
-
p i j : the probability of an event at time t j given no previous event for instance i .
-
Φ : the probit link function.
-
μ i j : the predicted event status.
-
μ 0 : the baseline, which equals to 0 for centered outcome.
-
z i j : truncated normal latent variable.
-
f : the nonlinear function that connects the binary response with the set of predictors. The f function [35] is modeled nonparametrically using the BARK architecture discussed previously in detail.
Please note that after our data transformation method as described in Section 2.2 and in Figure 2, we can write the model for yij as a nonparametric probit regression of yij on the time t(j) and the covariates xi [36], and then utilize the Albert–Chib (1993) truncated normal latent variables zij to reduce it to the continuous outcome BARK model of Equations (1)–(3) applied to z’s. That is why, in Equation (10), f() is a function of time t(j) and the covariates xi. The f   ~   BARK in Equation (10) means f [37] follows the same BARK architecture as explained in Equation (3) in Section 2.1. However, now the new covariate vector is x* = (t, x).
f x * = j = 1   β j K x * , χ j
The survival function at the discrete time points t j is given by (for details see [38,39,40])
S t j x = P T > t j x = l = 0 j 1 p t l , x
where
p t , x = Φ μ 0 + f t , x
The drawings of posteriors after burning in were used to create the survival function. For the sake of simplicity, we call our above-described BARK survival model sBARK or survival BARK.

2.3. MCMC Algorithm to Fit sBARK

Because the integral of the posterior distribution is analytically intractable, the MCMC method is used to fit our sBARK model. The full MCMC process is shown below:
  • Sample the latent variables z i j conditional on all other parameters from the truncated normal distribution described in (10). The truncation scheme is based on the value of y i j described above in (10). The truncated normal distribution has mean μ ^ i j , which is based on the posterior sample of f ^ i j from BARK output and variance is 1.
  • Update all parameters of BARK architecture with the MCMC process built in the BARK algorithm. The full MCMC scheme to generate the BARK parameters are documented in detail in [26] and in https://github.com/merliseclyde/bark, accessed on 20 November 2019.
  • The zij’s are centered around a known constant µ0, which is equivalent to centering the probabilities pij around p0 = Φ(µ0). By default, we set µ0 = 0, which means pij are centered around 0.5.

2.4. Prior and MCMC Parameter Selection for sBARK

In all our simulation studies and real data analysis in our sBARK model, we set α = 1 ,   γ = 5 ,   ϵ = 0.5 . These choices of prior parameters ensure that our prior is near-diffuse, but the posterior is a proper density. For automatic variable selection, we use the Option 4: different weights with hard shrinkage and selection prior with Bernoulli parameter set to 0.20. In all our simulation studies and real data analysis, we ran MCMC chain with 200,000 iterations and used the first 100,000 as burn in. To reduce autocorrelation, we keep every 5th draw. All our MCMC chain convergences are verified via inspection of trace plots and running the Gelman–Rubin diagnostic [34]. For the survival BART, we used 200 trees with near-diffuse prior settings exactly as suggested in [19]. For survival BART, we generated 200,000 MCMC iterations with the first 100,000 as burn in.

3. Simulation Study and a Benchmark Real Data Analysis

In this section, we conduct a simulation study and we also apply our model to a standard right-censored survival outcome data: the kidney catheter data [27].

3.1. Simulated Data

One simulated data set is created to test sBARK’s performance. The simulation was performed with the simple.surv.sim() function in the “survsim” package [35] in R. In this simulation, 80 cases are generated. The maximum observation time is set to be 365 days. The time to event follows a log-normal distribution. Each subject has seven covariates, with each following a standard normal distribution. The mean function of the log-normal distribution is given by f x = 0.1 x 1 + 0.2   x 2 2 + 0.3 ln x 3 + 0.4 x 4 + 0.5   x 5 3 + 0.6   x 6 + 0.7   x 7 2 + 0.2   x 1 x 2 . The simulated dataset has a total of 12 attributes, including the instance id “nid”, the observation start time “start”, the observation end time “stop”, the status indicator “status” (1: event, 0: censored), the individual heterogeneity “z”, and seven covariates. Of the 80 instances, 20 are censored. We chose to conduct the simulation study with 80, as in a typical data set we are often restricted to small samples. Traditionally, with machine learning models, they always tend to perform better with large data volume, but it is always a challenge working with reasonably small data. So, we showed our model can work well even with a small sample size. In Figure 3, we report the estimated survival function based on popular survival models such as Kaplan–Meier, survival Bayesian additive regression tree (BART), and our survival BARK or sBARK model. We can see clearly that our sBARK is equally effective in estimating the underlying survival probability curve. To be noted here is that the role of our Figure 3 is not to show better performance of sBARK but to validate that based on Equation (11) our sBARK model can also estimate the survival curve. The survival curve produced by our sBARK is like the ones provided by Kaplan–Meier and the survival BART. The point we are trying to make through Figure 3 is that our sBARK was implemented by transforming the survival analysis problem into a regular binary classification problem, which then handled by BARK architecture can retrieve/estimate the underlying survival curve/function as accurately as Kaplan–Meier and the survival BART. In Table 1 we list the root mean square error (RMSE) based on 5-fold cross-validation of our simulated data based on survival BARK, survival BART [19], and the more traditional Cox proportional hazards model [16]. From Table 1, it is clearly seen that our survival BARK model’s root mean square error across all included time points is always lower than its closest competitor, the survival BART. In the BART package/model, which is the main competitor of our sBARK model, the authors used the traditional RMSE formula with the original survival predicted outcome, which can be calculated using our Equation (11). We also adopted the same approach for the calculation in all reported RMSE calculations. The censoring information is implicitly taken into account in the construction of the survival function at the discrete time points. Like in our main competitor survival BART, in our sBARK calculation of RMSE we did not carry out anything additional to it for censored times.

3.2. Benchmark Data Analysis with Kidney Catheter Data

In this section, we demostrate the performance of our survival BART with respect to the kidney catheter data [36]. This data is available in the “survival” package [37] in R and has been widely used as a benchmark. The data set records the recurrence of infection in patients who are using a portable dialysis machine. The infection may occur when the catheter is inserted. In this case, the catheter must be removed from the patient. It can be re-inserted only after the infection has been treated. The recurrence time is the time from the insertion until the next infection. This data is right-censored because the catheter may be removed for other reasons. This data has 76 instances and seven attributes, including patient ID, survival time, status indicator (1: event, 0: censored), age, sex, disease type, and estimated frailty. Of the 76 instances, 18 are censored. Figure 4 shows the survival probability as a function of time (days) obtained by the nonparametric survival BARK, the nonparametric survival BART, and the Kaplan–Meier estimator on the kidney catheter dataset. The Kaplan–Meier method only considers the event time when constructing the survival function curve. The survival BARK and survival BART take both event time and covariates into consideration when estimating the survival probability. Using the Kaplan–Meier curve as a baseline, it can be seen that the survival probability estimated by the nonparametric survival BARK tends to be higher over time. The survival curve by the nonparametric survival BART is closer to the Kaplan–Meier curve, and the survival probability tends to be lower. Moreover, from Table 2, it is clearly demonstrated that our survival BARK has completely outperformed its best contender survival BART and traditional models such as the Cox proportional hazards model.

4. Analysis of 30-Day Hospital Readmission Data

The 30-day hospital readmission data was obtained from Medical Information Mart for Intensive Care (MIMIC III), a public accessible deidentified EHR database of critical care encounters at the Beth Israel Deaconess Medical Center between 2001 and 2012 (https://mimic.mit.edu/ (accessed on 20 November 2019)) [38]. It contains 58,977 encounters of 46,521 unique patients. A total of 43,407 encounters are unplanned (emergency or urgent). To be eligible as a readmission, the encounter needs to meet the following criteria: (1) the patient has to be older than 18 at the time of index admission; (2) the patient does not expire during the index admission or after discharge from the index admission; (3) the gap between index admission and readmission is between one day and 30 days; and (4) ff a patient has more than one eligible index admission–readmission episode, only the latest episode is kept. Of the 43,407 unplanned encounters, 1434 eligible index admission–readmission episodes were identified. Another 1434 admissions with censored outcome (not readmitted within 30 days) were randomly sampled from the remaining unplanned encounters to balance the data set. The data set has 45 attributes including patients’ socio-demographics information (age, sex, religion, marital status, ethnicity, and insurance), index admission type, index admission length of stay, index admission discharge disposition, minimum/mean/maximum of vital signs before discharge (heart rate, systolic blood pressure, diastolic blood pressure, and respiratory rate), and minimum/mean/maximum of laboratory test results before discharge (hemoglobin, serum sodium, magnesium, calcium, glucose, urea, creatinine, platelet count, and white blood cell count). Categorical variables were transformed into dummy variables, and as a result, the number of attributes increased to 85. Cases with missing values and outliers (greater than three times the standard deviation) in any attribute were removed. The cleaned data set has 1228 censored cases and 1307 uncensored cases.
In Table 3, we summarized our findings with respect to our survival BARK model and its competitors. In terms of out-of-sample prediction from Table 3, we observe that our survival BARK does at least 2 times better than its competitors in an early prediction model of readmission by incorporating detailed medical history such as diagnoses, procedures, medications, lab test results, vital signs, and healthcare utilization. In Figure 5, we report the posterior probability plot for each of the predictors for the purpose of variable selection. In this paper, we adopted a posterior probability threshold of 0.50, and from Figure 5 we see that out of 85, 10 variables are selected based on the 0.50 probability threshold. Based on our survival BARK model, the 10 selected predictors for 30-day hospital readmission are minimum heart rate, minimum respiratory rate, minimum urea content, Medicaid status, Medicare status, marital status, white race, long-term care status, age group 50 to 64, and age group 65 to 79.

5. Conclusions

Modelling of 30-day hospital readmission has been a popular research topic in recent years. Although survival analysis is more appropriate than traditional binary classification methods in this scenario, very few studies have treated readmission prediction as a survival analysis problem. In this paper, we developed a kernel-based machine learning method for readmission predictions. Our developed survival BARK or sBARK model supports survival analysis based on flexible kernel machines. The sBARK, which is based on a kernel machine learning model, has the flexibility of nonlinear methods. Our survival BARK has been implemented in a discrete-time fashion. The dataset is restructured to transform the survival analysis problem into a regular binary classification problem, which can be handled by BARK. In the restructured data, the binary response variable indicates the status at each distinct time up to and including the instance’s original observation time. The distinct discrete time points before and including the instance’s observation time are used as a new predictor attribute in BARK. In addition to that, Equation (11) provides a direct way to calculate the survival curves based on our sBARK model. The survival probability curves of the two datasets estimated by survival BARK and the survival BART have similar trends. In all simulation studies and benchmark data analyses, and on the real-life 30-day hospital readmission data, our survival BARK has convincingly outperformed its other nonlinear tree-based competitor survival BART with respect to prediction performance.
BARK can also be extended based on other survival analysis models, such as the parametric Weibull model and the semiparametric Cox proportional hazards model. Moreover, there is huge potential to make our survival BARK work with big data scenarios such as electronic health record data, which is in the order of hundreds of millions. However, to make our survival BARK model scalable for such big data sets, we need to modify the sBARK algorithm to optimize its time and space complexity. Along with that, the searching and ordering steps also require to be further optimized. It is also possible to parallelize the computations by taking advantage of multicore processors. It can also be ported to the Spark environment to allow for more flexible computing resource management. These are some of our potential directions of work for future extensions.

Author Contributions

Conceptualization, S.C. and P.Z.; methodology, S.C.; software, P.Z.; validation, S.C., T.D. and P.Z.; formal analysis, S.C. and P.Z.; data curation, P.Z.; writing—original draft preparation, S.C., P.Z., Y.H. and T.D.; writing—review and editing, S.C., P.Z., Y.H. and T.D.; visualization, S.C., P.Z., Y.H. and T.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Informed Consent Statement

Not Applicable.

Data Availability Statement

https://mimic.mit.edu/ (accessed on 7 June 2022).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Readmissions Reduction Program (HRRP). Available online: https://www.cms.gov/medicare/medicare-fee-for-service-payment/acuteinpatientpps/readmissions-reduction-program.html (accessed on 20 November 2019).
  2. Hospital Quality Initiative—Outcome Measures 2016 Chartbook. Available online: https://www.cms.gov/Medicare/Quality-Initiatives-Patient-Assessment-Instruments/HospitalQualityInits/OutcomeMeasures.html (accessed on 20 November 2019).
  3. Jencks, S.F.; Williams, M.V.; Coleman, E.A. Rehospitalizations among Patients in the Medicare Fee-for-Service Program. N. Engl. J. Med. 2009, 360, 1418–1428. [Google Scholar] [CrossRef] [PubMed]
  4. Weissman, J.S.; Ayanian, J.Z.; Chasan-Taber, S.; Sherwood, M.J.; Roth, C.; Epstein, A.M. Hospital Readmissions and Quality of Care. Med. Care 1999, 37, 490–501. [Google Scholar] [CrossRef] [PubMed]
  5. Polanczyk, C.A.; Newton, C.; Dec, G.W.; Di Salvo, T.G. Quality of care and hospital readmission in congestive heart failure: An explicit review process. J. Card. Fail. 2001, 7, 289–298. [Google Scholar] [CrossRef]
  6. Luthi, J.-C.; Lund, M.J.; Sampietro-Colom, L.; Kleinbaum, D.G.; Ballard, D.J.; McClellan, W.M. Readmissions and the quality of care in patients hospitalized with heart failure. Int. J. Qual. Health Care 2003, 15, 413–421. [Google Scholar] [CrossRef] [PubMed]
  7. Boccuti, C.; Casillas, G. Aiming for Fewer Hospital U-turns: The Medicare Hospital Readmission Reduction Program. Policy Brief. 2017, 1–10. [Google Scholar]
  8. Kansagara, D.; Englander, H.; Salanitro, A.; Kagen, D.; Theobald, C.; Freeman, M.; Kripalani, S. Risk prediction models for hospital readmission: A systematic review. JAMA 2011, 306, 1688–1698. [Google Scholar] [CrossRef]
  9. Miller, R.G. Survival Analysis; John Wiley & Sons: Hoboken, NJ, USA, 1997. [Google Scholar]
  10. Wang, P.; Li, Y.; Reddy, C.K. Machine Learning for Survival Analysis: A Survey. arXiv 2017, arXiv:1708.04649. Available online: https://arxiv.org/abs/1708.04649 (accessed on 28 February 2020).
  11. Klein, J.P.; Moeschberger, M.L. Survival Analysis: Techniques for Censored and Truncated Data, 2nd ed.; Springer: New York, NY, USA, 2003; Volume 9, pp. 302–308. [Google Scholar]
  12. Dätwyler, C.; Stucki, T. Parametric Survival Models. 2011. Available online: http://stat.ethz.ch/education/semesters/ss2011/seminar/contents/handout_9.pdf (accessed on 15 February 2020).
  13. Kaplan, E.L.; Meier, P. Nonparametric Estimation from Incomplete Observations. J. Am. Stat. Assoc. 1958, 53, 457. [Google Scholar] [CrossRef]
  14. Andersen, P.K.; Borgan, O.; Gill, R.D.; Keiding, N. Statistical Models Based on Counting Processes; Springer: Berlin/Heidelberg, Germany, 1993. [Google Scholar]
  15. Cutler, S.J.; Ederer, F. Maximum Utilization of the Life Table Method in Analyzing Survival. In Annals of Life Insurance Medicine; Springer: Berlin/Heidelberg, Germany, 1964; pp. 9–22. [Google Scholar]
  16. Cox, D.R. Regression Models and Life-Tables. J. R. Stat. Soc. Ser. B 1972, 34, 187–220. [Google Scholar] [CrossRef]
  17. Ishwaran, H.; Kogalur, U.B.; Blackstone, E.H.; Lauer, M.S. Random survival forests. Ann. Appl. Stat. 2008, 2, 841–860. [Google Scholar] [CrossRef] [PubMed]
  18. Chipman, H.A.; George, E.I.; McCulloch, R.E. BART: Bayesian additive regression trees. Ann. Appl. Stat. 2012, 6, 266–298. [Google Scholar] [CrossRef]
  19. Sparapani, R.; Logan, B.R.; McCulloch, R.E.; Laud, P.W. Nonparametric survival analysis using Bayesian Additive Regression Trees (BART). Stat. Med. 2016, 35, 2741–2753. [Google Scholar] [CrossRef] [PubMed]
  20. Bonato, V.; Baladandayuthapani, V.; Broom, B.M.; Sulman, E.P.; Aldape, K.D.; Do, K.-A.; Anh, K. Bayesian ensemble methods for survival prediction in gene expression data. Bioinformatics 2010, 27, 359–367. [Google Scholar] [CrossRef] [PubMed]
  21. Van Belle, V.; Pelckmans, K.; Van Huffel, S.; Suykens, J.A.K. Support vector methods for survival analysis: A comparison between ranking and regression approaches. Artif. Intell. Med. 2011, 53, 107–118. [Google Scholar] [CrossRef] [PubMed]
  22. Khan, F.M.; Zubek, V.B. Support Vector Regression for Censored Data (SVRc): A Novel Tool for Survival Analysis. In Proceedings of the 2008 Eighth IEEE International Conference on Data Mining, Pisa, Italy, 15–19 December 2008; Institute of Electrical and Electronics Engineers (IEEE): Piscataway, NJ, USA, 2008; pp. 863–868. [Google Scholar]
  23. Shivaswamy, P.K.; Chu, W.; Jansche, M. A support vector approach to censored targets. In Proceedings of the IEEE International Conference on Data Mining, ICDM 2007, Omaha, NE, USA, 28–31 October 2007; Institute of Electrical and Electronics Engineers (IEEE): Piscataway, NJ, USA, 2008; pp. 655–660. [Google Scholar]
  24. Kiaee, F.; Sheikhzadeh, H.; Eftekhari Mahabadi, S. Relevance Vector Machine for Survival Analysis. In IEEE Transactions on Neural Networks and Learning Systems; Institute of Electrical and Electronics Engineers (IEEE): Piscataway, NJ, USA, 2015; Volume 27, pp. 648–660. [Google Scholar]
  25. Tipping, M. Sparse Bayesian Learning and the Relevance Vector Mach. J. Mach. Learn. Res. 2001, 1, 211–244. [Google Scholar]
  26. Ouyang, Z. Bayesian Additive Regression Kernels; Duke University: Durham, NC, USA, 2008. [Google Scholar]
  27. Chakraborty, S.; Ghosh, M.; Mallick, B.K. Bayesian nonlinear regression for large p small n problems. J. Multivar. Anal. 2012, 108, 28–40. [Google Scholar] [CrossRef]
  28. Aronszajn, N. Theory of reproducing kernels. Trans. Am. Math. Soc. 1950, 68, 337–404. [Google Scholar] [CrossRef]
  29. Maity, A.; Mallick, B.K. Proportional Hazards Regression Using Bayesian Kernel Machines. In Bayesian Modeling in Bioinformatics; CRC Press Taylor & Francis Group: Boca Raton, FL, USA, 2011; pp. 317–336. [Google Scholar]
  30. Kalbfleisch, J.D. Non-Parametric Bayesian Analysis of Survival Time Data. J. R. Stat. Soc. Ser. B 1978, 40, 214–221. [Google Scholar] [CrossRef]
  31. Electronic Health Records. Available online: https://www.cms.gov/Medicare/E-Health/EHealthRecords/index.html (accessed on 20 November 2019).
  32. Hodgkins, A.J.; Bonney, A.; Mullan, J.; Mayne, D.; Barnett, S. Survival analysis using primary care electronic health record data: A systematic review of the literature. Health Inf. Manag. J. 2017, 47, 6–16. [Google Scholar] [CrossRef]
  33. Albert, J.H.; Chib, S. Bayesian Analysis of Binary and Polychotomous Response Data. J. Am. Stat. Assoc. 1993, 88, 669. [Google Scholar] [CrossRef]
  34. Gelman, A.; Rubin, D.B. Inference from Iterative Simulation Using Multiple Sequences. Stat. Sci. 1992, 7, 457–472. [Google Scholar] [CrossRef]
  35. Moriña, D.; Navarro, A. The R Package survsim for the Simulation of Simple and Complex Survival Data. J. Stat. Softw. 2014, 59, 1–20. [Google Scholar] [CrossRef]
  36. McGilchrist, C.A.; Aisbett, C.W. Regression with Frailty in Survival Analysis. Biometrics 1991, 47, 461. [Google Scholar] [CrossRef] [PubMed]
  37. Therneau, T.; Grambsch, P. Modeling Survival Data: Extending the Cox Model, 1st ed.; Springer: New York, NY, USA, 2000. [Google Scholar]
  38. Johnson, A.E.W.; Pollard, T.J.; Shen, L.; Lehman, L.-W.H.; Feng, M.; Ghassemi, M.; Moody, B.; Szolovits, P.; Celi, L.A.; Mark, R.G. MIMIC-III, a freely accessible critical care database. Sci. Data 2016, 3, 160035. [Google Scholar] [CrossRef] [PubMed]
  39. Fahrmeir, L. Encyclopedia of Biostatistics. Discrete Survival-Time Models; Wiley: Hoboken, NJ, USA, 1998; pp. 1163–1168. [Google Scholar]
  40. Masyn, K.E. Discrete-Time Survival Mixture Analysis for Single and Recurrent Events Using Latent Variables. Unpublished Doctoral Dissertation, University of California, Los Angeles, CA, USA, 2003. Available online: http://www.statmodel.com/download/masyndissertation.pdf (accessed on 20 April 2020).
Figure 1. Subject’s survival status within an observation time window. Subject #5 survives at the end of the observation and the data is missing outside the monitor window (right-censored).
Figure 1. Subject’s survival status within an observation time window. Subject #5 survives at the end of the observation and the data is missing outside the monitor window (right-censored).
Stats 05 00038 g001
Figure 2. Transform the original survival dataset into a binary classification dataset.
Figure 2. Transform the original survival dataset into a binary classification dataset.
Stats 05 00038 g002
Figure 3. Estimated survival probability curves for the simulated dataset by nonparametric survival BARK, survival BART, and Kaplan–Meier method.
Figure 3. Estimated survival probability curves for the simulated dataset by nonparametric survival BARK, survival BART, and Kaplan–Meier method.
Stats 05 00038 g003
Figure 4. Estimated survival probability curves for the kidney catheter dataset by survival BARK, survival BART, and Kaplan–Meier method.
Figure 4. Estimated survival probability curves for the kidney catheter dataset by survival BARK, survival BART, and Kaplan–Meier method.
Stats 05 00038 g004
Figure 5. Posterior probability plot for the 30-day hospital readmission data.
Figure 5. Posterior probability plot for the 30-day hospital readmission data.
Stats 05 00038 g005
Table 1. RMSE from 5-fold cross-validation performance of Survival BARK, Survival BART, and Cox-PH Model on the simulated dataset.
Table 1. RMSE from 5-fold cross-validation performance of Survival BARK, Survival BART, and Cox-PH Model on the simulated dataset.
Cross-Validation FoldSurvival BARKSurvival BARTCox PH-Model
10.76289940.77260982.9273658
20.74170080.91803552.9747698
30.92388480.97069462.9974077
40.74779131.04739692.9366146
50.86383850.87416332.9211262
Mean0.80802290.91658002.9514568
Table 2. RMSE from 5-fold cross-validation performance of Survival BARK, Survival BART, and Cox-PH Model on the kidney catheter dataset.
Table 2. RMSE from 5-fold cross-validation performance of Survival BARK, Survival BART, and Cox-PH Model on the kidney catheter dataset.
Cross-Validation FoldSurvival BARKSurvival BARTCox PH-Model
10.67382030.98655523.6232368
20.84002541.10559544.1958699
30.78213221.21448793.8610213
40.37097010.49962613.6559455
50.93088251.26463323.9148510
Mean0.71956611.01418003.8501849
Table 3. RMSE from 5-fold cross-validation performance of Survival BARK, Survival BART, and Cox- PH Model on the 30-day Hospital Readmission Data.
Table 3. RMSE from 5-fold cross-validation performance of Survival BARK, Survival BART, and Cox- PH Model on the 30-day Hospital Readmission Data.
Cross-Validation FoldSurvival BARKSurvival BARTCox PH-Model
11.78293733.85786048.1691553
21.68412023.88924098.0396034
31.63833233.84802788.1605448
41.85995753.63385038.2678329
51.72269403.78440238.3306097
Mean1.73760823.80267638.1935492
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Chakraborty, S.; Zhao, P.; Huang, Y.; Dey, T. Semiparametric Survival Analysis of 30-Day Hospital Readmissions with Bayesian Additive Regression Kernel Model. Stats 2022, 5, 617-630. https://doi.org/10.3390/stats5030038

AMA Style

Chakraborty S, Zhao P, Huang Y, Dey T. Semiparametric Survival Analysis of 30-Day Hospital Readmissions with Bayesian Additive Regression Kernel Model. Stats. 2022; 5(3):617-630. https://doi.org/10.3390/stats5030038

Chicago/Turabian Style

Chakraborty, Sounak, Peng Zhao, Yilun Huang, and Tanujit Dey. 2022. "Semiparametric Survival Analysis of 30-Day Hospital Readmissions with Bayesian Additive Regression Kernel Model" Stats 5, no. 3: 617-630. https://doi.org/10.3390/stats5030038

Article Metrics

Back to TopTop