Next Article in Journal
Imbedding Pd Nanoparticles into Porous In2O3 Structure for Enhanced Low-Concentration Methane Sensing
Next Article in Special Issue
Deep Learning Models for Stress Analysis in University Students: A Sudoku-Based Study
Previous Article in Journal
Efficient Data-Driven Machine Learning Models for Cardiovascular Diseases Risk Prediction
Previous Article in Special Issue
The Trade-Off between Airborne Pandemic Control and Energy Consumption Using Air Ventilation Solutions
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Two-Step Approach for Occupancy Estimation in Intensive Care Units Based on Bayesian Optimization Techniques

1
Galicia Sur Health Research Institute (IIS Galicia Sur), Álvaro Cunqueiro Hospital, 36310 Vigo, Spain
2
Department of Electronic Technology, University of Vigo, 36310 Vigo, Spain
3
Intensive Care Unit Department, Complexo Hospitalario Universitario de Vigo (SERGAS), Álvaro Cunqueiro Hospital, 36213 Vigo, Spain
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(3), 1162; https://doi.org/10.3390/s23031162
Submission received: 30 November 2022 / Revised: 14 January 2023 / Accepted: 17 January 2023 / Published: 19 January 2023
(This article belongs to the Special Issue Sensors Fusion in Digital Healthcare Applications)

Abstract

:
Due to the high occupational pressure suffered by intensive care units (ICUs), a correct estimation of the patients’ length of stay (LoS) in the ICU is of great interest to predict possible situations of collapse, to help healthcare personnel to select appropriate treatment options and to predict patients’ conditions. There has been a high amount of data collected by biomedical sensors during the continuous monitoring process of patients in the ICU, so the use of artificial intelligence techniques in automatic LoS estimation would improve patients’ care and facilitate the work of healthcare personnel. In this work, a novel methodology to estimate the LoS using data of the first 24 h in the ICU is presented. To achieve this, XGBoost, one of the most popular and efficient state-of-the-art algorithms, is used as an estimator model, and its performance is optimized both from computational and precision viewpoints using Bayesian techniques. For this optimization, a novel two-step approach is presented. The methodology was carefully designed to execute codes on a high-performance computing system based on graphics processing units, which considerably reduces the execution time. The algorithm scalability is analyzed. With the proposed methodology, the best set of XGBoost hyperparameters are identified, estimating LoS with a MAE of 2.529 days, improving the results reported in the current state of the art and probing the validity and utility of the proposed approach.

1. Introduction

The intensive care units (ICUs) of hospitals have a variety of devices to monitor the patients’ health states that generate a large amount of data, allowing healthcare personnel to be aware of the patient’s vital sign values and to make the most appropriate decisions to ensure their correct evolution [1]. It is very difficult to analyze all of this volume of information manually, so it is of great interest to use artificial intelligence (AI) tools that automate and help in these tasks [2], especially in extraordinary situations such as the one experienced due to the COVID-19 pandemic, where ICUs were overwhelmed.
The patient’s ICU length of stay (LoS) is an important metric from a clinical point of view. Due to the high occupational pressure suffered by the ICU, a correct estimation of the patients’ LoS in the ICU is of great interest to predict possible situations of collapse, to help healthcare personnel to select appropriate treatment options and to predict patients’ conditions. The design of a reliable estimator system is useful to anticipate future collapse situations and to take the corresponding actions [2], such as conditioning an area similar to the ICU urgently that accommodates future patients. In this field there are several studies. The vast majority of them approach the problem as a binary classification, predicting, for example, which patients will stay in the ICU for more or less than a defined number of days [3,4,5]. However, other studies approach the problem from the point of view of the exact calculation of LoS [6,7,8], which allows for a finer prediction of the ICU situation. In this article, it was decided to approach the problem from this second point of view, estimating the ICU stay duration from the monitoring data obtained during the first 24 h since the moment the patient accessed the ICU. In order to obtain more precise results, it was decided to delve into the optimization of the hyperparameters of the model, a task which requires a high computational load.
One of the goals of high-performance computing systems is to reduce the execution time of a given task [9]. Due to the increased computational load of artificial intelligence problems, it is necessary to use these systems to reduce the execution time. In recent years, performance enhancement from one processor generation to another has stagnated, making it necessary to find other ways to optimize the performance of algorithms. One of the main alternatives is the use of GPUs (graphics processing units), which help to continue to reduce execution times.
Occupancy prediction is a frequent topic within the healthcare ecosystem [10,11], as well as in other related fields, such as building energy systems [12], building performance [13] and heating, ventilation and air conditioning (HVAC) system control [14]. In this work, we propose a new methodology based on a novel two-step Bayesian optimization approach to improve LoS estimation methods used to predict ICU occupancy. This method allows one to optimize an LoS XGBoost predictor that uses clinical variables extracted from ICU monitoring, by finding the best set of hyperparameters of the model. The proposed approach improves the results of LoS predictions that would be obtained in the case of carrying out the optimization in the conventional way (regular Bayesian) and also improves the results obtained in other state-of-the-art works [3,4,5,6,7,8].
Due to the high computational load that this task entails, an important part of the methodology is the parallelization of the problem on a GPU architecture, allowing it to be solved in a computationally efficient way. XGBoost is used as estimator model. All these contributions are supported by experimental results.
The remainder of the article is structured as follows. First, the materials used are detailed. Then, the proposed methodology is explained and the experimental results are presented. Finally, the discussion and conclusions of the work are presented.

2. Materials

2.1. Data Source

In this work, the MIMIC-III (Medical Information Mart for Intensive Care III) [15] ICU database, developed by the MIT (Massachusetts Institute of Technology), was used. It contains data from 61,532 ICU stays at Beth Israel Deaconess Medical Center.
The database collected demographic data, vital sign measurements made at the bedside (1 data point per hour), laboratory test data, procedures, medications, caregiver notes, imaging reports, signals (electrocardiography (ECG), photoplethysmogram (PPG), arterial blood pressure (ABP)), etc. It contains the variable to be estimated in this work (LoS in the ICU) with a resolution of ± 10 4 days (±8.64 s). Figure 1 shows the different data sources of an ICU. The detailed structure of the database can be found in the MIMIC-III original publication [15].

2.2. XGBoost

XGBoost [16] is a gradient boosting technique based on ensemble learning. These techniques correct errors made by previous models in successive ones, optimizing a loss function. This function (1) is modified at each iteration t. The successive models are built using the exact greedy algorithm, which analyzes all possible split loss reduction options, until the stop condition is achieved. A more detailed explanation may be found in [16].
L ( t ) = i = 1 n l ( y i , y ^ i ( t 1 ) ) + Ω ( f t )
where l is a differentiable convex loss function that must be transformed into another one in a Euclidean domain using Taylor’s Theorem, the pair ( y i , x i ) represents the training set, y ^ i is the final prediction and Ω ( f t ) is the regularization term used to penalize more complex models through both Lasso and Ridge regularization and to prevent overfitting.
XGBoost is one of the most popular algorithms in the state of the art [17], highlighting its computational efficiency and GPU support. It is one of the most used algorithms in recent biomedical works based on tabular data [18,19,20].
As stated above [21,22], XGBoost uses three types of parameters: general, booster and learning task parameters. General parameters indicate the booster used: tree or linear model. Booster parameters are related with the booster employed and define the internal performance parameters, e.g., learning rate or number of estimators, while learning task parameters indicate the corresponding learning objective. The methodology presented in this work focuses on learning task parameters seeking to optimize the performance of the XGBoost regressor model. These parameters are:
  • Learning rate: Step-size shrinkage used in updates to prevent overfitting. After each boost step, the weights of the new features can be obtained directly, and the learning rate reduces the weights of the features to make the boost process more conservative.
  • Maximum delta step: the maximum delta step each leaf output is allowed to be.
  • Maximum depth of a tree: increasing this value will make the model more complex.
  • Maximum leaves: maximum number of nodes to be added.
  • Minimum child weight: Minimum sum of instance weight (hessian) needed in a child. If the tree partition step results in a leaf node with the sum of instance weight being less than the minimum child weight, then the building process will give up further partitioning.
  • Number of estimators: The number of trees in the ensemble. It is equivalent to the number of boosting rounds.
  • Region alpha: L1 regularization term on weights. Increasing this value will make the model more conservative.
  • Region lambda: L2 regularization term on weights. Increasing this value will make the model more conservative. It is normalized to number of training examples.
  • Scale pos weight: controls the balance of positive and negative weights, useful for unbalanced classes.
  • Subsample: Subsample ratio of the training instances. Setting it to 0.5 means that XGBoost would randomly sample half of the training data prior to growing trees. Subsampling will occur once in every boosting iteration.

2.3. Bayesian Optimization

One of the relevant parts of a machine learning pipeline is the optimization of the estimator model. This task implies a high computational load, so optimizing it efficiently is a key factor. Given a dataset D, the goal of hyperparameter optimization [23] is to find λ in (2):
λ * = arg min λ ϵ Λ I E D t r a i n , D t e s t ϵ D V ( L , A λ , D t r a i n , D t e s t )
where λ is a vector of hyperparameters from the hyperparameter search space Λ , A is the predictive model and V ( L , A λ , D t r a i n , D t e s t ) measures the loss of the model A .
In this work, Bayesian techniques [24] were used to carry out this task, which stand out for their computational efficiency when performing the search. The surrogate model (a Gaussian stochastic probabilistic model) was used. The optimization was divided into several stages: The first stage was the set-up of the surrogate model. Next, the best hyperparameter combination was sought, and it was applied to the real objective function. Finally, the surrogate model was updated. This process was repeated iteratively until the defined criteria were achieved.
One of the most used objective functions is that of expected improvements (3), because it can be calculated in closed form if the estimator model y with the configuration λ follows a normal distribution (4) [23].
I E [ I I ( λ ) ] = I E [ m a x ( f m i n y , 0 ) ]
I E [ I I ( λ ) ] = ( f m i n μ ( λ ) ) Φ ( f m i n μ ( λ ) σ ) + σ ϕ ( f m i n μ ( λ ) σ )
where ϕ ( · ) and Φ ( · ) are the standard normal density and standard normal distribution, and f m i n is the best observed value.
In this work, the open-source package Hyperopt [25], which uses Bayesian optimization as the search technique, was used.

3. Methodology

The proposed methodology to optimize the estimator model of the patient’s LoS is divided into several stages. Figure 2a shows an outline of these stages, while a more detailed explanation of the hyperparameter search using Bayesian optimization is shown in Figure 2b. The first stage is related to the data preprocessing and the feature extraction. The second one is devoted to building the LoS estimator. In this work, we propose to use the state-of-the-art approach based on gradient boosting, namely XGBoost, due to the good results obtained in estimation tasks and the high level of computational improvement that could be achieved on a GPU architecture during the optimization stage. The third stage is devoted to optimizing the model hyperparameters in order to improve the performance of the estimator. In this work, a novel two-step Bayesian optimization approach is proposed, implemented on a GPU architecture to reduce the execution time. Finally, the estimator model is validated.

3.1. Data Preprocessing and Feature Extraction

The first stage of the pipeline was devoted to data preprocessing to obtain a set of features, derived from the clinical data, used to fit and validate the estimator model. Such a stage required two tasks.
The first task consisted of selecting the variables used to produce the dataset from the original database. In this approach, all available clinical variables were considered, selecting only the ones that were present in at least 80% of the patients for building the model, as in other published works [21,26]. The reason why all of the variables were not available for all patients was because, depending on the pathology and the patient’s clinical condition, only a certain set of variables was monitored. Patients who did not have values in at least 2/3 of these variables were subsequently discarded. In order to deal with missing data, a schema for input data was required to fill values on such empty variables. Although there are other imputation methods, such as MissForest [27] or generative adversial networks [28], in this work we have proposed using K–Nearest Neighbors, as it is one of the most widely used in the current state of the art, in addition to its simplicity in implementation and usage.
The next task consisted of the conversion from variables to features, in this case by computing the mean value, standard deviation and maximum and minimum values of the clinical variables gathered during the first 24 h of the patient’s ICU stay, except for the volume of urine, for which only the total volume in this time interval was used.

3.2. LoS Estimation Model Building

Once the previous preprocessing stage was completed, it was necessary to proceed with the configuration stage of the ICU LoS estimator model. As mentioned above, the XGBoost model was used, which was fitted using the above described features. In order to train the model, a random split train and test of these data was carried out following a ratio of 80/20. To validate the model, the mean absolute error (MAE) (5) was used, which is one of the most popular metrics in the current state of the art, allowing for the comparison of the results.
M A E = i = 1 n | y ^ i y i |
where n is the number of patients, y ^ i is the LoS estimated by XGBoost model and y i is the golden standard for the LoS, obtained from the database.

3.3. XGBoost Optimization: Two-Step Approach

To improve the accuracy of the model, we delved into the automatic optimization of the hyperparameters of XGBoost instead of using the default values.
This stage was divided into two steps. A first search step was executed using the initially defined hyperparameter search space. Then, a second search was performed once the search space was modified after the hyperparameters’ evolution during the first search stage was analyzed.
The optimization started with the hyperparameter search space ( Λ ) definition, which consisted of indicating which parameters should be varied during the different iterations (2) and within which limits, in addition to indicating the type of search space. There are three types of search space distributions: uniform, log uniform and q uniform, which return real values uniformly distributed between defined limits. Log uniform is more suitable for geometric series, whereas uniform and q uniform are more suitable for arithmetic series, with the difference being that q uniform returns round values, so the selection of the search space depends on the hyperparameter type. This part was fundamental, since an incorrect definition of space can cause the process to not work efficiently in computational terms.
In addition to defining the search space, it was necessary to define an objective function to minimize in each iteration, that is, the statistical metric that defined the quality of the estimator model needed to be defined. In this work, it was decided to use the MAE (5) as a metric. Firstly, the maximum number of evaluations (m) was defined as 5000.
Once this first search was completed, the search space was modified to improve the results obtained avoiding the bottlenecks derived from the initial defined limits and refocusing on the area where the best value was found after the first search, but expanding the search limits, as shown in the Results section. Table 1 summarizes the search space distribution used in the first and second phases, and the types of distributions used.

4. Results

In this section, the results obtained using the proposed methodology are presented and analyzed, in terms of the set of hyperparameters identified during the optimization stage, the performance of the model estimator using MAE as the metric and the computational performance. For this analysis, the initial dataset is randomly split into train (27,077 patients) and test (6770 patients) sets, calculating the MAE for the test set.

4.1. Hyperparameter Tuning Stage

Regarding the hyperparameter optimization stage, the method permits one to reduce the MAE, and consequently improve the estimations of the model, by identifying the set of parameters that provides the lower MAE. Table 2 shows the best combination of hyperparameters obtained both after the first search and after the second search, explained and described in Section 3.
The whole evolution of interactions of both optimization steps are plotted in Figure 3, which shows the minimum value of the MAE obtained (grouped into intervals of 50 iterations), that clearly shows the downward trend. The figure also shows how after redefining the search space from the results of the first search step (black vertical line), there is a sharp drop in the value of the MAE, obtaining a minimum value of 2.529 in iteration 6962.
Another important element when analyzing the optimization stage of the model is the evolution in the search, that is, how the values of the hyperparameters vary throughout the different trials. Figure 4 shows the value of one of the model hyperparameters (the maximum depth) in each trial, with its corresponding MAE. The figures corresponding to the rest of the hyperparameters are shown in the Supplementary Materials. From these figures, what was indicated above can also be seen, meaning that there is no direct correlation between the value of the individual hyperparameter and the value of the MAE obtained. However, it shows that the intervals between the optimal MAE value obtained and the trend in the search are found. From these results, a second optimization step was performed by redefining the search space of the Bayesian model, modifying the search space limits. After this second optimization step, a MAE of 2.529 was obtained.

4.2. Estimator Validation

Only the first stay of each patient was selected in the database to avoid possible information leakage and to compare the results obtained with related works [8], resulting in a total of 46,476 stays. As indicated in Section 3, a second filtering was subsequently carried out to discard patients for whom more that 1/3 of the variables were unavailable, obtaining a total of 33,847 ICU stays. Table 3 shows the main characteristics of the resulting cohort.
Table 4 shows the clinical ICU variables and features extracted from each patient. It consists of a total of 35 variables as 139 features.
To quantify the precision of the model when estimating the patient’s LoS in the ICU, the MAE was used as the metric. As mentioned above, the initial dataset was split into train (27,077 patients) and test (6770 patients) datasets, calculating the MAE for the test set. The LoS was estimated with a resolution of ± 10 4 days, obtaining a MAE of 2.529 days, lower than the rest of the current state-of-the-art works consulted. Rouzbahman et al. [29] obtain a MAE of 5.07 days, while Alghatani et al. [30] obtain a MAE of 2.64 days, both using MIMIC as the database. Moreover, the final MAE obtained is lower than the MAE obtained with the default hyperparameters (3.040) and after the first optimization step (2.539). Although it is true that there are studies with a slightly lower MAE value, they focus on a specific group of patients, instead of addressing the problem from a generic point of view, as we do. Our work improves the results of those who face problems from a generic point of view. Table 5 shows the results obtained.

4.3. Computational Performance

Model fitting and optimization are extremely expensive processes. To carry out this work, a high-performance computing system was required. This architecture consisted of a 24-core CPU, 256 GB of RAM and one NVIDIA A100 GPU unit. The GPU had 40 GB of memory and consisted of 108 multiprocessors, allowing the execution of 2024 threads per multiprocessor, a total of 221,184 threads per GPU [31]. One of the main objectives of the methodology proposed in this article was computational efficiency. For this, it was necessary that the software used could be executed on the GPU.
In order to benchmark the effects that hardware and dataset dimensions have on the performance of the proposed method, several tests were conducted, both on the GPU and on the CPU (using one core and three cores). Each hardware configuration was tested with three different dataset versions, namely the original dataset used in a previous experiment (dataset), this dataset reduced to half of the number of patients (Subdataset 1) and a dataset with half of the features (Subdataset 2). Both the selection of patients and characteristics were performed randomly. Table 6 shows the results obtained in minutes using 100 iterations (m).
Once the drastic difference in time between the CPU and GPU was verified, the number of iterations was increased to 500, only using the GPU to analyze how it influenced the execution time. It was observed that the execution time increases practically in the same proportion as the number of iterations. The variation in this proportion is due to the fact that depending on the value of the hyperparameters in which it is iterating, the execution time differs.

5. Discussion

The proposed methodology allows one to identify the set of hyperparameters that provides the best performance of the predictor in terms of minimizing the mean absolute error (MAE). With the best combination of hyperparameters, the LoS was estimated with a MAE of 2.529 days, lower than the rest of the current state-of-the-art works consulted. It is also lower than the MAE obtained after the first search stage (MAE = 2.539 days) and lower than using the default XGBoost hyperparameters set without executing the optimization task (MAE = 3.04 days). Although it is true that there are studies with a slightly lower MAE value, they focus on a specific group of patients, instead of addressing the problem from a generic point of view, as was indicated in Section 1.
XGBoost was used as an estimator model due to being one of the ones that obtains the best results in other current state-of-the-art works and its computational efficiency. However, the same methodology could be applied to other models (Random Forest, Support Vector Machine, etc.). The feature extraction stage could be modified adding data from other ICU sensors, extracting other statistics or considering another time window instead of 24 h.
One of the fundamental reasons for obtaining better results is the optimization phase of the model. Adapting the search methodology to work in the GPU allowed us to intensify the search for the best combination of hyperparameters. In addition, as already indicated above, the search technique used based on Bayesian optimization was characterized by performing the search for hyperparameters more efficiently from a computational point of view. The results also confirm that the two-step approach proposed for the optimization of the estimator model improves the results obtained. This same proposal could be applied to another problem.
A stagnation of the MAE improvement is observed with the passing of the trials, which gives rise to an analysis of the relationship of computational cost–improvement obtained, in which it would be necessary to assess the available hardware and the allowable error in each particular situation. Regarding the relationship between the individual values of the hyperparameters and the value of the MAE, no clear patterns are observed, which justifies that what is really relevant is the combination of the values of the different hyperparameters, not the values of each one separately. This demonstrates the need to perform the optimization of the model automatically, as was the case in this work, instead of performing it manually, which would be unfeasible in terms of work times, and worse results would be obtained.
From the computational results, how the use of the GPU drastically reduces the execution time is observed. It is also interesting to analyze how the variations in the size of the dataset hardly imply variations in execution times in the GPU. This makes sense, since although the size of the dataset is reduced, GPUs are characterized by processing large amounts of data in parallel in a single clock cycle, so the variation in the size of the dataset does not always imply a variation in the timing of execution. If the same methodology was used in the classic architectures based solely on the CPU, the execution time would have been drastically higher, which would limit the ability to analyze the possible improvements to be made to the methodology for a real clinical application.
As for future lines of work, this methodology can be extended in several ways. The first line of research is related to the application domain, regarding using this novel methodology to estimate other time variables that are different to LoS, as well as to estimate other variables aside from time within the ICU, and extending the applicability to other hospital areas and by extension to any other social prediction. Another future line of research could be exploring the configuration domain, namely the modifications in the internal configuration of the methodology (the estimator model used to change the XGBoost predictor model by any other regressor model that could provide a time estimation). Modifications in the optimization stage and in the feature extraction stage are also of great interest, using more data sources or using different feature extraction techniques.

6. Conclusions

This article presents a methodology to estimate the LoS in the ICU using data collected by UCI sensors and other sources (laboratory and medical history) during the first 24 h of a patient’s admission, focusing on the optimization stage of the estimator model, both from a computational point of view and at the estimation precision level. To do this, the methodology was adapted so that the training of the model was executed on the GPU and the search for the best combination of hyperparameters was carried out automatically using Bayesian optimization techniques in a novel two-step approach.
The results using 33,847 patients demonstrate the validity of the proposed methodology, obtaining a MAE of 2.529 days, lower than that of other works from the current state of the art consulted. In addition, an improvement in model precision by dividing the model optimization phase into two steps instead of performing it in a single step is also demonstrated.
This work opens several future lines of research for applying the present methodology to predict other variables within the ICU, or in other hospital areas. Another future line of research could be exploring modifications in the internal configuration of the methodology (the estimator model used, modifications in the optimization stage, using different features, etc.).

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/s23031162/s1. Figures similar to Figure 4 are added as Supplementary Material for each hyperparameter.

Author Contributions

J.A.G.-N.: conceptualization, data curation, formal analysis, investigation, methodology, software, validation, visualization, writing—original draft, writing—review editing; L.B.: validation, investigation; S.C.: investigation; J.F.: investigation, supervision; J.J.R.-A.: investigation, supervision; D.V.: investigation, validation; C.V.: conceptualization, investigation, formal analysis, methodology, resources, writing—review editing, supervision, funding acquisition. All authors have reviewed the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

The research was partially supported by Agencia Gallega de Innovación (GAIN) through “Proxectos de investigación sobre o SARS-CoV-2 e a enfermidade COVID-19 con cargo ao Fondo COVID-19” Program, with code number IN845D-2020/29, and through “Axudas para a consolidación e estruturación de unidades de investigación competitivas e outras accións de fomento nos organismos públicos de investigación de Galicia e noutras entidades do sistema galego de I+D+i - GPC” with code number IN607B-2021/18.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets analyzed in the current study are available in the PhysioNet repository, https://physionet.org/content/mimiciii/1.4/ (accessed on 10 January 2021).

Conflicts of Interest

The authors declare no conflict of interest. The founders had no role in the design of the study; in the collection, analyses or interpretation of data; in the writing of the manuscript or in the decision to publish the results.

Abbreviations

The following abbreviations are used in this manuscript:
CPUcentral processing unit
GPUgraphics processing units
ICUintensive care unit
INRinternational normalized ratio
LoSlength of stay
MAEmean absolute error
MIMICMedical Information Mart for Intensive Care

References

  1. Gutierrez, G. Artificial Intelligence in the Intensive Care Unit. Crit. Care 2020, 24, 1–9. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. van de Sande, D.; van Genderen, M.E.; Huiskens, J.; Gommers, D.; van Bommel, J. Moving from bytes to bedside: A systematic review on the use of artificial intelligence in the intensive care unit. Intensive Care Med. 2021, 47, 750–760. [Google Scholar] [CrossRef] [PubMed]
  3. Peres, I.T.; Hamacher, S.; Oliveira, F.L.C.; Thomé, A.M.T.; Bozza, F.A. What factors predict length of stay in the intensive care unit? Systematic review and meta-analysis. J. Crit. Care 2020, 60, 183–194. [Google Scholar] [CrossRef] [PubMed]
  4. Awad, A.; Bader-El-Den, M.; McNicholas, J. Patient length of stay and mortality prediction: A survey. Health Serv. Manag. Res. 2017, 30, 105–120. [Google Scholar] [CrossRef] [Green Version]
  5. Su, L.; Xu, Z.; Chang, F.; Ma, Y.; Liu, S.; Jiang, H.; Wang, H.; Li, D.; Chen, H.; Zhou, X.; et al. Early Prediction of Mortality, Severity, and Length of Stay in the Intensive Care Unit of Sepsis Patients Based on Sepsis 3.0 by Machine Learning Models. Front. Med. 2021, 8, 883. [Google Scholar] [CrossRef]
  6. Alsinglawi, B.; Alnajjar, F.; Mubin, O.; Novoa, M.; Alorjani, M.; Karajeh, O.; Darwish, O. Predicting Length of Stay for Cardiovascular Hospitalizations in the Intensive Care Unit: Machine Learning Approach. In Proceedings of the 2020 42nd Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), Montreal, QC, Canada,, 20–24 July 2020; pp. 5442–5445. [Google Scholar] [CrossRef]
  7. Houthooft, R.; Ruyssinck, J.; van der Herten, J.; Stijven, S.; Couckuyt, I.; Gadeyne, B.; Ongenae, F.; Colpaert, K.; Decruyenaere, J.; Dhaene, T.; et al. Predictive modelling of survival and length of stay in critically ill patients using sequential organ failure scores. Artif. Intell. Med. 2015, 63, 191–207. [Google Scholar] [CrossRef] [Green Version]
  8. Purushotham, S.; Meng, C.; Che, Z.; Liu, Y. Benchmarking deep learning models on large healthcare datasets. J. Biomed. Inf. 2018, 83, 112–134. [Google Scholar] [CrossRef]
  9. Kindratenko, V.; Trancoso, P. Trends in High-Performance Computing. Comput. Sci. Eng. 2011, 13, 92–95. [Google Scholar] [CrossRef]
  10. Whitt, W.; Zhang, X. Forecasting arrivals and occupancy levels in an emergency department. Oper. Res. Health Care 2019, 21, 1–18. [Google Scholar] [CrossRef]
  11. Littig, S.J.; Isken, M.W. Short term hospital occupancy prediction. Health Care Manag. Sci. 2007, 10, 47–66. [Google Scholar] [CrossRef]
  12. Tekler, Z.D.; Chong, A. Occupancy prediction using deep learning approaches across multiple space types: A minimum sensing strategy. Build. Environ. 2022, 226, 109689. [Google Scholar] [CrossRef]
  13. Tekler, Z.D.; Low, R.; Gunay, B.; Andersen, R.K.; Blessing, L. A scalable Bluetooth Low Energy approach to identify occupancy patterns and profiles in office spaces. Build. Environ. 2020, 171, 106681. [Google Scholar] [CrossRef]
  14. Dong, J.; Winstead, C.; Nutaro, J.; Kuruganti, T. Occupancy-Based HVAC Control with Short-Term Occupancy Prediction Algorithms for Energy-Efficient Buildings. Energies 2018, 11, 2427. [Google Scholar] [CrossRef] [Green Version]
  15. Johnson, A.E.W.; Pollard, T.J.; Shen, L.; Lehman, L.w.H.; Feng, M.; Ghassemi, M.; Moody, B.; Szolovits, P.; Anthony Celi, L.; Mark, R.G. MIMIC-III, a freely accessible critical care database. Sci. Data 2016, 3, 1–9. [Google Scholar] [CrossRef] [Green Version]
  16. Chen, T.; Guestrin, C. XGBoost: A Scalable Tree Boosting System. In KDD ’16: Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining; Association for Computing Machinery: New York, NY, USA, 2016; pp. 785–794. [Google Scholar] [CrossRef] [Green Version]
  17. Nielsen, D. Tree Boosting With XGBoost—Why Does XGBoost Win “Every” Machine Learning Competition? Ph.D. Thesis, NTNU, Trondheim, Norway, 2016. [Google Scholar]
  18. González, B.S.; Conceição, R.; Pimenta, M.; Tomé, B.; Guillén, A. Tackling the muon identification in water Cherenkov detectors problem for the future Southern Wide-field Gamma-ray Observatory by means of machine learning. Neural Comput. Applic. 2022, 34, 5715–5728. [Google Scholar] [CrossRef]
  19. Site, A.; Vasudevan, S.; Afolaranmi, S.O.; Lastra, J.L.M.; Nurmi, J.; Lohan, E.S. A Machine-Learning-Based Analysis of the Relationships between Loneliness Metrics and Mobility Patterns for Elderly. Sensors 2022, 22, 4946. [Google Scholar] [CrossRef]
  20. Yang, J.; Clifton, D.; Hirst, J.E.; Kavvoura, F.K.; Farah, G.; Mackillop, L.; Lu, H. Machine Learning-Based Risk Stratification for Gestational Diabetes Management. Sensors 2022, 22, 4805. [Google Scholar] [CrossRef]
  21. González-Nóvoa, J.A.; Busto, L.; Rodríguez-Andina, J.J.; Fariña, J.; Segura, M.; Gómez, V.; Vila, D.; Veiga, C. Using Explainable Machine Learning to Improve Intensive Care Unit Alarm Systems. Sensors 2021, 21, 7125. [Google Scholar] [CrossRef]
  22. González-Nóvoa, J.A.; Busto, L.; Santana, P.; Fariña, J.; Rodríguez-Andina, J.J.; Juan-Salvadores, P.; Jiménez, V.; Íñiguez, A.; Veiga, C. Using Bayesian Optimization and Wavelet Decomposition in GPU for Arterial Blood Pressure Estimation. In Proceedings of the 2022 44th Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), Glasgow, UK, 11–15 July 2022; pp. 1012–1015. [Google Scholar] [CrossRef]
  23. Feurer, M.; Hutter, F. Hyperparameter Optimization. In Automated Machine Learning; Springer: Cham, Switzerland, 2019; pp. 3–33. [Google Scholar] [CrossRef] [Green Version]
  24. Shahriari, B.; Swersky, K.; Wang, Z.; Adams, R.P.; de Freitas, N. Taking the Human Out of the Loop: A Review of Bayesian Optimization. Proc. IEEE 2015, 104, 148–175. [Google Scholar] [CrossRef] [Green Version]
  25. Bergstra, J.; Yamins, D.; Cox, D.D. Making a Science of Model Search. arXiv 2012, arXiv:1209.5111. [Google Scholar]
  26. Jiang, Z.; Bo, L.; Xu, Z.; Song, Y.; Wang, J.; Wen, P.; Wan, X.; Yang, T.; Deng, X.; Bian, J. An explainable machine learning algorithm for risk factor analysis of in-hospital mortality in sepsis survivors with ICU readmission. Comput. Methods Programs Biomed. 2021, 204, 106040. [Google Scholar] [CrossRef] [PubMed]
  27. Stekhoven, D.J.; Bühlmann, P. MissForest—non-parametric missing value imputation for mixed-type data. Bioinformatics 2012, 28, 112–118. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  28. Low, R.; Tekler, Z.D.; Cheah, L. Predicting Commercial Vehicle Parking Duration using Generative Adversarial Multiple Imputation Networks. Transp. Res. Rec. 2020, 2674, 820–831. [Google Scholar] [CrossRef]
  29. Rouzbahman, M.; Jovicic, A.; Chignell, M. Can Cluster-Boosted Regression Improve Prediction of Death and Length of Stay in the ICU? IEEE J. Biomed. Health Inf. 2016, 21, 851–858. [Google Scholar] [CrossRef] [PubMed]
  30. Alghatani, K.; Ammar, N.; Rezgui, A.; Shaban-Nejad, A. Predicting Intensive Care Unit Length of Stay and Mortality Using Patient Vital Signs: Machine Learning Model Development and Validation. JMIR Med. Informat. 2021, 9, e21347. [Google Scholar] [CrossRef]
  31. NVIDIA. NVIDIA A100 GPUs Power the Modern Data Center. 2022. Available online: https://www.nvidia.com/en-us/data-center/a100/ (accessed on 23 May 2022).
Figure 1. Intensive care unit structure.
Figure 1. Intensive care unit structure.
Sensors 23 01162 g001
Figure 2. Outline of the proposed methodology. (a) Shows a general outline, while a more detailed explanation of the hyperparameter search using Bayesian optimization is shown in (b).
Figure 2. Outline of the proposed methodology. (a) Shows a general outline, while a more detailed explanation of the hyperparameter search using Bayesian optimization is shown in (b).
Sensors 23 01162 g002
Figure 3. Evolution of minimum MAE across 50 iteration intervals.
Figure 3. Evolution of minimum MAE across 50 iteration intervals.
Sensors 23 01162 g003
Figure 4. Evolution of maximum depth value and its corresponding MAE across the trials.
Figure 4. Evolution of maximum depth value and its corresponding MAE across the trials.
Sensors 23 01162 g004
Table 1. Hyperparameter distributions.
Table 1. Hyperparameter distributions.
HyperparameterDistribution TypeFirst Step DistributionSecond Step Distribution
MinMaxMinMax
Learning rateloguniform−80−7.5−4.5
Max. depthquniform11514.525
Min. child weightquniform010612
Max. delta stepquniform01000.5
Subsample ratiouniform0.110.40.7
Lambda regionuniform0.110.71.2
Alpha regionuniform0.110.40.8
Scale pos weightuniform0.110.450.85
Max. number of leavesquniform0100.11
Number of estimatorsquniform110,0001002500
Table 2. Hyperparameter optimal values.
Table 2. Hyperparameter optimal values.
HyperparameterBest Value
First SearchSecond Search
Learning rate0.001350.00132
Max. depth1524
Min. child weight912
Max. delta step00
Subsample ratio0.5250.529
Lambda region0.8911.184
Alpha region0.6610.559
Scale pos weight0.6740.755
Max. number of leaves00
Number of estimators745812
Table 3. Characteristics of the data source.
Table 3. Characteristics of the data source.
MIMIC-IIIResulting Cohort
Number of patients46,47633,847
Number of ICU stays61,53233,847
Average age64.9374.65
GenderF: 20,380
M: 26,096
F: 19,319
M: 14,528
LoS average4.924.32
LoS standard deviation9.646.21
LoS variance92.9138.55
Table 4. Features extracted from each clinical variable.
Table 4. Features extracted from each clinical variable.
Clinical VariablesStatistics
AgeValue in admission
GenderF/M
Urine outputAccumulated value after 24 h
Glasgow coma motor scale
Glasgow coma verbal scale
Glasgow coma eyes scale
Systolic blood pressure
Heart rate
Body temperature
PaO2
FiO2
Serum urea nitrogen level
Sodium level
Potassium level
Bilirubin level
Respiratory rate
Glucose
Albumin
Anion gapAverage, maximum, minimum
and standard deviation value
after 24 h
Chloride
Creatinine
Lactate
Calcium
Haematocrit
Hemoglobin
INR 1
Platelets
Prothrombin time test
Activated thromboplastin time
Base excess
PaCO2
FiCO2
PH
Total CO2
1 International normalized ratio of prothrombin time.
Table 5. Comparison of the MAE obtained using the default model with respect to that obtained after the first and second steps of the optimization phase.
Table 5. Comparison of the MAE obtained using the default model with respect to that obtained after the first and second steps of the optimization phase.
Default Estimator
(Without Optimization)
Optimized Model
(After First Step)
Optimized Model
(After Second Step)
3.0402.5392.529
Table 6. Time comparison. m is the number of iterations.
Table 6. Time comparison. m is the number of iterations.
CPUGPU
1 Core
(m = 100)
3 Core
(m = 100)
m = 100m = 500
Dataset85.86’39.26’9.70’75.23’
Subdataset 182.56’41.86’11.82’63.06’
Subdataset 244.63’24.50’9.07’44.63’
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

González-Nóvoa, J.A.; Busto, L.; Campanioni, S.; Fariña, J.; Rodríguez-Andina, J.J.; Vila, D.; Veiga, C. Two-Step Approach for Occupancy Estimation in Intensive Care Units Based on Bayesian Optimization Techniques. Sensors 2023, 23, 1162. https://doi.org/10.3390/s23031162

AMA Style

González-Nóvoa JA, Busto L, Campanioni S, Fariña J, Rodríguez-Andina JJ, Vila D, Veiga C. Two-Step Approach for Occupancy Estimation in Intensive Care Units Based on Bayesian Optimization Techniques. Sensors. 2023; 23(3):1162. https://doi.org/10.3390/s23031162

Chicago/Turabian Style

González-Nóvoa, José A., Laura Busto, Silvia Campanioni, José Fariña, Juan J. Rodríguez-Andina, Dolores Vila, and César Veiga. 2023. "Two-Step Approach for Occupancy Estimation in Intensive Care Units Based on Bayesian Optimization Techniques" Sensors 23, no. 3: 1162. https://doi.org/10.3390/s23031162

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop