ABSTRACT
OBJECTIVE:
To develop and validate a predictive model utilizing machine-learning techniques for estimating the length of hospital stay among patients who underwent coronary artery bypass grafting.
METHODS:
Three machine learning models (random forest, extreme gradient boosting and neural networks) and three traditional regression models (Poisson regression, linear regression, negative binomial regression) were trained in a dataset of 9,584 patients who underwent coronary artery bypass grafting between January 2017 and December 2021. The data were collected from hospital discharges from 133 centers in Brazil. Algorithms were ranked by calculating the root mean squared logarithmic error (RMSLE). The top performing algorithm was validated in a never-before-seen database of 2,627 patients. We also developed a model with the top ten variables to improve usability.
RESULTS:
The random forest technique produced the model with the lowest error. The RMLSE was 0.412 (95%CI 0.405–0.419) on the training dataset and 0.454 (95%CI 0.441–0.468) on the validation dataset. Non-elective surgery, admission to a public hospital, heart failure, and age had the greatest impact on length of hospital stay.
CONCLUSIONS:
The predictive model can be used to generate length of hospital stay indices that could be used as markers of efficiency and identify patients with the potential for prolonged hospitalization, helping the institution in managing beds, scheduling surgeries, and allocating resources.
DESCRIPTORS:
Length of Stay; Machine Learning; Coronary Artery Bypass
INTRODUCTION
Coronary artery bypass grafting (CABG) is the most commonly performed cardiac surgery in Brazil, representing 54.1% of cases11 Gomes WJ, Moreira RS, Zilli AC, Bettiati Jr LC, Figueira FAMS, D’ Azevedo SSP, et al. The brazilian registry of adult patient undergoing cardiovascular surgery, the BYPASS project: results of the first 1,722 patients. Braz J Cardiovasc Surg. 2017;32(2):71-6. https://doi.org/10.21470/1678-9741-2017-0053
https://doi.org/10.21470/1678-9741-2017-... . The length of hospital stay (LOS) related to CABG serves as a significant indicator of the quality of institutions providing this service, as it directly impacts resource utilization and healthcare costs. Timely discharge of patients following CABG can facilitate hospital workflow by freeing up beds and optimizing healthcare professionals’ time22 Organisation for Economic Cooperation and Development. Hospital discharges and average length of stay. In: Health at a Glance 2021: OECD Indicators. Paris: OECD Publishing; 2021. https://doi.org/10.1787/18faaea9-en
https://doi.org/10.1787/18faaea9-en... . Furthermore, an unnecessarily prolonged hospital stay can escalate the likelihood of adverse events such as falls, hospital infections and medication errors. The development of a predictive model capable of estimating LOS can be an effective tool for managing hospital expenses, enhancing service efficiency, and ensuring better patient care33 Stone K, Zwiggelaar R, Jones P, Mac Parthaláin NM. A systematic review of the prediction of hospital length of stay: Towards a unified framework. PLOS Digit Health. 2022;1(4):e0000017. https://doi.org/10.1371/journal.pdig.0000017
https://doi.org/10.1371/journal.pdig.000... .
Traditionally, predictive models in medicine are made by algorithms based on traditional statistics44 Deo RC. Machine learning in medicine: Will this time be different? Circulation. 2020;142(16):1521-3. https://doi.org/10.1161/circulationaha.120.050583
https://doi.org/10.1161/circulationaha.1... . However, in recent times, machine learning algorithms have emerged as a viable approach for developing predictive models55 Obermeyer Z, Emanuel EJ. Predicting the future – big data, machine learning, and clinical medicine. N Engl J Med. 2016;375(13):1216-9. https://doi.org/10.1056/NEJMp1606181
https://doi.org/10.1056/NEJMp1606181...
6 Rajkomar A, Dean J, Kohane I. Machine learning in medicine. N Engl J Med. 2019;380(14):1347-58. https://doi.org/10.1056/NEJMra1814259
https://doi.org/10.1056/NEJMra1814259... -77 Stevens LM, Mortazavi BJ, Deo RC, Curtis L, Kao DP. Recommendations for reporting machine learning analyses in clinical research. Circ Cardiovasc Qual Outcomes. 2020;13(10):e006556. https://doi.org/10.1161/CIRCOUTCOMES.120.006556
https://doi.org/10.1161/CIRCOUTCOMES.120... . Determining the most effective model for generating accurate predictions involves comparing different algorithms’ predictive abilities.
A potential limitation of predictive models is that they may not be applicable to a patient population that significantly differs from the population in which the model was developed88 Austin PC, van Klaveren D, Vergouwe Y, Nieboer D, Lee DS, Steyerberg EW. Geographic and temporal validity of prediction models: different approaches were useful to examine model performance. J Clin Epidemiol. 2016;79:76-85. https://doi.org/10.1016/j.jclinepi.2016.05.007
https://doi.org/10.1016/j.jclinepi.2016.... . To the best of our knowledge, no predictive models specifically developed to estimate LOS for patients undergoing CABG have been trained and validated using a Brazilian database.
The primary objective of this study was to develop and validate a predictive model utilizing machine learning techniques for estimating LOS among patients who underwent CABG, utilizing data from a Brazilian administrative database.
METHODS
Source of Data and Study Participants
For training and validation, we used an administrative database collected from hospital discharges from 133 hospitals, in all regions of Brazil, from January 2017 to December 2022. The database is derived from public and private insurance hospitals that employ DRG for the management of healthcare processes in healthcare systems. The dataset contains information for each hospitalization, encompassing demographic details, adverse events, primary and secondary diagnosis codes, as well as procedure codes. Data were collected by healthcare professionals trained in data collection. The entire database of patients who underwent CABG, comprising 12,211 admissions of individuals aged 18 years and older, was utilized to enhance the power of the predictive model and its generalization. Data from 9,584 patients who underwent CABG with or without valve replacement between January 2017 and December 2021 were used for training and internal validation of the predictive model. Then, the predictive model with the best performance was validated in a new sample of 2,627 patients who had undergone CABG between January 2022 and December 2022. Exclusion criteria were patients younger than 18 years, patients transferred to another hospital and patients who died in the same hospitalization. The study followed the Transparent Reporting of the Multivariable Prediction Model for Individual Prognosis or Diagnosis (TRIPOD) guidelines99 Moons KGM, Altman DG, Reitsma JB, Loannidis JPA, Macaskill P, Steyerberg EW, et al. Transparent Reporting of a multivariable prediction model for Individual Prognosis Or Diagnosis (TRIPOD): Explanation and Elaboration. Translation in to Russian. Dig Diagn. 2022;3(3):232-322. https://doi.org/10.17816/DD110794
https://doi.org/10.17816/DD110794... .
Outcome and Variable Selection
The variables were collected at hospital discharge using the International Classification of Diseases, 10th Revision (ICD-10), coding. Coders were instructed to identify conditions present on admission. Because there is a great number of ICD-10 codes, we categorize them into clinically meaningful groups. In total, 59 independent variables were included in the predictive model. We also developed a model with the top ten variables identified by Shapley Additive Explanations (SHAP) technique1010 Lundberg SM, Lee SI. A unified approach to interpreting model predictions. arXiv. 2017;arXiv:1705.07874v2. https://doi.org/10.48550/arXiv.1904.07204
https://doi.org/10.48550/arXiv.1904.0720... . Variables that the authors considered not to be clinically plausible predictors of LOS stay were excluded. The outcome was LOS for patients undergoing CABG. LOS in days was calculated from the difference between the date and time of admission and discharge.
Missing Data
Data from 88 patients were missing in the database. Patients with missing data were excluded from the analysis. We considered that the missing data occurred completely at random and that they represented a small part of the sample.
Model Training, Selection, and Validation
The training sample was used to train three machine learning models (random forest, extreme gradient boosting and neural networks) and three traditional regression models (Poisson regression, linear regression, negative binomial regression). To guide algorithm selection, we ranked the predictive capacities of models. The model with the best performance was selected and validated in a new sample of patients undergoing CABG at a distinct time. For training the predictive models, a sample of 9,584 patients undergoing CABG between January 2017 and December 2021 was used. For machine learning models, 30% of the training base was selected for a hold out for internal validation and excluded from the database. In the remaining portion of the database, we used the K-fold cross validation method1111 Bradshaw TJ, Huemann Z, Hu J, Rahmim A. A guide to cross-validation for artificial intelligence in medical imaging. Radiol Artif Intell. 2023;5(4): e220232. https://doi.org/10.1148/ryai.220232
https://doi.org/10.1148/ryai.220232... : the dataset was divided into ten folds, with each algorithm being trained ten times. In each iteration, one fold was utilized for validation while the remaining nine folds were used as the training dataset. The process continues until all parts have participated in both the training and validation processes. This procedure has a single parameter k, which refers to the number of groups the training dataset should be divided into for training and validation purposes. The most common values used for k range from 5 to 10. We used k = 10 in all tested models.
The model with the best performance was defined through the root mean squared and logarithmic error (RMLSE) in the training dataset. The RMLSE was chosen as our metric because it penalizes the model more severely when the predicted value is less than the actual value compared to when the predicted value is more than the actual value, but we also included other common regression metrics of the models in a supplemental materiala. The closer the RMLSE is to 0, the smaller the error of the model1212 Muhlestein WE, Akagi DS, Davies JM, Chambless LB. Predicting inpatient length of stay after brain tumor surgery: developing machine learning ensembles to improve predictive performance. Neurosurgery. 2019;85(3):384-93. https://doi.org/10.1093/neuros/nyy343
https://doi.org/10.1093/neuros/nyy343... . While it may not be the most optimal method for evaluating the predictive accuracy of machine learning models, we calculated the R² value for the top-performing model to facilitate a comparative analysis against alternative predictive models. Internal validation was performed on 30% of the training base that had been separated for this purpose. We also developed a model with the top ten variables identified by SHAP technique in the validation dataset, to improve usability.
The algorithm that exhibited the highest performance using all variables and the model employing the top ten variables were both trained on the entire training dataset. The models were then validated in a never-before-seen database of 2,627 patients who had undergone CABG between January 2022 and December 2022.
To set the hyperparameters of the chosen model, we used the Grid Search technique. A set of possible values for each hyperparameter was selected and all possible combinations of these hyperparameters were tested. The hyperparameters that obtained the best performance were used for the final model. Tuning hyperparameters prevents the model from learning solely from the presented data (avoiding overfitting and underfitting), enabling it to generalize to other possible scenarios1313 Hastie T, Tibshirani R, Friedman J. The elements of statistical learning: data mining, inference, and prediction. New York: Springer; 2008.
The SHAP technique was used to define the relative importance of each model variable in the outcome. This is done by comparing the predictive ability of a model with and without a given variable. The larger the difference between the predictions, the more important the variable is to the model. The advantage of using SHAP values lies in the fact that they add interpretability to complex models1010 Lundberg SM, Lee SI. A unified approach to interpreting model predictions. arXiv. 2017;arXiv:1705.07874v2. https://doi.org/10.48550/arXiv.1904.07204
https://doi.org/10.48550/arXiv.1904.0720... .
A calibration curve was also built to assess the predictive ability of the selected model. The main objective of a calibration curve is to assess the alignment of model predictions with the true values of the output variable. It provides an idea of how well the model's predictions are aligned with the actual values when comparing numerical predictions with the true values of the outcome variable. To do this, predictions are grouped into intervals or bins, creating groups of similar predictions. An ideal calibration curve would have all points located on the diagonal line, indicating perfect calibration of predictions with respect to actual values. By observing the calibration curve, it is possible to identify whether the model is consistently underestimating or overestimating the values, which may indicate the need for adjustments or refinements in the model1414 Alba AC, Agoritsas T, Walsh M, Hanna S, Iorio A, Devereaux PJ, et al. Discrimination and calibration of clinical prediction models: users’ guides to the medical literature. JAMA. 2017;318(14):1377-84. https://doi.org/10.1001/jama.2017.12126
https://doi.org/10.1001/jama.2017.12126... .
To describe the characteristics of patients, we used the Mann-Whitney test to compare continuous variables and Pearson's χ² test to compare categorical variables. For each of the most important variables, we compared the LOS between patients with and without that characteristic. Since the LOS did not follow a normal distribution, we employed a non-parametric test (Mann-Whitney) to assess whether the difference in hospital stay was significant between patients with the characteristic and those without it, using a significance level of 5%.
This study was approved by the Ethics and Research Committee of the Medical Sciences School of Minas Gerais, Brazil (Certificate of Presentation for Ethical Appreciation — CAEE: 29000819·0.0000·5134). It was classified as a low-risk study, since it used anonymous convenience samples extracted from a database which is used for managerial purposes. The study did not require participants to sign the informed consent form. Python software (version 3.7) was used for training and validation of the model.
RESULTS
Patient Characteristics
Data from 9,584 patients (75% male, mean age 63.2 (9.6) years) who underwent CABG between January 2017 and December 2021 were utilized for training the algorithm. Subsequently, the algorithm's performance was assessed using a new dataset comprising 2,627 patients (74.2% male, mean age 63.7 (9.4) years) who underwent CABG between January 2022 and December 2022. The flow of participants through the study is depicted in Figure 1. Baseline characteristics of both the training and validation cohorts, along with their unadjusted association with the primary outcome, are presented in Table 1. The characteristics of the training and validation groups were similar. The average duration of hospital stay was 14.8 days (standard deviation - SD = 10.3) for the training cohort and 15.9 days (SD = 12.6) for the validation sample. Out of the total surgeries considered in this study, 55.9% (6,826 patients — 5,363 on the training and 1,463 on the validation dataset) were classified as non-elective procedures. Valve replacement was performed in 11.8% (1,436 patients — 1,136 on the training and 300 on the validation dataset) of the surgeries, and 52.0% (6,345 patients — 4,398 on the training and 1,407 on the validation dataset) of the hospitalizations occurred within the Brazilian public health system (SUS). The average number of chronic health conditions per patient was 4.26 (SD = 1.97). Among the patients included in this study, 2,654 patients (21.7%) had more than five chronic health conditions. The in-hospital mortality rate for patients who underwent CABG was 6.5% (854 patients).
Model Performance
We trained three machine learning models (random forest, extreme gradient boosting and neural networks) and three traditional regression models (Poisson regression, linear regression, negative binomial regression) in the training dataset. The random forest model had the best performance and was then validated in a new dataset, followed by the extreme gradient boosting model.
The RMLSE of the random forest model was 0.412 (95%CI 0.405-0.419) on the training dataset, 0.429 (95%CI 0.417-0.440) on the hold-out and 0.454 (95%CI 0.441-0.468) on the validation on a distinct dataset collected between January 2022 and December 2022. The RMLSE of the top ten model was 0.485 (95%CI 0.470-0.500) on the validation on a distinct data set and was not significantly different from the model which included all the variables (Signed Wicoxon Test, p-value = 1). Model calibration curves on the validation dataset are shown in Figure 2. The model with the top ten variables is available for use through the link: https://rvsm.reiks.tec.br/
Calibration plot of validation dataset. Calibration curve demonstrating the predicted lenght of stay relative to actual lenght of stay for the model.
Feature Importance
The SHAP analysis showed that the variables with the greatest impact on the patient's LOS in decreasing order of strength were non-elective surgery, admission to a public hospital, presence of heart failure, age, and Non-STE-ACS. The strongest variables for model prediction were similar in the training and validation datasets (Figure 3).
A summary plot showing the variables’ relative importance and their effect on the predicted outcomes within the model in the validation dataset is shown (Figure 4).
Summary plot showing variables’ relative importance and their effect on the predicted outcomes within the model in the validation dataset.
DISCUSSION
The machine-learning system may experience diminished performance when there is a mismatch between the dataset with which it was developed and the data on which it is deployed, in a process called dataset shift1515 Finlayson SG, Subbaswamy A, Singh K, Bowers J, Kupke A, Zittrain J, et al. The clinician and dataset shift in artificial intelligence. N Engl J Med. 2021;385(3):283-6. https://doi.org/10.1056/nejmc2104626
https://doi.org/10.1056/nejmc2104626...
16 Subbaswamy A, Saria S. From development to deployment: dataset shift, causality, and shift-stable models in health AI. Biostatistics. 2020;21(2):345-52. https://doi.org/10.1093/biostatistics/kxz041
https://doi.org/10.1093/biostatistics/kx... -1717 Saria S, Subbaswamy A. Tutorial: safe and reliable machine learning. arXiv. 2019;arXiv:1904.07204. https://doi.org/10.48550/arXiv.1904.07204
https://doi.org/10.48550/arXiv.1904.0720... . A machine learning model trained on a specific population or healthcare structure can underperform when applied in a new setting88 Austin PC, van Klaveren D, Vergouwe Y, Nieboer D, Lee DS, Steyerberg EW. Geographic and temporal validity of prediction models: different approaches were useful to examine model performance. J Clin Epidemiol. 2016;79:76-85. https://doi.org/10.1016/j.jclinepi.2016.05.007
https://doi.org/10.1016/j.jclinepi.2016.... . To the best of our knowledge, this is the first predictive model developed to estimate LOS in patients undergoing CABG using a Brazilian database.
Our study used an administrative database to produce an machine learning model capable of estimating LOS in patients undergoing CABG with RMLSE = 0.454 for the validation of the model including all variables and RMLSE = 0.485 for the validation of the top ten model. The model can reliably predict the LOS in groups of patients, as demonstrated in the calibration curve. The algorithm was developed using data collected from multiple centers across all regions of Brazil, encompassing a large population. Through direct comparison of different predictive models of machine learning and traditional statistics, it was possible to avoid bias associated with algorithm selection. The random forest technique produced the model with the lowest error. Other strengths of our study include a large sample size and the utilization of LOS as a continuous variable rather than a categorical one. A model with the top ten variables was not significantly different from the model which included all the variables, improving usability. Our algorithm with the top ten variables is freely accessible via a provided weblink, facilitating the dissemination of the algorithm to centers that can utilize a model adjusted to local realities and further validate it.
Traditional statistics models may incorrectly presuppose linear interactions among the variables influencing outcomes, thereby constraining the effectiveness of predictive models1818 Ryo M, Rillig MC. Statistically reinforced machine learning for nonlinear patterns and variable interactions. Ecosphere. 2017;8(11):e01976. https://doi.org/10.1002/ecs2.1976
https://doi.org/10.1002/ecs2.1976... . In contrast, machine learning approaches consider the dynamic interplay of variables in ways that are nonlinear and nonparametric, leading to enhanced prediction tools1919 Breiman L. Random forests. Mach Learn. 2001;45:5-32. https://doi.org/10.1023/a:1010933404324
https://doi.org/10.1023/a:1010933404324... . Machine learning employs more flexible techniques that facilitate the incorporation of extensive volumes of multidimensional data2020 Ambale-Venkatesh B, Yang X, Wu CO, Liu K, Hundley WG, McClelland R, et al. Cardiovascular event prediction by machine learning: the multi-ethnic study of atherosclerosis. Circ Res. 2017;121(9):1092-101. https://doi.org/10.1161/CIRCRESAHA.117.311312
https://doi.org/10.1161/CIRCRESAHA.117.3... . Supporting this assertion, the random forest technique yielded a model with better performance compared to the conventional statistical models (Poisson regression, linear regression, negative binomial regression) we developed in our database.
The random forest algorithm is a powerful Machine learning technique that can be used for both regression and classification models. It is a special case of an ensemble method called «bagging» (bootstrap aggregating). Bagging involves training multiple models on different random subsets of the original dataset and combining their predictions to produce a more robust result. In the case of random forest, bagging is applied to a set of decision trees. Multiple training subsets are created from the original dataset using a technique called «bootstrap sampling» for each subset. For each training subset, a decision tree is trained. The predictions from all individual trees are combined to form a final prediction2020 Ambale-Venkatesh B, Yang X, Wu CO, Liu K, Hundley WG, McClelland R, et al. Cardiovascular event prediction by machine learning: the multi-ethnic study of atherosclerosis. Circ Res. 2017;121(9):1092-101. https://doi.org/10.1161/CIRCRESAHA.117.311312
https://doi.org/10.1161/CIRCRESAHA.117.3... . In the case of random forest, the combination often involves averaging (for regression) or voting (for classification) the predictions of individual trees. The key feature of random forest is the introduction of randomness during tree construction. This includes the random selection of features at each node split and the use of random feature subsets for each tree. The introduced randomness helps reduce the correlation between individual trees, making the ensemble more diverse and, therefore, less prone to overfitting. This diversity is fundamental for improving the generalization of the model1919 Breiman L. Random forests. Mach Learn. 2001;45:5-32. https://doi.org/10.1023/a:1010933404324
https://doi.org/10.1023/a:1010933404324... .
Most of the studies examining LOS in patients who undergo CABG solely focused on the duration of their stay in the intensive care unit2121 Shahian DM, Edwards FH, Ferraris VA, Haan CK, Rich JB, Normand SLT, et al. Quality measurement in adult cardiac surgery: part 1 - conceptual framework and measure selection. Ann Thorac Surg. 2007;83(4 Suppl):S3-12. https://doi.org/10.1016/j.athoracsur.2007.01.053
https://doi.org/10.1016/j.athoracsur.200...
22 Cocker JD, Messaoudi J, Stockman BA, Bossaert LL, Rodrigus IER. Preoperative prediction of intensive care unit stay following cardiac surgery. Eur J Cardiothorac Surg. 2011;39(1):60-7. https://doi.org/10.1016/j.ejcts.2010.04.015
https://doi.org/10.1016/j.ejcts.2010.04....
23 Nilsson J, Algotsson L, Hoglund P, Luhrs C, Brandt J. EuroSCORE predicts intensive care unit stay and costs of open-heart surgery. Ann Thorac Surg. 2004;78(5):1528-34. https://doi.org/10.1016/j.athoracsur.2004.04.060
https://doi.org/10.1016/j.athoracsur.200... -2424 Peterson ED, Coombs LP, Ferguson TB, Shroyer AL, DeLong ER, Grover FL, et al. Hospital variability in length of stay after coronary artery bypass surgery: results from the Society of Thoracic Surgeon's National Cardiac Database. Ann Thorac Surg. 2002;74(2):464-73. https://doi.org/10.1016/s0003-4975(02)03694-9
https://doi.org/10.1016/s0003-4975(02)03... . Alshakhs et al.2525 Alshakhs F, Alharthi H, Aslam N, Khan IU, Elasheri M. Predicting postoperative length of stay for isolated coronary artery bypass graft patients using machine learning. Int J Gen Med. 2020;13:751-62. https://doi.org/10.2147/IJGM.S250334
https://doi.org/10.2147/IJGM.S250334... conducted a study with a predictive model using Machine learning for hospital LOS after CABG in Saudi Arabia with 621 patients. In contrast to our study, LOS was categorized into high and low risk. The random forest technique also produced the model with the best accuracy. Osnabrugge et al.2626 Osnabrugge RL, Speir AM, Head SJ, Jones PG, Ailawadi G, Fonner CE, et al. Prediction of costs and length of stay in coronary artery bypass grafting. Ann Thorac Surg. 2014;98(4):1286-93. https://doi.org/10.1016/j.athoracsur.2014.05.073
https://doi.org/10.1016/j.athoracsur.201... developed a predictive model for hospital LOS and costs in patients undergoing CABG using linear regression analysis in a large database. Triana et al.2727 Triana AJ, Vyas R, Shah AS, Tiwari V. Predicting length of stay of coronary artery bypass grafting patients using machine learning. J Surg Res. 2021;264:68-75. https://doi.org/10.1016/j.jss.2021.02.003
https://doi.org/10.1016/j.jss.2021.02.00... developed a Machine learning model for LOS using data from a single institution's Society of Thoracic Surgeons (STS) Registry. Our model had a better R² coefficient (R² = 0.277 for the validation of the model including all variables, R² = 0.1933 for the top ten model) than these other models that considered only preoperative variables (R² = 0.10 and R² = 0.058, respectively). Both studies also developed a model including preoperative, intraoperative, and postoperative variables, with R² values of 0.51 and 0.232, respectively. Recognizing that predictions should be adjusted for variations in risk factors as opposed to differences in outcomes, we made a choice to exclude these variables. This decision aligns with a key focal point of our model, which is benchmarking to be used in management improvement programs and the establishment of pay-for-performance mechanisms. It is important to highlight that the utilization of R² might not be the most suitable approach for assessing the predictive precision of intricate models such as random forests. R² quantifies the extent to which independent variables in the model account for the variability in the dependent variable. However, machine learning methods can grasp non-linear and interactive elements—features that conventional linear regression models might miss. This distinction leads to R2's limitations when it comes to capturing the nuances of complex relationships that random forests can handle2828 Fávero LP, Belfiore P. Manual de análise de dados: estatística e modelagem multivariada com Excel, SPSS e Stata. Rio de Janeiro: Elsevier; 2017..
The risk factors for prolonged LOS that we found are similar to previous studies. Osnabrugge et al.2626 Osnabrugge RL, Speir AM, Head SJ, Jones PG, Ailawadi G, Fonner CE, et al. Prediction of costs and length of stay in coronary artery bypass grafting. Ann Thorac Surg. 2014;98(4):1286-93. https://doi.org/10.1016/j.athoracsur.2014.05.073
https://doi.org/10.1016/j.athoracsur.201... and Almashrafi et al.2929 Almashrafi A, Alsabti H, Mukaddirov M, Balan B, Aylin P. Factors associated with prolonged length of stay following cardiac surgery in a major referral hospital in Oman: a retrospective observational study. BMJ Open. 2016;6(6):e010764. https://doi.org/10.1136/bmjopen-2015-010764
https://doi.org/10.1136/bmjopen-2015-010... also described non-elective surgery as a risk factor for increased LOS. Patients with heart failure also had longer LOS in the studies conducted by Almashrafi et al.2929 Almashrafi A, Alsabti H, Mukaddirov M, Balan B, Aylin P. Factors associated with prolonged length of stay following cardiac surgery in a major referral hospital in Oman: a retrospective observational study. BMJ Open. 2016;6(6):e010764. https://doi.org/10.1136/bmjopen-2015-010764
https://doi.org/10.1136/bmjopen-2015-010... and Lazar et al.3030 Lazar HL, Fitzgerald C, Gross S, Heeren T, Aldea GS, Shemin RJ. Determinants of length of stay after coronary artery bypass graft surgery. Circulation. 1995;92(9 Suppl):II20-4. https://doi.org/10.1161/01.cir.92.9.20
https://doi.org/10.1161/01.cir.92.9.20... Combined CABG and valve surgery was also found to increase LOS in the study conducted by Almashrafi et al.2929 Almashrafi A, Alsabti H, Mukaddirov M, Balan B, Aylin P. Factors associated with prolonged length of stay following cardiac surgery in a major referral hospital in Oman: a retrospective observational study. BMJ Open. 2016;6(6):e010764. https://doi.org/10.1136/bmjopen-2015-010764
https://doi.org/10.1136/bmjopen-2015-010... Triana et al.2727 Triana AJ, Vyas R, Shah AS, Tiwari V. Predicting length of stay of coronary artery bypass grafting patients using machine learning. J Surg Res. 2021;264:68-75. https://doi.org/10.1016/j.jss.2021.02.003
https://doi.org/10.1016/j.jss.2021.02.00... also found age to increase LOS. One of the main risk factors for prolonged LOS found in our study was performing CABG in the public health system, which may reflect the influence of the low income of the population assisted in this system, or its poorer quality of structure and hospital processes. We did not find another study that analyzed this variable.
Our study has several limitations. Despite using a database collected in a different period to validate the model, this was a retrospective database collected from the same institutions where the model was trained. Furthermore, the characteristics of the patients in the training and validation samples are very similar, probably leading to overly optimistic results. It would be important to carry out an external and prospective validation, ideally assessing the application of the model by clinicians in clinical practice. Access to other variables could increase the explanatory power of our model. As this is a database collected for administrative purposes, we did not have access to risk factors that influenced LOS in other studies, such as the use of intra-aortic balloon pump, laboratory results, use of intravenous nitroglycerin and pulmonary artery systolic pressure2222 Cocker JD, Messaoudi J, Stockman BA, Bossaert LL, Rodrigus IER. Preoperative prediction of intensive care unit stay following cardiac surgery. Eur J Cardiothorac Surg. 2011;39(1):60-7. https://doi.org/10.1016/j.ejcts.2010.04.015
https://doi.org/10.1016/j.ejcts.2010.04.... ,2424 Peterson ED, Coombs LP, Ferguson TB, Shroyer AL, DeLong ER, Grover FL, et al. Hospital variability in length of stay after coronary artery bypass surgery: results from the Society of Thoracic Surgeon's National Cardiac Database. Ann Thorac Surg. 2002;74(2):464-73. https://doi.org/10.1016/s0003-4975(02)03694-9
https://doi.org/10.1016/s0003-4975(02)03... ,2525 Alshakhs F, Alharthi H, Aslam N, Khan IU, Elasheri M. Predicting postoperative length of stay for isolated coronary artery bypass graft patients using machine learning. Int J Gen Med. 2020;13:751-62. https://doi.org/10.2147/IJGM.S250334
https://doi.org/10.2147/IJGM.S250334... ,2727 Triana AJ, Vyas R, Shah AS, Tiwari V. Predicting length of stay of coronary artery bypass grafting patients using machine learning. J Surg Res. 2021;264:68-75. https://doi.org/10.1016/j.jss.2021.02.003
https://doi.org/10.1016/j.jss.2021.02.00... . Building a specific database for conducting studies in patients undergoing cardiac surgery could further enhance the predictive capacity of the model. Other studies have indicated that postoperative factors and adverse events are significant contributors to increased length of stay (LOS)2626 Osnabrugge RL, Speir AM, Head SJ, Jones PG, Ailawadi G, Fonner CE, et al. Prediction of costs and length of stay in coronary artery bypass grafting. Ann Thorac Surg. 2014;98(4):1286-93. https://doi.org/10.1016/j.athoracsur.2014.05.073
https://doi.org/10.1016/j.athoracsur.201... ,2727 Triana AJ, Vyas R, Shah AS, Tiwari V. Predicting length of stay of coronary artery bypass grafting patients using machine learning. J Surg Res. 2021;264:68-75. https://doi.org/10.1016/j.jss.2021.02.003
https://doi.org/10.1016/j.jss.2021.02.00... . However, incorporating such variables would hinder the model suitability for hospital benchmarking and pay-for-performance measures. We excluded non-survivors to predict LOS. Since the characteristics of patients who survive are different from those of patients who die in hospital, our findings could be altered in a population including non-survivors. This affects the generalizability of our algorithm.
Hospital LOS is an important marker of quality of care. The developed model can be used to compare the observed and expected LOS to generate LOS indices that could be used as markers of efficiency, allowing healthcare systems to manage resources and optimize hospital payment models. In addition, identifying patients with the potential for prolonged hospitalization can help the institution in managing beds, scheduling surgeries, and allocating resources. The predictive model can also help in the psychological preparation and planning of the patient and his family.
Funding:
No funding was received for conducting this study.- a
REFERENCES
- 1Gomes WJ, Moreira RS, Zilli AC, Bettiati Jr LC, Figueira FAMS, D’ Azevedo SSP, et al. The brazilian registry of adult patient undergoing cardiovascular surgery, the BYPASS project: results of the first 1,722 patients. Braz J Cardiovasc Surg. 2017;32(2):71-6. https://doi.org/10.21470/1678-9741-2017-0053
» https://doi.org/10.21470/1678-9741-2017-0053 - 2Organisation for Economic Cooperation and Development. Hospital discharges and average length of stay. In: Health at a Glance 2021: OECD Indicators. Paris: OECD Publishing; 2021. https://doi.org/10.1787/18faaea9-en
» https://doi.org/10.1787/18faaea9-en - 3Stone K, Zwiggelaar R, Jones P, Mac Parthaláin NM. A systematic review of the prediction of hospital length of stay: Towards a unified framework. PLOS Digit Health. 2022;1(4):e0000017. https://doi.org/10.1371/journal.pdig.0000017
» https://doi.org/10.1371/journal.pdig.0000017 - 4Deo RC. Machine learning in medicine: Will this time be different? Circulation. 2020;142(16):1521-3. https://doi.org/10.1161/circulationaha.120.050583
» https://doi.org/10.1161/circulationaha.120.050583 - 5Obermeyer Z, Emanuel EJ. Predicting the future – big data, machine learning, and clinical medicine. N Engl J Med. 2016;375(13):1216-9. https://doi.org/10.1056/NEJMp1606181
» https://doi.org/10.1056/NEJMp1606181 - 6Rajkomar A, Dean J, Kohane I. Machine learning in medicine. N Engl J Med. 2019;380(14):1347-58. https://doi.org/10.1056/NEJMra1814259
» https://doi.org/10.1056/NEJMra1814259 - 7Stevens LM, Mortazavi BJ, Deo RC, Curtis L, Kao DP. Recommendations for reporting machine learning analyses in clinical research. Circ Cardiovasc Qual Outcomes. 2020;13(10):e006556. https://doi.org/10.1161/CIRCOUTCOMES.120.006556
» https://doi.org/10.1161/CIRCOUTCOMES.120.006556 - 8Austin PC, van Klaveren D, Vergouwe Y, Nieboer D, Lee DS, Steyerberg EW. Geographic and temporal validity of prediction models: different approaches were useful to examine model performance. J Clin Epidemiol. 2016;79:76-85. https://doi.org/10.1016/j.jclinepi.2016.05.007
» https://doi.org/10.1016/j.jclinepi.2016.05.007 - 9Moons KGM, Altman DG, Reitsma JB, Loannidis JPA, Macaskill P, Steyerberg EW, et al. Transparent Reporting of a multivariable prediction model for Individual Prognosis Or Diagnosis (TRIPOD): Explanation and Elaboration. Translation in to Russian. Dig Diagn. 2022;3(3):232-322. https://doi.org/10.17816/DD110794
» https://doi.org/10.17816/DD110794 - 10Lundberg SM, Lee SI. A unified approach to interpreting model predictions. arXiv. 2017;arXiv:1705.07874v2. https://doi.org/10.48550/arXiv.1904.07204
» https://doi.org/10.48550/arXiv.1904.07204 - 11Bradshaw TJ, Huemann Z, Hu J, Rahmim A. A guide to cross-validation for artificial intelligence in medical imaging. Radiol Artif Intell. 2023;5(4): e220232. https://doi.org/10.1148/ryai.220232
» https://doi.org/10.1148/ryai.220232 - 12Muhlestein WE, Akagi DS, Davies JM, Chambless LB. Predicting inpatient length of stay after brain tumor surgery: developing machine learning ensembles to improve predictive performance. Neurosurgery. 2019;85(3):384-93. https://doi.org/10.1093/neuros/nyy343
» https://doi.org/10.1093/neuros/nyy343 - 13Hastie T, Tibshirani R, Friedman J. The elements of statistical learning: data mining, inference, and prediction. New York: Springer; 2008
- 14Alba AC, Agoritsas T, Walsh M, Hanna S, Iorio A, Devereaux PJ, et al. Discrimination and calibration of clinical prediction models: users’ guides to the medical literature. JAMA. 2017;318(14):1377-84. https://doi.org/10.1001/jama.2017.12126
» https://doi.org/10.1001/jama.2017.12126 - 15Finlayson SG, Subbaswamy A, Singh K, Bowers J, Kupke A, Zittrain J, et al. The clinician and dataset shift in artificial intelligence. N Engl J Med. 2021;385(3):283-6. https://doi.org/10.1056/nejmc2104626
» https://doi.org/10.1056/nejmc2104626 - 16Subbaswamy A, Saria S. From development to deployment: dataset shift, causality, and shift-stable models in health AI. Biostatistics. 2020;21(2):345-52. https://doi.org/10.1093/biostatistics/kxz041
» https://doi.org/10.1093/biostatistics/kxz041 - 17Saria S, Subbaswamy A. Tutorial: safe and reliable machine learning. arXiv. 2019;arXiv:1904.07204. https://doi.org/10.48550/arXiv.1904.07204
» https://doi.org/10.48550/arXiv.1904.07204 - 18Ryo M, Rillig MC. Statistically reinforced machine learning for nonlinear patterns and variable interactions. Ecosphere. 2017;8(11):e01976. https://doi.org/10.1002/ecs2.1976
» https://doi.org/10.1002/ecs2.1976 - 19Breiman L. Random forests. Mach Learn. 2001;45:5-32. https://doi.org/10.1023/a:1010933404324
» https://doi.org/10.1023/a:1010933404324 - 20Ambale-Venkatesh B, Yang X, Wu CO, Liu K, Hundley WG, McClelland R, et al. Cardiovascular event prediction by machine learning: the multi-ethnic study of atherosclerosis. Circ Res. 2017;121(9):1092-101. https://doi.org/10.1161/CIRCRESAHA.117.311312
» https://doi.org/10.1161/CIRCRESAHA.117.311312 - 21Shahian DM, Edwards FH, Ferraris VA, Haan CK, Rich JB, Normand SLT, et al. Quality measurement in adult cardiac surgery: part 1 - conceptual framework and measure selection. Ann Thorac Surg. 2007;83(4 Suppl):S3-12. https://doi.org/10.1016/j.athoracsur.2007.01.053
» https://doi.org/10.1016/j.athoracsur.2007.01.053 - 22Cocker JD, Messaoudi J, Stockman BA, Bossaert LL, Rodrigus IER. Preoperative prediction of intensive care unit stay following cardiac surgery. Eur J Cardiothorac Surg. 2011;39(1):60-7. https://doi.org/10.1016/j.ejcts.2010.04.015
» https://doi.org/10.1016/j.ejcts.2010.04.015 - 23Nilsson J, Algotsson L, Hoglund P, Luhrs C, Brandt J. EuroSCORE predicts intensive care unit stay and costs of open-heart surgery. Ann Thorac Surg. 2004;78(5):1528-34. https://doi.org/10.1016/j.athoracsur.2004.04.060
» https://doi.org/10.1016/j.athoracsur.2004.04.060 - 24Peterson ED, Coombs LP, Ferguson TB, Shroyer AL, DeLong ER, Grover FL, et al. Hospital variability in length of stay after coronary artery bypass surgery: results from the Society of Thoracic Surgeon's National Cardiac Database. Ann Thorac Surg. 2002;74(2):464-73. https://doi.org/10.1016/s0003-4975(02)03694-9
» https://doi.org/10.1016/s0003-4975(02)03694-9 - 25Alshakhs F, Alharthi H, Aslam N, Khan IU, Elasheri M. Predicting postoperative length of stay for isolated coronary artery bypass graft patients using machine learning. Int J Gen Med. 2020;13:751-62. https://doi.org/10.2147/IJGM.S250334
» https://doi.org/10.2147/IJGM.S250334 - 26Osnabrugge RL, Speir AM, Head SJ, Jones PG, Ailawadi G, Fonner CE, et al. Prediction of costs and length of stay in coronary artery bypass grafting. Ann Thorac Surg. 2014;98(4):1286-93. https://doi.org/10.1016/j.athoracsur.2014.05.073
» https://doi.org/10.1016/j.athoracsur.2014.05.073 - 27Triana AJ, Vyas R, Shah AS, Tiwari V. Predicting length of stay of coronary artery bypass grafting patients using machine learning. J Surg Res. 2021;264:68-75. https://doi.org/10.1016/j.jss.2021.02.003
» https://doi.org/10.1016/j.jss.2021.02.003 - 28Fávero LP, Belfiore P. Manual de análise de dados: estatística e modelagem multivariada com Excel, SPSS e Stata. Rio de Janeiro: Elsevier; 2017.
- 29Almashrafi A, Alsabti H, Mukaddirov M, Balan B, Aylin P. Factors associated with prolonged length of stay following cardiac surgery in a major referral hospital in Oman: a retrospective observational study. BMJ Open. 2016;6(6):e010764. https://doi.org/10.1136/bmjopen-2015-010764
» https://doi.org/10.1136/bmjopen-2015-010764 - 30Lazar HL, Fitzgerald C, Gross S, Heeren T, Aldea GS, Shemin RJ. Determinants of length of stay after coronary artery bypass graft surgery. Circulation. 1995;92(9 Suppl):II20-4. https://doi.org/10.1161/01.cir.92.9.20
» https://doi.org/10.1161/01.cir.92.9.20
Publication Dates
- Publication in this collection
16 Sept 2024 - Date of issue
2024
History
- Received
19 Feb 2024 - Accepted
31 Mar 2024