Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Survival machine learning methods for mortality prediction after heart transplantation in the contemporary era

  • Lathan Liou,

    Roles Conceptualization, Formal analysis, Investigation, Methodology, Software, Visualization, Writing – original draft, Writing – review & editing

    Affiliations Department of Epidemiology, Harvard T.H. Chan School of Public Health, Boston, Massachusetts, United States of America, Department of Genetics and Genomics Sciences, Icahn School of Medicine at Mount Sinai, New York, New York, United States of America

  • Elizabeth Mostofsky,

    Roles Data curation, Project administration, Resources, Writing – review & editing

    Affiliation Department of Epidemiology, Harvard T.H. Chan School of Public Health, Boston, Massachusetts, United States of America

  • Laura Lehman,

    Roles Writing – review & editing

    Affiliations Department of Epidemiology, Harvard T.H. Chan School of Public Health, Boston, Massachusetts, United States of America, Harvard Medical School, Boston, Massachusetts, United States of America, Department of Neurology, Boston Children’s Hospital, Boston, Massachusetts, United States of America

  • Soziema Salia,

    Roles Writing – review & editing

    Affiliations Department of Epidemiology, Harvard T.H. Chan School of Public Health, Boston, Massachusetts, United States of America, Department of Internal Medicine, Cape Coast Teaching Hospital, Cape Coast, Ghana

  • Francisco J. Barrera,

    Roles Writing – review & editing

    Affiliation Department of Epidemiology, Harvard T.H. Chan School of Public Health, Boston, Massachusetts, United States of America

  • Ying Wei,

    Roles Writing – review & editing

    Affiliation Department of Epidemiology, Harvard T.H. Chan School of Public Health, Boston, Massachusetts, United States of America

  • Amal Cheema,

    Roles Writing – review & editing

    Affiliations Department of Epidemiology, Harvard T.H. Chan School of Public Health, Boston, Massachusetts, United States of America, Geisel School of Medicine, Dartmouth College, Hanover, New Hampshire, United States of America

  • Anuradha Lala,

    Roles Methodology, Supervision, Writing – review & editing

    Affiliation Zena and Michael A. Wiener Cardiovascular Institute and Department of Population Health Science and Policy, Mount Sinai, New York, New York, United States of America

  • Andrew Beam ,

    Contributed equally to this work with: Andrew Beam, Murray A. Mittleman

    Roles Methodology, Supervision, Writing – review & editing

    Affiliation Department of Epidemiology, Harvard T.H. Chan School of Public Health, Boston, Massachusetts, United States of America

  • Murray A. Mittleman

    Contributed equally to this work with: Andrew Beam, Murray A. Mittleman

    Roles Conceptualization, Investigation, Methodology, Project administration, Resources, Supervision, Writing – review & editing

    mmittlem@hsph.harvard.edu

    Affiliations Department of Epidemiology, Harvard T.H. Chan School of Public Health, Boston, Massachusetts, United States of America, Harvard Medical School, Boston, Massachusetts, United States of America, Department of Medicine, Division of Cardiovascular Medicine, Beth Israel Deaconess Medical Center, Boston, Massachusetts, United States of America

Abstract

Although prediction models for heart transplantation outcomes have been developed previously, a comprehensive benchmarking of survival machine learning methods for mortality prognosis in the most contemporary era of heart transplants following the 2018 donor heart allocation policy change is warranted. This study assessed seven statistical and machine learning algorithms–Lasso, Ridge, Elastic Net, Cox Gradient Boost, Extreme Gradient Boost Linear, Extreme Gradient Boost Tree, and Random Survival Forests in a post-policy cohort of 7,160 adult heart-only transplant recipients in the Scientific Registry of Transplant Recipients (SRTR) database who received their first transplant on or after October 18, 2018. A cross-validation framework was designed in mlr. Model performance was also compared in a seasonally-matched pre-policy cohort. In the post-policy cohort, Random Survival Forests and Cox Gradient Boost had the highest performances with C-indices of 0.628 and 0.627. The relative importance of some predictive variables differed between the pre-policy and post-policy cohorts, such as the absence of ECMO in the post-policy cohort. Survival machine learning models provide reasonable prediction of 1-year posttransplant mortality outcomes and continual updating of prediction models is warranted in the contemporary era.

Introduction

Orthotopic heart transplantation (HTx) remains the gold standard therapy for end-stage heart failure. Careful patient selection has allowed for a contemporary median survival of 12.5 years with an additional 2 years conditional upon 1-year survival post-transplant [1]. Accurate determination of which patient characteristics should be considered contraindications to transplantation for optimal outcomes remains one of the greatest challenges in our field. Historically, heart transplant risk prediction models have been generated using multivariable logistic regression [2] or Cox proportional hazard regression [3,4] and have yielded the identification of certain key risk factors. The limitation of such models lies in the assumption of purely linear relationships that do not account for more complex associations. Although many factors associated with post-transplantation survival have been identified [5,6], there remains variability in models exploring more complex relationships. Recently, several studies have attempted to improve the prediction of heart transplant outcomes using machine learning (ML) methods [711] and survival ML methods [1214] in both adult and pediatric populations. Compared to traditional regression‐based modeling approaches, ML algorithms can capture more complex interactions between variables. Accounting for time-to-event information tends to result in more statistically powerful prediction estimates [15]. Furthermore, posttransplantation survival is known to be non-proportional across different strata, making conventional statistical methods such as Cox proportional hazards regression potentially invalid and less powerful for assessing post-transplantation mortality among heart transplant recipients.

Another potentially important consideration is the revision of the donor heart allocation system on October 18, 2018, to “reduce wait-list mortality, enhance geographic organ sharing, and improve organ distribution equity” [16]. Although it has been shown that 1-year posttransplant survival under the new heart allocation policy was not significantly different from before [17], it is possible that both the clinical profile of transplant recipients and clinical care practices have shifted in the almost 5 years since policy implementation [18]. Thus, in this study, we explore and benchmark the performance of machine learning survival algorithms in the post-2018 transplant era (henceforth referred to as “post-policy”) to better understand prognostic factors associated with one-year mortality. We compare these results to a parallel analysis run in a seasonally-matched pre-2018 transplant era cohort (henceforth referred to as “pre-policy”).

Methods

Data source

We used data from the Scientific Registry of Transplant Recipients (SRTR). The SRTR data system includes data on all donor, waitlisted candidates, and transplant recipients in the US, submitted by the Organ Procurement and Transplantation Network (OPTN) members. The data is anonymized such that it is not possible to identify individual participants. The Health Resources and Services Administration (HRSA), U.S. Department of Health and Human Services, provides oversight of the activities of the OPTN and SRTR contractors. We had approval from SRTR under DUA #9899 to conduct this research study. Informed consent is obtained by OPTN for living donors. As formally declared by the UNOS Ethics Committee, no organs were procured from prisoners. We confirm that all research was performed in accordance with relevant guidelines/regulations. The data was last accessed on May 5, 2023. This study was reviewed and informed consent was waived by the Institutional Review Board at Harvard T.H. Chan School of Public Health due to the retrospective and anonymized nature of this registry dataset.

Study population

We excluded recipients <18 years of age at transplant (n = 1,275) and an additional 8,187 recipients of multi-organ transplants because allocation criteria are different for multi-organ and pediatric candidates. This study includes 7,160 adult heart-only transplant recipients who received their first transplant on or after October 18, 2018, with at least one recorded follow-up visit. We had follow-up data up until June 3, 2021. This study was reviewed and approved by the Institutional Review Board at the Harvard T.H. Chan School of Public Health.

Data preparation

The outcome of interest was one-year all-cause mortality assessed from the time of heart transplantation to death or end of the one-year follow-up. Survival data for recipients in both cohorts were administratively censored at 1 year after transplant to prevent bias from the differential length of follow-up between cohorts. For each cohort, we split 90% as a training set and 10% as a holdout set. In order of operation, we performed z-score standardization on continuous variables, multiple imputation, and one-hot encoding on categorical variables separately in the validation set and holdout set to minimize data leakage. We included predictor variables based on expert opinion and literature review. We excluded potential predictors with over 20% missingness. For those with less than 20% missingness, we used multiple imputation with chained equations with 5 iterations using the R package mice [19], which preserves the associations in the data and the uncertainty in those associations. We assumed that the data are missing at random and that censoring is non-informative. We used one-hot encoding to convert categorical variables into binary variables for each level of each categorical feature. Certain variables that had many levels with only a small number of observations were combined (Supplementary Appendix). This data pre-processing led to a final set of 75 demographic, clinical, recipient, waitlist, donor, and procedural variables (114 after one-hot encoding), of which 9 were categorical, 44 binary, and the remaining 22 numeric (S1 Table in S1 File).

We also prepared a seasonally matched pre-policy cohort of patients transplants from November 1, 2014, with follow-up time to June 3, 2017, as a sensitivity analysis to account for known seasonal trends in decreased donor heart donation [20]. This cohort ends 1 year before policy implementation to avoid bias from any anticipatory practice changes before the policy implementation in October 2018.

Machine learning analysis

We compared a Cox proportional hazards model and 7 machine learning algorithms. Lasso, Elastic Net, and Ridge regression were selected as the penalized Cox regression methods [21,22]. Boosted Cox regression methods included a Cox model with gradient boosting (Cox Boost) [23], Extreme Gradient Boosting (XGBoost) [24] with linear model-based boosting (XGBL), and tree-based boosting XGBoost (XGBT) [25]. Random Survival Forests (RSF) [26] was chosen as the tree-based ensemble method. Details on the models and selected parameters are described in the Supplementary Appendix. Each model was trained using a nested 5-repeat, 5-fold cross-validation within the validation cohort (80/20 split), and tuning was automatically performed for the machine learning algorithms to select optimal hyperparameters. A random search with 25 iterations was used to select values for the hyperparameters in the inner loop and model performance was evaluated in the outer loop. For each method, we took the parameters from the best-performing model within the cross-validation and used the entirety of the training cohort to produce the final model. The final model was then tested on the holdout set. The R package mlr (Machine Learning in R) [27] was used as a framework to carry out benchmarking of these experiments. A graphical schematic of our benchmark framework is presented in Fig 1. To evaluate model performance, we calculated both the cross-validated and holdout concordance index (Harrell’s C-index). The C-index measures the proportion of pairs where the observation with the higher survival time has the higher probability of survival as predicted by the model. A C-index of 1 means the predictions were perfect: higher-risk patients are ranked ahead of lower-risk patients. A C-index of 0.5 means the predictions are random: half the time higher-risk patients are correctly ranked ahead, half the time lower-risk patients are incorrectly ranked ahead. Variable weights were computed as the beta coefficients from regression-based models and importance from boosted and ensemble models. Variable relevance was determined as the number of times it was selected across 25 cross-validation iterations. Variable significance was defined as either the empirical cross-validated 95% confidence interval of the hazard ratio not crossing 1 or the importance weight not crossing 0. To assess whether there were statistical differences between cross-validated models, we calculated the corrected resampled paired t-test, proposed by Nadeau and Bengio [28], using the R package correctR. A standard t-test on results from a repeated k-fold cross-validation inflates the Type I error is inflated because the violation of the independence assumption leads to an underestimation in the variance. A Bonferroni adjustment was applied to p-values to control the false discovery rate at 0.05. All analyses were performed using R 4.3.0. The code used in the analysis can be found at https://github.com/latlio/srtr_mlr.

thumbnail
Fig 1. Schematic of machine learning benchmarking study design.

https://doi.org/10.1371/journal.pone.0313600.g001

Results

Cohort characteristics

In the post-policy cohort, 506 patients died within one year of their heart transplant during the study period from October 18, 2018, to June 3, 2021. The crude one-year survival was 90.9% (95% CI: 90.2%, 91.7%) with 5864 total person-years of follow-up time. The average recipient age was 53.5, and the average donor age was 32.5. A total of 72.5% (n = 5193) of the cohort was male, and 72.1% (n = 5162) of the cohort was white. 27.1% had diabetes. 37.6% of patients were bridged with ventricular assist device (VAD), 4.9% of patients were bridged with extracorporeal membrane oxygenation (ECMO), and 26.8% of patients were bridged with intra-aortic balloon pump therapy (IABP). Patients spent a median of 33 days on the wait-list and received organs that spent 3.4 hours of ischemia time on average during transport. A comprehensive summary of demographic, clinical, recipient, waitlist, donor, and procedural variables for the post-policy cohort is detailed in Table 1. We note that in the pre-policy cohort, 512 patients died within one year, and the crude one-year survival rate was 91.5% (95% CI: 90.8%, 92.2%).

thumbnail
Table 1. Baseline characteristics of heart transplant recipients from October 18, 2018 to June 3, 2021.

https://doi.org/10.1371/journal.pone.0313600.t001

Machine learning benchmark performance

In the post-policy holdout data, RSF performed the best with a C-index of 0.628, while Cox performed the worst with a C-index of 0.585 (Table 2). Cox Boost performed similarly well with a C-index of 0.627. Both Lasso and Elastic Net performed moderately well in the holdout with C-indices of 0.613 each, although their performance in the cross-validation was poor with cross-validated mean C-indices of 0.516 and 0.508 respectively (S3 Table in S1 File). The average cross-validated C-index ranged from 0.615 to 0.508 (Fig 2). Although XGBL had the highest cross-validation C-index (mean = 0.615 [SD = 0.025]), it did not have the highest holdout C-index. The most variable algorithms were Lasso (SD = 0.033) and Ridge (SD = 0.031) across cross-validation repeats. Except for Lasso and Elastic Net, there were no statistically significant differences in cross-validation C-indices between Cox and the ML methods (S4 Table in S1 File). Lasso and Elastic Net had significantly worse cross-validated performance than the other ML methods.

thumbnail
Fig 2. Violin plots comparing cross-validation C-indices for 7 different machine learning survival models against Cox PH in the A) post-policy cohort and B) pre-policy cohort.

Black bar indicates the mean c-index.

https://doi.org/10.1371/journal.pone.0313600.g002

In the pre-policy cohort, the algorithms performed similarly, with Cox Boost (holdout C-index = 0.633) performing the best and Cox (holdout C-index = 0.573) the worst (Table 2). Cross-validation C-indices ranged from 0.628 to 0.514, and the most variable algorithm was Lasso (SD = 0.044) (S4 Table in S1 File). There were also largely no statistically significant differences in cross-validation C-indices between Cox and the ML methods except for Lasso (S4 Table in S1 File). All of the pre-policy cohort-trained models applied to the post-policy holdout set performed more poorly compared to the pre-policy holdout set.

Prognostic variables for posttransplant one-year mortality

Sparse models such as Lasso, Elastic Net, and Cox Boost use a regularization term to shrink coefficients of variables to zero, effectively generating a subset of variables. Hereafter, we refer to this as a model’s “selection”, although we clarify that we did not perform computational feature selection before training the model. Since we performed 5 repeats of 5-fold cross-validation, a variable can be selected 25 times at most. We use this crude measure of selection to first determine which variables tend to be included in the majority of model replications [29], although we note that there exists inherent variability in sparse variable selection [30]. Age, total ischemic time, bilirubin, BMI, and ECMO were the top 5 most selected variables in the post-policy cohort (Fig 3). The top 5 in the pre-policy cohort were similar except for HIV status. VAD, GFR, prior transfusions, and Medicaid were other variables that were frequently selected in the post- and pre-policy cohorts. Comprehensive heatmap summaries of the total number of times each input predictor was selected in each model in the post- and pre-policy cohorts are provided in S1 and S2 Figs in S1 File respectively.

thumbnail
Fig 3. Average number of times variable is selected across all sparse machine learning models in the A) post-policy cohort and B) pre-policy cohort.

Sparse machine learning models include Lasso, Elastic Net, and Cox Boost.

https://doi.org/10.1371/journal.pone.0313600.g003

Plots of the top 20 most important features for the top-performing models are shown in Fig 4. The most important predictor variables are fairly constant between post-policy and pre-policy cohorts, although there are some differences. Some of the most important variables RSF selected in the post-policy cohort were bilirubin, age, pulmonary diastolic pressure, donor height, and wait time (Fig 4A). In the pre-policy cohort, bilirubin, donor age, eGFR measured at waitlist and pre-transplant, and age were selected (Fig 4B). Cox Boost selected HCV positive as the most important variable in the post-policy cohort and HIV negative as the most important variable in the pre-policy cohort (Fig 4C and 4D). XGBT selected bilirubin, eGFR, age, and BMI as the most important variables in both pre and post-policy cohorts (Fig 4E and 4F). Boxplots for all variables hazard ratios and importance scores for all models are shown in S3-9 Figs in S1 File.

thumbnail
Fig 4. Top 20 significant predictor variables.

A) RSF in post-policy cohort B) RSF in pre-policy cohort C) Cox Boost in post-policy cohort D) Cox Boost in pre-policy cohort E) XGBT in post-policy cohort F) XGBT in pre-policy cohort.

https://doi.org/10.1371/journal.pone.0313600.g004

Discussion

In this study, we benchmarked survival machine learning methods for the prognosis of one-year all-cause mortality in adult heart transplant recipients transplanted after the UNOS allocation policy change in October 2018 [16]. We also compared the models in a pre-policy era to assess whether the characteristics of predictor variables changed. In both the post- and pre-policy eras, the discriminatory power of machine learning models for one-year all-cause mortality tended to be higher compared to Cox, although not statistically different according to cross-validated C-indices. While the ML models were not statistically superior to Cox, these complex models likely have an advantage over Cox regression in two aspects: they do not require the proportional hazards assumption and they can capture nonlinear interactions between outcome and variables. Thus, these models can potentially provide a level of insight into the different predictive factors in these two eras that linear models cannot. It is also important to state that first, we only used information available pretransplant since that is where a post-transplant mortality prediction model is most likely useful, unlike other prediction models [31]. This may have limited the maximum possible performance of machine learning performance on mortality prediction. Second, based on our study design, we did not use the full range of data available, as other predictive models have done, so our model performance may be affected by lower statistical power. The pre-policy cohort-trained models had lower C-indices when applied to a post-policy holdout set compared to a pre-policy holdout set. This could be explained by underlying changes in treatment practices following the policy change in 2018 such as increased IABP usage and increased pre-transplant temporary mechanical circulatory support use [17,18,3234].

Although we begin with the caveat that the variables assessed by these survival machine learning models should not be taken as causal nor be interpreted too heavily, we briefly highlight two observations. The first is that ECMO was considered an important variable by RSF in the pre-policy cohort, but not in the post-policy cohort. Given the ongoing debate on whether ECMO is associated with worse post-transplant outcomes [32,3439], the absence of ECMO in the contemporary post-policy model may suggest that it is no longer a highly discriminative variable of one-year mortality relative to other variables. Whether this is due to improvements in ECMO management or increased relative usage of IABP remains a future direction of research. Additionally, HIV status was no longer considered an important variable in RSF and Cox Boost in the post-policy cohort. Although one study looked found no differences in transplantation outcomes for HIV-positive and -negative patients overall since 1987 [40], a question arises as to whether contemporary management of HIV patients has improved.

Study limitations

First, as with most national registries, SRTR is susceptible to data entry errors and missing data regarding organ recovery or failure necessary to create continuously updated mortality estimates. However, the SRTR conducts edit checks, validation of data at time of entry, and internal verification when there are outliers. Second, the data did not include known prognostic risk factors such as natriuretic peptides, serum sodium, and specific hemodynamic data such as heart rate, although these are usually available and collected at individual centers. In addition, SRTR does not include other variables such as patient adherence and granular clinical decision-making variables that likely affect survival. Notably, status justification form variables had a high proportion of missingness (above 90%), so they were not included in the predictor set. We would also like to mention that we did not have data that adequately capture social determinants of health relating to access to care or socioeconomic status. Third, this study is exploratory, and the intent was not to build a prediction model for individual prediction of mortality risk but rather to identify potentially important variables associated with mortality. Fourth, machine learning models are often more difficult to interpret than linear models such as Cox since many hyperparameters can be tuned and there is no easily interpretable measure of directional association. Fifth, we assumed that there was only missingness at random and non-informative censoring. However, there may be informative censoring, since early post-transplant deaths are more likely to be recorded than routine follow-up appointments in the first year after transplantation. Sixth, we did not have access to an external heart transplant dataset, which is needed to generalize our findings to populations outside of the United States. Future studies can test survival machine learning methods on prospectively collected SRTR data or more carefully assess how associations of individual predictive variables have changed following the UNOS 2018 policy change.

Conclusions

In this benchmark prognostic study, we demonstrated that machine learning models demonstrate reasonable one-year mortality prediction and can help reveal complex relationships between predictor variables and mortality. We show differences in important variables such as ECMO and HIV status between the post-policy and pre-policy eras, which showcases the ability of machine learning models to generate future hypotheses for research and suggest the need for continual updating of existing models. As clinical care practices continue to evolve, exploring computational time-to-event algorithms to develop mortality risk prediction models with more accumulated data is warranted.

Acknowledgments

We would like to thank Annette Spooner for sharing her machine learning benchmark code with us and for helping with some of the technical components of the analysis.

References

  1. 1. Khush KK, Cherikh WS, Chambers DC, Harhay MO, Hayes D, Hsich E, et al. The International Thoracic Organ Transplant Registry of the International Society for Heart and Lung Transplantation: Thirty-sixth adult heart transplantation report—2019; focus theme: Donor and recipient size match. J Heart Lung Transplant. 2019 Oct 1;38(10):1056–66. pmid:31548031
  2. 2. Dani A, Heidel JS, Qiu T, Zhang Y, Ni Y, Hossain MM, et al. External validation and comparison of risk score models in pediatric heart transplants. Pediatr Transplant. 2021 Dec 8;e14204. pmid:34881481
  3. 3. Weiss ES, Allen JG, Arnaoutakis GJ, George TJ, Russell SD, Shah AS, et al. Creation of a quantitative recipient risk index for mortality prediction after cardiac transplantation (IMPACT). Ann Thorac Surg. 2011 Sep;92(3):914–21; discussion 921–922. pmid:21871277
  4. 4. Aleksova N, Alba AC, Molinero VM, Connolly K, Orchanian-Cheff A, Badiwala M, et al. Risk prediction models for survival after heart transplantation: A systematic review. Am J Transplant Off J Am Soc Transplant Am Soc Transpl Surg. 2020 Apr;20(4):1137–51. pmid:31733026
  5. 5. Weiss ES, Allen JG, Kilic A, Russell SD, Baumgartner WA, Conte JV, et al. Development of a quantitative donor risk index to predict short-term mortality in orthotopic heart transplantation. J Heart Lung Transplant Off Publ Int Soc Heart Transplant. 2012 Mar;31(3):266–73. pmid:22093382
  6. 6. Hsich EM, Blackstone EH, Thuita LW, McNamara DM, Rogers JG, Yancy CW, et al. Heart Transplantation: An In-Depth Survival Analysis. JACC Heart Fail. 2020 Jul;8(7):557–68. pmid:32535125
  7. 7. Zhou Y, Chen S, Rao Z, Yang D, Liu X, Dong N, et al. Prediction of 1-year mortality after heart transplantation using machine learning approaches: A single-center study from China. Int J Cardiol. 2021 Sep 15;339:21–7. pmid:34271025
  8. 8. Linse B, Ohlsson M, Stehlik J, Lund LH, Andersson B, Nilsson J. A machine learning model for prediction of 30-day primary graft failure after heart transplantation. Heliyon. 2023 Mar;9(3):e14282. pmid:36938431
  9. 9. Kampaktsis PN, Tzani A, Doulamis IP, Moustakidis S, Drosou A, Diakos N, et al. State-of-the-art machine learning algorithms for the prediction of outcomes after contemporary heart transplantation: Results from the UNOS database. Clin Transplant. 2021 Aug;35(8):e14388. pmid:34155697
  10. 10. Killian MO, Payrovnaziri SN, Gupta D, Desai D, He Z. Machine learning-based prediction of health outcomes in pediatric organ transplantation recipients. JAMIA Open. 2021 Jan;4(1):ooab008. pmid:34075353
  11. 11. Naruka V, Arjomandi Rad A, Subbiah Ponniah H, Francis J, Vardanyan R, Tasoudis P, et al. Machine learning and artificial intelligence in cardiac transplantation: A systematic review. Artif Organs. 2022 Sep;46(9):1741–53. pmid:35719121
  12. 12. Ashfaq A, Gray GM, Carapelluci J, Amankwah EK, Rehman M, Puchalski M, et al. Survival analysis for pediatric heart transplant patients using a novel machine learning algorithm—A UNOS Analysis. J Heart Lung Transplant Off Publ Int Soc Heart Transplant. 2023 Jun 14;S1053-2498(23)01899-5. pmid:37327979
  13. 13. Nilsson J, Ohlsson M, Höglund P, Ekmehag B, Koul B, Andersson B. The International Heart Transplant Survival Algorithm (IHTSA): a new model to improve organ sharing and survival. PloS One. 2015;10(3):e0118644.
  14. 14. Ayers B, Sandholm T, Gosev I, Prasad S, Kilic A. Using machine learning to improve survival prediction after heart transplantation. J Card Surg. 2021 Nov;36(11):4113–20. pmid:34414609
  15. 15. George B, Seals S, Aban I. Survival analysis and regression models. J Nucl Cardiol Off Publ Am Soc Nucl Cardiol. 2014 Aug;21(4):686–94. pmid:24810431
  16. 16. Adult heart allocation—OPTN [Internet]. [cited 2022 Jul 29]. https://optn.transplant.hrsa.gov/professionals/by-organ/heart-lung/adult-heart-allocation/
  17. 17. Lazenby KA, Narang N, Pelzer KM, Ran G, Parker WF. An updated estimate of posttransplant survival after implementation of the new donor heart allocation policy. Am J Transplant Off J Am Soc Transplant Am Soc Transpl Surg. 2022 Jun;22(6):1683–90. pmid:34951528
  18. 18. Varshney AS, Berg DD, Katz JN, Baird-Zars VM, Bohula EA, Carnicelli AP, et al. Use of Temporary Mechanical Circulatory Support for Management of Cardiogenic Shock Before and After the United Network for Organ Sharing Donor Heart Allocation System Changes. JAMA Cardiol. 2020 Jun 1;5(6):703–8. pmid:32293644
  19. 19. Zhang Z. Multiple imputation with multivariate imputation by chained equation (MICE) package. Ann Transl Med. 2016 Jan;4(2):30. pmid:26889483
  20. 20. Kamalia MA, Ramamurthi A, Rein L, Mohammed A, Joyce DL. Detection of Seasonal Trends in National Donor Heart Availability Using the UNOS Dataset. J Card Fail. 2019 Aug 1;25(8, Supplement):S174.
  21. 21. Tibshirani R. The lasso method for variable selection in the Cox model. Stat Med. 1997 Feb 28;16(4):385–95. pmid:9044528
  22. 22. Simon N, Friedman J, Hastie T, Tibshirani R. Regularization Paths for Cox’s Proportional Hazards Model via Coordinate Descent. J Stat Softw. 2011 Mar;39(5):1–13. pmid:27065756
  23. 23. Friedman JH. Stochastic gradient boosting. Comput Stat Data Anal. 2002 Feb 28;38(4):367–78.
  24. 24. Chen T, Guestrin C. XGBoost: A Scalable Tree Boosting System. In: Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining [Internet]. New York, NY, USA: Association for Computing Machinery; 2016 [cited 2022 Jul 29]. p. 785–94. (KDD ‘16). https://doi.org/10.1145/2939672.2939785
  25. 25. Barnwal A, Cho H, Hocking T. Survival regression with accelerated failure time model in XGBoost. J Comput Graph Stat. 2022;31(4):1292–302.
  26. 26. Ishwaran Hemant, Kogalur Udaya B., Blackstone Eugene H., Lauer Michael S. Random survival forests. Ann Appl Stat. 2008 Sep 1;2(3):841–60.
  27. 27. Bischl B, Lang M, Kotthoff L, Schiffner J, Richter J, Studerus E, et al. mlr: Machine Learning in R. J Mach Learn Res. 2016;17(170):1–5.
  28. 28. Nadeau C, Bengio Y. Inference for the Generalization Error. Mach Learn. 2003 Sep 1;52(3):239–81.
  29. 29. Spooner A, Chen E, Sowmya A, Sachdev P, Kochan NA, Trollor J, et al. A comparison of machine learning methods for survival analysis of high-dimensional clinical data for dementia prediction. Sci Rep. 2020 Nov 23;10(1):20410. pmid:33230128
  30. 30. Strobl C, Boulesteix AL, Zeileis A, Hothorn T. Bias in random forest variable importance measures: Illustrations, sources and a solution. BMC Bioinformatics. 2007 Jan 25;8(1):25. pmid:17254353
  31. 31. Tian D, Yan HJ, Huang H, Zuo YJ, Liu MZ, Zhao J, et al. Machine Learning–Based Prognostic Model for Patients After Lung Transplantation. JAMA Netw Open. 2023 May 5;6(5):e2312022. pmid:37145595
  32. 32. Kilic A, Mathier MA, Hickey GW, Sultan I, Morell VO, Mulukutla SR, et al. Evolving Trends in Adult Heart Transplant With the 2018 Heart Allocation Policy Change. JAMA Cardiol. 2021 Feb 1;6(2):159–67. pmid:33112391
  33. 33. Parker WF, Chung K, Anderson AS, Siegler M, Huang ES, Churpek MM. Practice Changes at U.S. Transplant Centers After the New Adult Heart Allocation Policy. J Am Coll Cardiol. 2020 Jun 16;75(23):2906–16.
  34. 34. Trivedi JR, Slaughter MS. “Unintended” Consequences of Changes in Heart Transplant Allocation Policy: Impact on Practice Patterns. ASAIO J Am Soc Artif Intern Organs 1992. 2020 Feb;66(2):125–7. pmid:31977354
  35. 35. Jani M, Lee S, Hoeksema S, Acharya D, Boeve T, Manandhar-Shrestha N, et al. Changes in Wait List Mortality, Transplantation Rates and Early Post-Transplant Outcomes in LVAD BTT with New Heart Transplant Allocation Score. A UNOS Database Analysis. J Heart Lung Transplant. 2021 Apr 1;40(4):S17.
  36. 36. Xia Y, Kim JS, Eng IK, Nsair A, Ardehali A, Shemin RJ, et al. Outcomes of heart transplant recipients bridged with percutaneous versus durable left ventricular assist devices. Clin Transplant. 2023 Apr;37(4):e14904. pmid:36594638
  37. 37. Bradbrook K, Goff RR, Lindblad K, Daly RC, Hall S. A national assessment of one-year heart outcomes after the 2018 adult heart allocation change. J Heart Lung Transplant. 2023 Feb 1;42(2):196–205. pmid:36184382
  38. 38. Cogswell R, John R, Estep JD, Duval S, Tedford RJ, Pagani FD, et al. An early investigation of outcomes with the new 2018 donor heart allocation system in the United States. J Heart Lung Transplant Off Publ Int Soc Heart Transplant. 2020 Jan;39(1):1–4.
  39. 39. Kim ST, Xia Y, Tran Z, Hadaya J, Dobaria V, Choi CW, et al. Outcomes of extracorporeal membrane oxygenation following the 2018 adult heart allocation policy. PLoS ONE. 2022 May 20;17(5):e0268771. pmid:35594315
  40. 40. Doberne JW, Jawitz OK, Raman V, Bryner BS, Schroder JN, Milano CA. Heart Transplantation Survival Outcomes of HIV Positive and Negative Recipients. Ann Thorac Surg. 2021 May;111(5):1465–71. pmid:32946847