Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

In-hospital real-time prediction of COVID-19 severity regardless of disease phase using electronic health records

  • Hyungjun Park,

    Roles Conceptualization, Formal analysis, Investigation, Methodology, Software, Validation, Visualization, Writing – original draft, Writing – review & editing

    Affiliation Division of pulmonology and Critical Care Medicine, Department of Internal Medicine, Gumdan top hospital, Incheon, South Korea

  • Chang-Min Choi,

    Roles Conceptualization, Validation

    Affiliations Division of Pulmonology and Critical Care Medicine, Department of Internal Medicine, Asan Medical Center, University of Ulsan College of Medicine, Seoul, South Korea, Division of Oncology, Department of Internal Medicine, University of Ulsan College of Medicine, Asan Medical Center, Seoul, South Korea

  • Sung-Hoon Kim,

    Roles Conceptualization, Funding acquisition, Supervision

    Affiliation Department of Anesthesiology and Pain Medicine, Asan Medical Center, Ulsan College of Medicine, Seoul, South Korea

  • Su Hwan Kim,

    Roles Conceptualization, Data curation, Supervision

    Affiliations Department of Internal Medicine, Seoul National University College of Medicine, Seoul, South Korea, Division of Gastroenterology, Department of Internal Medicine, Seoul Metropolitan Government Seoul National University Boramae Medical Center, Seoul, South Korea

  • Deog Kyoem Kim,

    Roles Conceptualization, Data curation, Supervision

    Affiliations Department of Internal Medicine, Seoul National University College of Medicine, Seoul, South Korea, Division of Pulmonary and Critical Care Medicine, Department of Internal Medicine, Seoul Metropolitan Government Seoul National University Boramae Medical Center, Seoul, South Korea

  • Ji Bong Jeong

    Roles Conceptualization, Data curation, Writing – review & editing

    jibjeong@snu.ac.kr

    Affiliations Department of Internal Medicine, Seoul National University College of Medicine, Seoul, South Korea, Division of Gastroenterology, Department of Internal Medicine, Seoul Metropolitan Government Seoul National University Boramae Medical Center, Seoul, South Korea

Abstract

Coronavirus disease 2019 (COVID-19) has strained healthcare systems worldwide. Predicting COVID-19 severity could optimize resource allocation, like oxygen devices and intensive care. If machine learning model could forecast the severity of COVID-19 patients, hospital resource allocation would be more comfortable. This study evaluated machine learning models using electronic records from 3,996 COVID-19 patients to forecast mild, moderate, or severe disease up to 2 days in advance. A deep neural network (DNN) model achieved 91.8% accuracy, 0.96 AUROC, and 0.90 AUPRC for 2-day predictions, regardless of disease phase. Tree-based models like random forest achieved slightly better metrics (random forest: 94.1% of accuracy, 0.98 AUROC, 0.95 AUPRC; Gradient boost: 94.1% of accuracy, 0.98 AUROC, 0.94 AUPRC), prioritizing treatment factors like steroid use. However, the DNN relied more on fixed patient factors like demographics and symptoms in aspect to SHAP value importance. Since treatment patterns vary between hospitals, the DNN may be more generalizable than tree-based models (random forest, gradient boost model). The results demonstrate accurate short-term forecasting of COVID-19 severity using routine clinical data. DNN models may balance predictive performance and generalizability better than other methods. Severity predictions by machine learning model could facilitate resource planning, like ICU arrangement and oxygen devices.

Introduction

The COVID-19 infection swiftly propagated on a global scale, leading the World Health Organization (WHO) to officially declare it as a pandemic [1]. Although more than half of all infections are subclinical in young ages, 43% of older than 65 get hospitalized and 76% were died until September 2021 [2]. To control the spread of COVID-19 infection, pharmaceutical (vaccination) [3] and non-pharmaceutical interventions such as using face masks, isolation, quarantining, and social distancing have been exercised [4, 5]. Although the difference of stringency and period of covid-19 restriction, most of the countries conducted severe restriction more than 10 months [4]. At the peak of the pandemic, these interventions were necessary to reduce shortages in hospital care resources, particularly among critical care resources such as intensive care units (ICUs) and oxygen supplies [6, 7]. As a result of the number of COVID-19 patients exceeding the hospital’s capacity, some patients received inadequate care, resulting in poorer results than expected for the natural course of COVID-19 [6]. Moreover, severe COVID-19 patients require high-flow oxygen devices and mechanical ventilation, and preparation of those facilities was warranted [8]. Therefore, the Korean government replaced public hospitals with infectious disease-dedicated hospitals, monitored the daily number of severe COVID-19 patients, and determined whether this number would exceed the hospital capacity [9].

Prediction of future need for intensive care is required for controlling capacities for facilities in medical centers. Timely preparation of essential medical devices based on short-term forecasts (1 day or 2 days) is critical for ensuring efficient management of healthcare resources. By predicting short-term patient outcomes, informed decisions can be made regarding the potential transfer of patients to alternate hospitals or care units within the same facility’s ICU. With the development of artificial intelligence, several prediction models were established to predict the short-term outcomes of COVID-19 [1012]. In the early phase of the pandemic, many models were developed using static predictions—these predicted future events at admission [1012], but did not indicate when the event would occur [13]. In contrast to static predictions, dynamic predictions or real-time predictions estimate a temporal event, such as the daily risk of acute kidney injury [14] or need for ICU care [15]. Dynamic prediction in the medical context pertains to the daily forecasting of impending outcomes, such as complications or mortality. Concerning COVID-19, the near-future possible outcome include patient mortality [11], the requirement for mechanical ventilation [16], as well as discharge or mortality [17]. The predictive capacity for complications or mortality empowers healthcare workers to concentrate on addressing correctable issues in patients and to adequately prepare for future intensive care unit (ICU) demands related to mechanical ventilation. Given the restrictions on available resources, forecasting not only adverse outcomes but also predicting the potential recovery of patients from mechanical ventilation or oxygen support assumes importance in hospital resource management. Consequently, making accurate predictions about patients’ potential recovery from mechanical ventilation and their subsequent discharge from the ICU enables informed decision-making concerning the transfer of other patients in need of mechanical ventilation. However, previous studies focused only on the start date of the event and not on recovery.

We examined various machine learning and deep learning models to predict the daily severity of COVID-19 at a hospital dedicated to COVID-19 patients. In addition, we assessed the importance of input variables in each model and considered the possible hazard of bias associated with models that displayed exceptional performance.

Material and methods

Clinical dataset description

Deidentified patients with a confirmed diagnosis of COVID-19 from January 1, 2020, to October 31, 2021 at Boramae Medical Center (BMC) were consecutively included. The authors obtained access to this data set from February 1, 2022, following approval from the Institutional Review Board (IRB). BMC was dedicated to infectious diseases and only admitted COVID-19 patients. During the pandemic, BMC used telemonitoring to monitor patients with COVID-19 in residential treatment centers. In case of symptom progression at a residential treatment center, the patient was referred to BMC. All patients identified with COVID-19 were diagnosed by nasopharyngeal swab reverse transcription-polymerase chain reaction. Patients younger than 18 years of age were excluded from this study. This study was approved by the ethics committee of the BMC and was conducted as per the Declaration of Helsinki. The requirement for informed consent was waived by the ethics committee of the BMC (approval number 10-2021-130), considering the retrospective nature of the study.

COVID-19 severity as the target outcome

The disease severity defined by World Health Organization (WHO) clinical progression scales, was recorded daily at BMC at midnight [18]. The primary objective of our study was to forecast the severity of the condition prior to the event, focusing on predictions for day 0, 1, and 2. It was deemed sufficient to forecast the patient’s severity up to 2 days in advance to adequately prepare the necessary oxygen and ICU resources. The WHO clinical progression scale ranges from 0 (not infected) to 10 (dead) [18]. In our model, the score was divided into 3 scores, i.e., mild cases (monitored, no oxygen therapy; WHO scale < 5), moderate disease (hospitalized, need oxygen by mask or nasal prong; WHO scale = 5), and severe disease (oxygen by high flow, non-invasive ventilation, or mechanical ventilation; WHO scale > 5). Utilization of the oxygen device was meticulously documented with corresponding timestamps, thereby ensuring that all daily fluctuations in severity were accurately captured on an daily basis.

Data preprocessing and feature selection

From the electronic medical record system at BMC, we extracted the following data: demographic features (age, sex, underlying disease, smoking status), daily symptom records, laboratory data, drug use, vital signs. (S1 and S2 Tables) The laboratory data encompassed a range of parameters, including complete blood cell count, chemistry markers (such as protein, albumin, liver function tests, BUN/Creatinine), electrolyte levels, inflammatory markers (CRP, LDH, ferritin, procalcitonin), and coagulation measurements. The selection of these specific variables was determined by the attending clinicians responsible for patient care, who deemed them relevant for the management and treatment of COVID-19 patients. All these data served as input variables utilized in the prediction of daily severity fluctuations.

The binary and categorical features such as the presence of symptoms and type of oxygen device used were categorized as 0 or 1 by one-hot encoding. Although information on the type of oxygen device and fraction of oxygen was available, it was not included in the model to prevent data leakage because it directly represented the outcome. Antiviral agent use was treated as a binary variable and steroid use, which was represented as daily summed doses, was treated as a continuous variable. For continuous variables, data cleansing was conducted by removing error values (such as negative blood pressure values) and capping outlier values. Data clipping was usually conducted in the 1st and 99th percentile values [15]; however, we used outlier values decided by the clinicians as the data clipping points because several 1st and 99th percentile values did not represent severity of the laboratory values. Each data point is described in S1 Table. Among the extracted features, rarely checked laboratory examinations or vital signs, such as HbA1c and neurologic examinations, were excluded from our input dataset. Patients’ symptoms were recorded daily and categorized into 18 types. Positive symptoms during admission are described in S2 Table. Every day, all daily symptom records were scrutinized and marked either as present or absent. If a patient was intubated, the corresponding symptom was deemed to be absent.

Data exclusion and imputation

Data integrity is important for precise prediction [19]. Laboratory test were less conducted than vital sign data because patients with mild COVID-19 infection did not frequently undergo laboratory tests. The admission events were excluded if >50% of their vital signs data, >80% of their laboratory data, or their COVID-19 severity outcome data were missing during the entire admission period (S1 Fig).

Missing value imputation was conducted in two steps. First, if the variable had an existing previous value, the last observation carried forward method was used (such as laboratory data and vital signs). Second, when the patient was hospitalized without laboratory results, we assumed that the prediction based on normal values until abnormal data were observed would be robust. As many data distributions were skewed, the mode was appropriate for representing normal values in our dataset. (S2 and S3 Figs) Thus, no previously observed data were imputed with each mode value.

Models

We evaluated several models for predicting the daily severity of COVID-19, classifying each model as a non-temporal or temporal model. The non-temporal model that received daily data consisted of logistic regression, random forest [20], gradient boost (XG boost) [21], and deep neural networks (DNN) [22]. The detailed hyperparameters of the models were described in Supplement methods. On the day of prediction, the model receives information regarding vital signs in the last 8 hours (16–24 hour) and laboratory and symptom data along with daily information. As temporal model, Transformer received trend information for prediction [23]. The model receives input as trend data, thus, the vital sign trend of n-day and lab data of n-day. The vital sign and laboratory data had different frequencies; the index time window of the vital sign was defined as 8 hours, and that of the lab was 24 hours. The detailed input pipeline is explained in Supplemental methods. (S1 File) We evaluated the model using different input lengths by changing it every two hospital days to determine its effect on the model.

The source code used for developing our model and for drawing the figure was implemented by Python (version 3.8), numpy (version 1.22), scikit-learn (version 1.0.2), pandas (version 1.4.0), Pytorch (version 1.10), and matplotlib (version 3.5.1).

Statistical analysis

The baseline characteristics were described as mean and standard deviation for continuous variables and number and percentage for binary variables. The difference between continuous variables was calculated by ANOVA; binary variables were calculated using the chi-square test. Model performance was evaluated by area under the receiver operating curve (AUROC) and area under the precision-recall curve (AUPRC), which involved multi-class classification. AUROC and AUPRC were calculated by binarization to the target and the rest of the target. Each target severity was plotted by each receiver operating characteristic curve and precision-recall curve. For a more comprehensive comparison of the models, we implemented a decision curve analysis [24]. Given that this methodology was originally conceptualized for binary classification, it necessitated the modification of the data format to a "one versus the rest" approach for our predictions classified as mild, moderate, or severe. The patients were divided into training (60%), validation (20%), and test (20%) sets by stratified splitting, and the model was trained based on the validation set using early stopping. To calculate the mean and standard deviation of performance metrics according to the input length, we repeated splitting and training five times per input length.

The model importance was computed using Shapley additive explanation (SHAP) values, which is model-independent. Using this strategy, each of the random forest, gradient boost, and deep neural network models assessed the importance of the variable. Input variable importance was assessed based on the data type, distinguishing between hospital/treatment-dependent factors (extrinsic factors) and patient-dependent factors (intrinsic factors). Hospital/treatment-dependent factors encompassed the number of vital sign occurrences (count, variance) and the dosage of drugs (max, min, median). On the other hand, patient-dependent factors included vital sign values (max, mean, median, min), demographics (operation history, previous admission, age, sex), and the presence of symptoms.

Results

Among the included patients, 4,660 patients were identified and 3,996 patients remained following exclusion criteria regarding age and data integrity. The baseline characteristics were classified by the maximum COVID-19 severity during admission. (Table 1) Older age, presence of underlying disease, previous operation, and previous admission were associated with severe COVID-19 outcomes. The median duration of hospital admission was 9 days (interquartile range: 7–12 days; S4 Fig). A good example of the model prediction is provided in Fig 1. The model can predict the aggravation and recovery of COVID-19 1 and 2 days in advance.

thumbnail
Fig 1. Example of prediction of daily severity during admission.

The target is actual daily severity, and day 0 to day 2 represent the model predictions. The patient’s status aggravated to severe on day 2, recovered to moderate on day 9, and became mild on day 13. The model (day 2) predicts disease aggravation and recovery 2 days before the infection.

https://doi.org/10.1371/journal.pone.0294362.g001

Model performances

The model predicted the disease severity per day, including the present day (day 0), day 1, and day 2. The random forest model had the best performance, followed by the XG boost model, the deep neural network model, and the logistic regression model. The model’s prediction performance peaked at day 0 with an AUROC of 0.981 and an AUPRC of 0.940 (DNN), and it decreased as the prediction horizon expanded (Table 2, S3 Table).

thumbnail
Table 2. The best model performance for daily COVID-19 severity.

https://doi.org/10.1371/journal.pone.0294362.t002

The accuracy of prediction decreased with increasing time horizon (accuracy: 92.6% [day 1], 91.8% [day 2], DNN). The transformer, which is more complex than DNN, did not show better performance. (Table 3) Regarding severity, the DNN model showed AUROC values of 0.958, 0.928, and 0.987 for mild, moderate, and severe COVID-19, respectively. In AUPRC, the model showed prediction performance of 0.985, 0.823, and 0.947 for mild, moderate, and severe COVID-19 after 2 days (Fig 2).

thumbnail
Fig 2. The AUROC and AUPRC for daily prediction of COVID-19 severity.

The upper figures denote the receiver operating curve for predicting day 0, day 1, and day 2. The lower figures denote the precision-recall curve for each severity outcome.

https://doi.org/10.1371/journal.pone.0294362.g002

thumbnail
Table 3. Comparison of the model performances between DNN and transformer.

https://doi.org/10.1371/journal.pone.0294362.t003

In addition, our decision curve analysis highlighted the net-benefit of the model based on varying threshold probabilities. Consistent with our previous AUROC comparisons, the random forest and XG boost models demonstrated higher net-benefits for the mild, moderate, and severe predictions for day 2, as displayed in Fig 3.

thumbnail
Fig 3. The comparison of the model by decision curve analysis.

The figure represents the decision curve analysis for severity predictions for the period two days ahead. The baseline comparison (All positive) is premised on the assumption that every two-day event will be mild, moderate, or severe. All of our models demonstrated benefits exceeding the basic assumption made without any model (All positive).

https://doi.org/10.1371/journal.pone.0294362.g003

Feature importance of different models

The SHAP method was used to demonstrate the models’ feature importance. (Fig 4) We selected the highest feature importance up to the 20th variable for predicting severe COVID-19 after 2 days. Although feature importance was determined in the same subset of the test set, the selected features varied among models. DNN showed higher feature importance in the following order: operation history, male sex, older age, hypertension, and nausea. The XG boost model used the following ranking: steroid (maximum, mean), lactate dehydrogenase (minimum), old age, and SpO2 (minimum, count). Additionally, the random forest model used the following: steroid (maximum), SpO2 (count), Pulse rate (count), steroid (minimum), and SpO2 (minimum).

thumbnail
Fig 4. The feature importance among the models.

The three model’s feature importance were compared by the SHAP method. The tree-based model places a greater focus on steroid usage, SpO2 count, and vital signs (Pulse rates, and diastolic BP). The DNN model placed a greater focus on demographic data and symptoms.

https://doi.org/10.1371/journal.pone.0294362.g004

The factors were categorized as patient-dependent and hospital/treatment-dependent. For predicting patient’s severity, the patient factor cannot be altered by the hospital where the patient was admitted. However, hospital/treatment factors, such as the number of vital signs that should be monitored and the dosage of steroid administered, can differ between hospitals. Regarding the SHAP variable analysis, the patient-dependent percentage for each model was 95% in DNN, 75% in XGBoost, and 35% in random forest (p-value: 0.065). Thus, the DNN prioritized patient factors, whereas the random forest and XG boost prioritized hospital/treatment factors. The same phenomenon was also observed for predicting mild and moderate COVID-19 infection (S5 Fig).

Discussion

We examined a range of models to forecast daily COVID-19 severity. The DNN model had an accuracy, AUROC, and AUPRC of 91.8%, 0.96, and 0.90, respectively for predicting the severity of COVID-19 as either mild, moderate, or severe 2 days in advance. The transformer model that received trend data did not perform better than the DNN model that received only daily data. The tree model (random forest, XG boost) outperformed DNN; however, the model’s feature importance was heavily weighted towards hospital/treatment factors, resulting in unclear generalizability in other hospitals.

Severe COVID-19 shows a similar disease course as that observed in acute respiratory distress syndrome, which requires high-level oxygen support from high-flow nasal oxygen, ventilator support, and ICU care [25]. Respiratory devices are restricted resources; thus, treatment of severe COVID-19 patients is limited by device and healthcare worker availability [6]. Our model can help identify patients that show aggravation of the infection and recovery in near future; thus, clinicians could be advised to start or wean respiratory devices in a limited resource setting. Previous studies focused on predicting disease aggravation at admission along with real-time prediction [16, 26]. While it is essential to identify potential patients with infection aggravation in future, identifying those who will recover is also important in managing overall hospital resources. By integrating our model into clinical practice, clinicians can readily monitor patients who may be at risk of deterioration due to COVID-19 as well as those showing signs of recovery. Such insights prove valuable in anticipating potential strains on medical facilities and allow for proactive arrangements, including network preparations for transferring patients requiring ventilator support and intensive care unit (ICU) facilities.

A static prediction implies prediction of any future event at the onset of specific time [15]. For example, a logistic model incorporating clinical symptoms and abnormal laboratory tests at admission could predict whether patients would develop a serious illness within 30 days [10]. Dynamic or real-time prediction predicts outcomes at each time-point [15]. Recently, a dynamic prediction model for forecasting future COVID-19-related aggravation [16] or mortality [11] was developed. Although temporal model (RNN) improved the performance of numerous tasks [13, 27, 28], daily severity prediction tasks did not benefit from the use of temporal model in this study. We anticipated improved performance by incorporating temporal information into the transformer model, but the improvement was task-dependent. In this study, the AUROC and AUPRC for predicting severity 2 days in advance were >0.9, which could indicate that the task was rather simple and that the trend information provides no additional information compared to daily information.

Among non-temporal models, tree-based models, such as random forest and XG boost, performed better than DNN and logistic regression models. When logistic regression demonstrates performance equivalent to that of machine learning models, including deep learning model, a basic model may be preferred over a machine learning model because it is more interpretable and understandable [29]. In this study, as the dimension of the input data exceeded 300, machine learning enhanced model performance, suggesting that complex computation is more advantageous for prediction tasks than linear computation [30]. Regarding model performance, the random forest model should be preferred because it exhibited the best results among non-temporal models. However, the SHAP method evaluated model reliability based on the significance of features rather than model performance. The tree model revealed that drug use and frequency of vital sign checks are of higher importance, and these factors are dependent on treatment pattern and hospital rules. Moreover, in random forest and XG boost models, steroid usage was the most important factor. Steroid usage was considered a surrogate marker for severe COVID-19 infection, even if it is intended to treat the disease. Therefore, performance may be diminished if the treatment pattern or hospital rule for vital sign changes. This risk was identified in a prior study that employed the ensemble model based on two tree models; higher internal validation performance was not maintained in other hospitals [31]. This is because the ensemble model was hospital- and treatment-dependent, and the prediction was not transferable to other hospitals. However, in this study, the deep learning model placed greater emphasis on patient demographics and symptoms. As the feature cannot be altered even if the patient is transferred to a different hospital, the model’s predictions are transferable to other hospital settings. Thus, we determined that the DNN model is superior than the tree model due to greater generalizability despite its inferior performance.

The strength of this study is that our deep learning model predicts COVID-19 severity regardless of the phase of the disease. To the best of our knowledge, this is the first study to predict future severity using a time-series dataset, regardless of the course of COVID-19. We assessed a variety of models, including machine learning and deep learning models, as well as data structures derived from non-temporal and temporal input. The model’s interpretation was evaluated using the SHAP method, which considers each model’s important features and their generalizability using a domain-wise knowledge approach. Furthermore, our model’s predictive capability is particularly beneficial during nighttime when timely interventions are pivotal. By accurately estimating the potential risk of patient deterioration, clinicians are better equipped to allocate essential medical devices, ensuring that patients receive the necessary support even during off-peak hours

This study had some limitations. First, it was based on single-centered retrospective cohort data; thus, the model may not show high performance using other hospital datasets. Due to the difficulty of accessing other hospitals with a comparable patient group, external validation was restricted. The variability in the frequency of vital sign and laboratory tests may result in a higher incidence of missing data. Additionally, the unavailability of daily symptom records in other hospitals could potentially present challenges to the generalizability of our model. Nevertheless, we evaluated the model’s essential components using the SHAP method and its generalizability with variable characteristics. We followed to a common technique for processing time-series data; thus, the problem regarding hospital and treatment-related factors would be similar to that of another task prediction. Consequently, our method for determining the importance of a variable will benefit from a future study focusing on generalizability. Further validation is required to support our findings. Second, the severity of COVID-19 was determined by WHO guidelines and the application of oxygen. The application of oxygen could depend on the clinician’s preference. If the clinical pattern of oxygen usage in another hospital differed from that in ours, our model would not accurately predict the application and weaning of oxygen at that hospital. If severity could be independently established as per the physician’s preference, the target severity would be appropriate for any future research. Third, this study has not yet undergone prospective validation, hence, it does not provide guidance on what steps should be taken in instances where there is a discrepancy between clinical judgments and model predictions. A subsequent prospective study is needed to collect actual clinical experiences and reassess how to proceed in such scenarios.

Conclusions

Our hierarchical transformer model showed higher predictive performance in the early and late periods of hospital admission. Our model could be useful for COVID-19 patients by predicting their future outcomes and aids the distribution of respiratory devices.

Supporting information

S1 Fig. Missing value proportion per patient.

https://doi.org/10.1371/journal.pone.0294362.s002

(TIF)

S2 Fig. Exploration of laboratory distribution.

https://doi.org/10.1371/journal.pone.0294362.s003

(TIF)

S3 Fig. Difference of mode, median, and mean in ferritin value.

https://doi.org/10.1371/journal.pone.0294362.s004

(TIF)

S4 Fig. Histogram of the length of hospital admission duration.

https://doi.org/10.1371/journal.pone.0294362.s005

(TIF)

S5 Fig. The feature importance of the models for predicting mild and moderate COVID-19.

https://doi.org/10.1371/journal.pone.0294362.s006

(TIF)

S1 Table. Percentile and clipping values of continuous variables.

https://doi.org/10.1371/journal.pone.0294362.s007

(DOCX)

S2 Table. Symptom presentation of COVID-19 patients during hospital admission.

https://doi.org/10.1371/journal.pone.0294362.s008

(DOCX)

S3 Table. Detailed model performance comparison.

https://doi.org/10.1371/journal.pone.0294362.s009

(DOCX)

References

  1. 1. Nka AD, Ka’e AC, Bouba Y, Semengue ENJ, Tchouaket MCT, Takou D, et al. Global burden of SARS-CoV-2 infection, hospitalization and case fatality rate among COVID-19 vaccinated individuals and its associated factors: A systematic review and meta-analysis protocol. PLoS One. 2022;17: 1–7. pmid:35943981
  2. 2. Centers for Disease Control and Prevention (CDC). Cases, Data, and Surveillance. [cited 25 May 2023]. Available: https://www.cdc.gov/coronavirus/2019-ncov/cases-updates/burden.html
  3. 3. Ssentongo P, Ssentongo AE, Voleti N, Groff D, Sun A, Ba DM, et al. SARS-CoV-2 vaccine effectiveness against infection, symptomatic and severe COVID-19: a systematic review and meta-analysis. BMC Infect Dis. 2022;22: 1–12. pmid:35525973
  4. 4. Fink G, Tediosi F, Felder S. Burden of Covid-19 restrictions: National, regional and global estimates. eClinicalMedicine. 2022;45. pmid:35224471
  5. 5. Bo Y, Guo C, Lin C, Zeng Y, Li HB, Zhang Y, et al. Effectiveness of non-pharmaceutical interventions on COVID-19 transmission in 190 countries from 23 January to 13 April 2020. Int J Infect Dis. 2021;102: 247–253. pmid:33129965
  6. 6. Zhuang Z, Cao P, Zhao S, Han L, He D, Yang L. The shortage of hospital beds for COVID-19 and non-COVID-19 patients during the lockdown of Wuhan, China. Ann Transl Med. 2021;9: 200–200. pmid:33708827
  7. 7. Overwhelmed Hospitals Are Shipping COVID-19 Patients To Far-Off Cities: NPR. [cited 1 Sep 2022]. Available: https://www.npr.org/2021/08/19/1029378744/hospital-beds-shortage-covid-coronavirus-states
  8. 8. Alhazzani W, Evans L, Alshamsi F, Møller MH, Ostermann M, Prescott HC, et al. Surviving Sepsis Campaign Guidelines on the Management of Adults with Coronavirus Disease 2019 (COVID-19) in the ICU: First Update. Crit Care Med. 2021;2019: E219–E234. pmid:33555780
  9. 9. Her M. Repurposing and reshaping of hospitals during the COVID-19 outbreak in South Korea. One Heal. 2020;10: 100137. pmid:32373705
  10. 10. Liang W, Liang H, Ou L, Chen B, Chen A, Li C, et al. Development and validation of a clinical risk score to predict the occurrence of critical illness in hospitalized patients with COVID-19. JAMA Intern Med. 2020;180: 1081–1089. pmid:32396163
  11. 11. Yan L, Zhang H-T, Goncalves J, Xiao Y, Wang M, Guo Y, et al. An interpretable mortality prediction model for COVID-19 patients. Nat Mach Intell. 2020;2: 283–288.
  12. 12. Ji D, Xu J, Chen Z, Yang T, Zhao P, Chen G, et al. Prediction for Progression Risk in Patients with COVID-19 Pneumonia: the CALL Score. Clin Infect Dis. 2012;0954162: 1–4.
  13. 13. Rasmy L, Nigo M, Kannadath BS, Xie Z, Mao B, Patel K, et al. Recurrent neural network models (CovRNN) for predicting outcomes of patients with COVID-19 on admission to hospital: model development and validation using electronic health record data. Lancet Digit Heal. 2022;4: e415–e425. pmid:35466079
  14. 14. Tomašev N, Glorot X, Rae JW, Zielinski M, Askham H, Saraiva A, et al. A clinically applicable approach to continuous prediction of future acute kidney injury. Nature. 2019;572: 116–119. pmid:31367026
  15. 15. Roy S, Mincu D, Loreaux E, Mottram A, Protsyuk I, Harris N, et al. Multitask prediction of organ dysfunction in the intensive care unit using sequential subnetwork routing. J Am Med Informatics Assoc. 2021;28: 1936–1946. pmid:34151965
  16. 16. Hao B, Hu Y, Sotudian S, Zad Z, Adams WG, Assoumou SA, et al. Development and validation of predictive models for COVID-19 outcomes in a safety-net hospital population. J Am Med Informatics Assoc. 2022;29: 1253–1262. pmid:35441692
  17. 17. Rodrigues DS, Nastri ACS, Magri MM, Oliveira MS de, Sabino EC, Figueiredo PHMF, et al. Predicting the outcome for COVID-19 patients by applying time series classification to electronic health records. BMC Med Inform Decis Mak. 2022;22: 1–15. pmid:35843930
  18. 18. Marshall JC, Murthy S, Diaz J, Adhikari N, Angus DC, Arabi YM, et al. A minimal common outcome measure set for COVID-19 clinical research. Lancet Infect Dis. 2020;20: e192–e197. pmid:32539990
  19. 19. Jin Y, Schneeweiss S, Merola D, Lin KJ. Impact of longitudinal data-completeness of electronic health record data on risk score misclassification. J Am Med Inform Assoc. 2022;29: 1225–1232. pmid:35357470
  20. 20. Breiman L. Random_Forest. Mach Learn. 2001;45: 5–32.
  21. 21. Natekin A, Knoll A. Gradient boosting machines, a tutorial. Front Neurorobot. 2013;7. pmid:24409142
  22. 22. Murtagh F. Multilayer perceptrons for classification and regression. Neurocomputing. 1991;2: 183–197.
  23. 23. Phuong M, Hutter M. Formal Algorithms for Transformers. 2022; 1–16. Available: http://arxiv.org/abs/2207.09238
  24. 24. Vickers AJ, van Calster B, Steyerberg EW. A simple, step-by-step guide to interpreting decision curve analysis. Diagnostic Progn Res. 2019;3: 1–8. pmid:31592444
  25. 25. Matthay MA, Thompson BT, Ware LB. The Berlin definition of acute respiratory distress syndrome: should patients receiving high-flow nasal oxygen be included? Lancet Respir Med. 2021;9: 933–936. pmid:33915103
  26. 26. Schwab P, Mehrjou A, Parbhoo S, Celi LA, Hetzel J, Hofer M, et al. Real-time prediction of COVID-19 related mortality using electronic health records. Nat Commun. 2021;12. pmid:33594046
  27. 27. Lee J, Ta C, Kim JH, Liu C, Weng C. Severity Prediction for COVID-19 Patients via Recurrent Neural Networks. AMIA. Annu Symp proceedings AMIA Symp. 2021;2021: 374–383.
  28. 28. Park HJ, Jung DY, Ji W, Choi CM. Detection of bacteremia in surgical in-patients using recurrent neural network based on time series records: Development and validation study. J Med Internet Res. 2020;22: 1–10. pmid:32669261
  29. 29. Volovici V, Syn NL, Ercole A, Zhao JJ, Liu N. Steps to avoid overuse and misuse of machine learning in clinical research. Nat Med. 2022;28: 1996–1999. pmid:36097217
  30. 30. Grinsztajn L, Oyallon E, Varoquaux G. Why do tree-based models still outperform deep learning on tabular data? 2022. Available: http://arxiv.org/abs/2207.08815
  31. 31. Roimi M, Neuberger A, Shrot A, Paul M, Geffen Y, Bar-Lavie Y. Early diagnosis of bloodstream infections in the intensive care unit using machine-learning algorithms. Intensive Care Med. 2020;46: 454–462. pmid:31912208