Accepted for/Published in: Journal of Medical Internet Research
Date Submitted: Sep 1, 2020
Date Accepted: Oct 2, 2020
Date Submitted to PubMed: Oct 7, 2020
Machine Learning to Predict Mortality and Critical Events in COVID-19 Positive New York City Patients: A Cohort Study
ABSTRACT
Background:
Coronavirus disease 2019 (COVID-19) has infected millions of patients worldwide and has been responsible for several hundred thousand fatalities. This has necessitated thoughtful resource allocation and early identification of high-risk patients. However, effective methods for achieving this are lacking.
Objective:
We analyze Electronic Health Records from COVID-19 positive hospitalized patients admitted to the Mount Sinai Health System in New York City (NYC). We present machine learning models for making predictions about the hospital course over clinically meaningful time horizons based on patient characteristics at admission. We assess performance of these models at multiple hospitals and time points.
Methods:
We utilized XGBoost and baseline comparator models, for predicting in-hospital mortality and critical events at time windows of 3, 5, 7 and 10 days from admission. Our study population included harmonized electronic health record (EHR) data from five hospitals in NYC for 4,098 COVID-19+ patients admitted from March 15, 2020 to May 22, 2020. Models were first trained on patients from a single hospital (N=1514) before or on May 1, externally validated on patients from four other hospitals (N=2201) before or on May 1, and prospectively validated on all patients after May 1 (N=383). Finally, we establish model interpretability to identify and rank variables that drive model predictions.
Results:
On the training set, the XGBoost classifier outperformed baseline models, with area under the receiver operating characteristic curve (AUC-ROC) for mortality at 0.89 at 3 days, 0.85 at 5 and 7 days, and 0.84 at 10 days; with area under the precision-recall curve (AU-PRC) of 0.45 at 3 days, 0.33 at 5 days, 0.44 at 7 days, and 0.48 at 10 days. XGBoost performed well for critical event prediction with AUC-ROC of 0.80 at 3 days, 0.79 at 5 days, 0.80 at 7 days, and 0.81 at 10 days and AU-PRC of 0.61 at 3 days, 0.62 at 5 days, 0.66 at 7 days, and 0.70 at 10 days. The trends in performance on both external and prospective validation sets were also similar to that of the training set. At 7 days, acute kidney injury on admission, elevated LDH, tachypnea, and hyperglycemia were the strongest drivers of critical event prediction, while higher age, anion gap, and C-reactive protein were the strongest drivers for mortality prediction.
Conclusions:
We trained and validated (both externally and prospectively) machine-learning models for mortality and critical events at different time horizons. These models identify at-risk patients, as well as uncover underlying relationships predicting outcomes.
Citation
Request queued. Please wait while the file is being generated. It may take some time.
Copyright
© The authors. All rights reserved. This is a privileged document currently under peer-review/community review (or an accepted/rejected manuscript). Authors have provided JMIR Publications with an exclusive license to publish this preprint on it's website for review and ahead-of-print citation purposes only. While the final peer-reviewed paper may be licensed under a cc-by license on publication, at this stage authors and publisher expressively prohibit redistribution of this draft paper other than for review purposes.