Machine learning models for predicting emergency department outcomes and routing
This retrospective observational study analyzed data from the MIMIC-IV-ED database, which includes more than 440,000 emergency visits, to compare machine learning models with standard clinical scoring systems for predicting outcomes and supporting dynamic patient routing. Models evaluated included gradient boosting, interpretable ML, classical algorithms, and deep learning. The primary outcome was not reported; secondary outcomes were hospitalization post-ED visit, critical deterioration (defined as ICU transfer or death within 12 hours), and 72-hour ED re-attendance. For all three secondary outcomes, gradient boosting algorithms performed better than standard clinical scoring systems and complex deep learning models. Predictive performance was reported as AUROC 0.820 for hospitalization, AUROC 0.881 for critical deterioration, and AUROC 0.699 for 72-hour re-attendance. The authors note that this is not a primary clinical trial and that results are based on a single database, limiting generalizability and real-world implementation. No adverse events, follow-up duration, or absolute outcome rates were reported. The authors provide evidence-based recommendations for intelligent patient routing systems to enhance emergency care efficiency and resource utilization, but these findings reflect predictive associations rather than causal improvements in patient outcomes or workflow.