RESUMEN
Clinical deterioration of hospitalized patients is common and can lead to critical illness and death. Rapid response teams (RRTs) assess and treat high-risk patients with signs of clinical deterioration to prevent further worsening and subsequent adverse outcomes. Whether activation of the RRT early in the course of clinical deterioration impacts outcomes, however, remains unclear. We sought to characterize the relationship between increasing time to RRT activation after physiologic deterioration and short-term patient outcomes. DESIGN: Retrospective multicenter cohort study. SETTING: Three academic hospitals in Pennsylvania. PATIENTS: We included the RRT activation of a hospitalization for non-ICU inpatients greater than or equal to 18 years old. INTERVENTIONS: None. MEASUREMENTS AND MAIN RESULTS: The primary exposure was time to RRT activation after physiologic deterioration. We selected four Cardiac Arrest Risk Triage (CART) score thresholds a priori from which to measure time to RRT activation (CART score ≥ 12, ≥ 16, ≥ 20, and ≥ 24). The primary outcome was 7-day mortality-death or discharge to hospice care within 7 days of RRT activation. For each CART threshold, we modeled the association of time to RRT activation duration with 7-day mortality using multivariable fractional polynomial regression. Increased time from clinical decompensation to RRT activation was associated with higher risk of 7-day mortality. This relationship was nonlinear, with odds of mortality increasing rapidly as time to RRT activation increased from 0 to 4 hours and then plateauing. This pattern was observed across several thresholds of physiologic derangement. CONCLUSIONS: Increasing time to RRT activation was associated in a nonlinear fashion with increased 7-day mortality. This relationship appeared most marked when using a CART score greater than 20 threshold from which to measure time to RRT activation. We suggest that these empirical findings could be used to inform RRT delay definitions in further studies to determine the clinical impact of interventions focused on timely RRT activation.
RESUMEN
OBJECTIVES: The National Early Warning Score, Modified Early Warning Score, and quick Sepsis-related Organ Failure Assessment can predict clinical deterioration. These scores exhibit only moderate performance and are often evaluated using aggregated measures over time. A simulated prospective validation strategy that assesses multiple predictions per patient-day would provide the best pragmatic evaluation. We developed a deep recurrent neural network deterioration model and conducted a simulated prospective evaluation. DESIGN: Retrospective cohort study. SETTING: Four hospitals in Pennsylvania. PATIENTS: Inpatient adults discharged between July 1, 2017, and June 30, 2019. INTERVENTIONS: None. MEASUREMENTS AND MAIN RESULTS: We trained a deep recurrent neural network and logistic regression model using data from electronic health records to predict hourly the 24-hour composite outcome of transfer to ICU or death. We analyzed 146,446 hospitalizations with 16.75 million patient-hours. The hourly event rate was 1.6% (12,842 transfers or deaths, corresponding to 260,295 patient-hours within the predictive horizon). On a hold-out dataset, the deep recurrent neural network achieved an area under the precision-recall curve of 0.042 (95% CI, 0.04-0.043), comparable with logistic regression model (0.043; 95% CI 0.041 to 0.045), and outperformed National Early Warning Score (0.034; 95% CI, 0.032-0.035), Modified Early Warning Score (0.028; 95% CI, 0.027- 0.03), and quick Sepsis-related Organ Failure Assessment (0.021; 95% CI, 0.021-0.022). For a fixed sensitivity of 50%, the deep recurrent neural network achieved a positive predictive value of 3.4% (95% CI, 3.4-3.5) and outperformed logistic regression model (3.1%; 95% CI 3.1-3.2), National Early Warning Score (2.0%; 95% CI, 2.0-2.0), Modified Early Warning Score (1.5%; 95% CI, 1.5-1.5), and quick Sepsis-related Organ Failure Assessment (1.5%; 95% CI, 1.5-1.5). CONCLUSIONS: Commonly used early warning scores for clinical decompensation, along with a logistic regression model and a deep recurrent neural network model, show very poor performance characteristics when assessed using a simulated prospective validation. None of these models may be suitable for real-time deployment.
Asunto(s)
Deterioro Clínico , Cuidados Críticos/normas , Aprendizaje Profundo/normas , Puntuaciones en la Disfunción de Órganos , Sepsis/terapia , Adulto , Humanos , Masculino , Persona de Mediana Edad , Pennsylvania , Estudios Retrospectivos , Medición de RiesgoRESUMEN
OBJECTIVES: Develop and implement a machine learning algorithm to predict severe sepsis and septic shock and evaluate the impact on clinical practice and patient outcomes. DESIGN: Retrospective cohort for algorithm derivation and validation, pre-post impact evaluation. SETTING: Tertiary teaching hospital system in Philadelphia, PA. PATIENTS: All non-ICU admissions; algorithm derivation July 2011 to June 2014 (n = 162,212); algorithm validation October to December 2015 (n = 10,448); silent versus alert comparison January 2016 to February 2017 (silent n = 22,280; alert n = 32,184). INTERVENTIONS: A random-forest classifier, derived and validated using electronic health record data, was deployed both silently and later with an alert to notify clinical teams of sepsis prediction. MEASUREMENT AND MAIN RESULT: Patients identified for training the algorithm were required to have International Classification of Diseases, 9th Edition codes for severe sepsis or septic shock and a positive blood culture during their hospital encounter with either a lactate greater than 2.2 mmol/L or a systolic blood pressure less than 90 mm Hg. The algorithm demonstrated a sensitivity of 26% and specificity of 98%, with a positive predictive value of 29% and positive likelihood ratio of 13. The alert resulted in a small statistically significant increase in lactate testing and IV fluid administration. There was no significant difference in mortality, discharge disposition, or transfer to ICU, although there was a reduction in time-to-ICU transfer. CONCLUSIONS: Our machine learning algorithm can predict, with low sensitivity but high specificity, the impending occurrence of severe sepsis and septic shock. Algorithm-generated predictive alerts modestly impacted clinical measures. Next steps include describing clinical perception of this tool and optimizing algorithm design and delivery.
Asunto(s)
Algoritmos , Sistemas de Apoyo a Decisiones Clínicas , Diagnóstico por Computador , Aprendizaje Automático , Sepsis/diagnóstico , Choque Séptico/diagnóstico , Estudios de Cohortes , Registros Electrónicos de Salud , Hospitales de Enseñanza , Humanos , Estudios Retrospectivos , Sensibilidad y Especificidad , Envío de Mensajes de TextoRESUMEN
OBJECTIVE: To assess clinician perceptions of a machine learning-based early warning system to predict severe sepsis and septic shock (Early Warning System 2.0). DESIGN: Prospective observational study. SETTING: Tertiary teaching hospital in Philadelphia, PA. PATIENTS: Non-ICU admissions November-December 2016. INTERVENTIONS: During a 6-week study period conducted 5 months after Early Warning System 2.0 alert implementation, nurses and providers were surveyed twice about their perceptions of the alert's helpfulness and impact on care, first within 6 hours of the alert, and again 48 hours after the alert. MEASUREMENTS AND MAIN RESULTS: For the 362 alerts triggered, 180 nurses (50% response rate) and 107 providers (30% response rate) completed the first survey. Of these, 43 nurses (24% response rate) and 44 providers (41% response rate) completed the second survey. Few (24% nurses, 13% providers) identified new clinical findings after responding to the alert. Perceptions of the presence of sepsis at the time of alert were discrepant between nurses (13%) and providers (40%). The majority of clinicians reported no change in perception of the patient's risk for sepsis (55% nurses, 62% providers). A third of nurses (30%) but few providers (9%) reported the alert changed management. Almost half of nurses (42%) but less than a fifth of providers (16%) found the alert helpful at 6 hours. CONCLUSIONS: In general, clinical perceptions of Early Warning System 2.0 were poor. Nurses and providers differed in their perceptions of sepsis and alert benefits. These findings highlight the challenges of achieving acceptance of predictive and machine learning-based sepsis alerts.