AI Article Synopsis

  • Interpretability in healthcare is crucial, but deep learning models often lack it, hindering their use; this study introduces an attention layer in an LSTM neural network to improve model transparency in predicting patient outcomes.
  • Using data from 10,616 cardiovascular patients in the MIMIC III dataset, the model analyzes 48 clinical parameters over 10-hour sequences to predict death within a week, achieving an AUC of 0.790.
  • The study finds that attention weights align well with known risk factors, demonstrating the effectiveness of attention mechanisms in enhancing deep learning interpretability for electronic health records analysis.

Article Abstract

Interpretability is fundamental in healthcare problems and the lack of it in deep learning models is currently the major barrier in the usage of such powerful algorithms in the field. The study describes the implementation of an attention layer for Long Short-Term Memory (LSTM) neural network that provides a useful picture on the influence of the several input variables included in the model. A cohort of 10,616 patients with cardiovascular diseases is selected from the MIMIC III dataset, an openly available database of electronic health records (EHRs) including all patients admitted to an ICU at Boston's Medical Centre. For each patient, we consider a 10-length sequence of 1-hour windows in which 48 clinical parameters are extracted to predict the occurrence of death in the next 7 days. Inspired from the recent developments in the field of attention mechanisms for sequential data, we implement a recurrent neural network with LSTM cells incorporating an attention mechanism to identify features driving model's decisions over time. The performance of the LSTM model, measured in terms of AUC, is 0.790 (SD = 0.015). Regard our primary objective, i.e. model interpretability, we investigate the role of attention weights. We find good correspondence with driving predictors of a transparent model (r = 0.611, 95% CI [0.395, 0.763]). Moreover, most influential features identified at the cohort-level emerge as known risk factors in the clinical context. Despite the limitations of study dataset, this work brings further evidence of the potential of attention mechanisms in making deep learning model more interpretable and suggests the application of this strategy for the sequential analysis of EHRs.

Download full-text PDF

Source
http://dx.doi.org/10.1016/j.jbi.2021.103876DOI Listing

Publication Analysis

Top Keywords

deep learning
12
learning models
8
patients admitted
8
neural network
8
attention mechanisms
8
attention
5
model
5
interpretability time-series
4
time-series deep
4
models study
4

Similar Publications

Want AI Summaries of new PubMed Abstracts delivered to your In-box?

Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!