Clinically relevant pretraining is all you need.

J Am Med Inform Assoc

Department of Biomedical Informatics, Columbia University, New York, New York, USA.

Published: August 2021

Clinical notes present a wealth of information for applications in the clinical domain, but heterogeneity across clinical institutions and settings presents challenges for their processing. The clinical natural language processing field has made strides in overcoming domain heterogeneity, while pretrained deep learning models present opportunities to transfer knowledge from one task to another. Pretrained models have performed well when transferred to new tasks; however, it is not well understood if these models generalize across differences in institutions and settings within the clinical domain. We explore if institution or setting specific pretraining is necessary for pretrained models to perform well when transferred to new tasks. We find no significant performance difference between models pretrained across institutions and settings, indicating that clinically pretrained models transfer well across such boundaries. Given a clinically pretrained model, clinical natural language processing researchers may forgo the time-consuming pretraining step without a significant performance drop.

Download full-text PDF

Source
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC8363791PMC
http://dx.doi.org/10.1093/jamia/ocab086DOI Listing

Publication Analysis

Top Keywords

institutions settings
12
pretrained models
12
clinical domain
8
domain heterogeneity
8
clinical natural
8
natural language
8
language processing
8
well transferred
8
transferred tasks
8
clinically pretrained
8

Similar Publications

Want AI Summaries of new PubMed Abstracts delivered to your In-box?

Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!