AI Article Synopsis

  • The study examines how stigma and bias related to race and other minoritized statuses affect pregnancy and birth outcomes by analyzing stigmatizing language in electronic health records.
  • Researchers developed automated natural language processing (NLP) methods to identify two types of stigmatizing language in labor and birth notes: marginalizing language and power/privilege language.
  • The results showed that Decision Trees were most effective for marginalizing language with an F-score of 0.73, while Support Vector Machines excelled in identifying power/privilege language with an F-score of 0.91, marking a significant advancement in using NLP to detect bias in medical documentation.

Article Abstract

Introduction: Stigma and bias related to race and other minoritized statuses may underlie disparities in pregnancy and birth outcomes. One emerging method to identify bias is the study of stigmatizing language in the electronic health record. The objective of our study was to develop automated natural language processing (NLP) methods to identify two types of stigmatizing language: marginalizing language and its complement, power/privilege language, accurately and automatically in labor and birth notes.

Methods: We analyzed notes for all birthing people > 20 weeks' gestation admitted for labor and birth at two hospitals during 2017. We then employed text preprocessing techniques, specifically using TF-IDF values as inputs, and tested machine learning classification algorithms to identify stigmatizing and power/privilege language in clinical notes. The algorithms assessed included Decision Trees, Random Forest, and Support Vector Machines. Additionally, we applied a feature importance evaluation method (InfoGain) to discern words that are highly correlated with these language categories.

Results: For marginalizing language, Decision Trees yielded the best classification with an F-score of 0.73. For power/privilege language, Support Vector Machines performed optimally, achieving an F-score of 0.91. These results demonstrate the effectiveness of the selected machine learning methods in classifying language categories in clinical notes.

Conclusion: We identified well-performing machine learning methods to automatically detect stigmatizing language in clinical notes. To our knowledge, this is the first study to use NLP performance metrics to evaluate the performance of machine learning methods in discerning stigmatizing language. Future studies should delve deeper into refining and evaluating NLP methods, incorporating the latest algorithms rooted in deep learning.

Download full-text PDF

Source
http://dx.doi.org/10.1007/s10995-023-03857-4DOI Listing

Publication Analysis

Top Keywords

stigmatizing language
20
machine learning
16
language
13
labor birth
12
clinical notes
12
power/privilege language
12
learning methods
12
natural language
8
language processing
8
identify stigmatizing
8

Similar Publications

Background: Stigmatising language is used commonly in healthcare, affecting healthcare providers' perceptions of patients and care delivery. Using person-first language is best practice, however, it does not reflect reality.

Method: This study examined medical students' perspectives on stigmatising language in healthcare.

View Article and Find Full Text PDF

Despite increased attempts to express equality in speech, biases often leak out through subtle linguistic cues. For example, the subject-complement statement (SCS, "Girls are as good as boys at math") is used to advocate for equality but often reinforces gender stereotypes (boys are the standard against which girls are judged). We ask whether stereotypes conveyed by SCS can be counteracted by gesture.

View Article and Find Full Text PDF

Background: Patients in the United States have recently gained federally mandated, free, and ready electronic access to clinicians' computerized notes in their medical records ("open notes"). This change from longstanding practice can benefit patients in clinically important ways, but studies show some patients feel judged or stigmatized by words or phrases embedded in their records. Therefore, it is imperative that clinicians adopt documentation techniques that help both to empower patients and minimize potential harms.

View Article and Find Full Text PDF

Many labels are used within and across subfields to describe personality disorder (PD) and interpersonally-oriented trait dimensions. For example, "interpersonal disorders" is a suggested alternative label to "personality disorders" in clinical research. Other "dark trait" terms, though not proposed as formal labels for PDs, also are used in different research areas for describing externalizing traits.

View Article and Find Full Text PDF

Want AI Summaries of new PubMed Abstracts delivered to your In-box?

Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!