Skip to main content

Measuring Implicit Bias in ICU Notes Using Word-Embedding Neural Network Models.

Publication ,  Journal Article
Cobert, J; Mills, H; Lee, A; Gologorskaya, O; Espejo, E; Jeon, SY; Boscardin, WJ; Heintz, TA; Kennedy, CJ; Ashana, DC; Chapman, AC; Smith, AK ...
Published in: Chest
June 2024

BACKGROUND: Language in nonmedical data sets is known to transmit human-like biases when used in natural language processing (NLP) algorithms that can reinforce disparities. It is unclear if NLP algorithms of medical notes could lead to similar transmissions of biases. RESEARCH QUESTION: Can we identify implicit bias in clinical notes, and are biases stable across time and geography? STUDY DESIGN AND METHODS: To determine whether different racial and ethnic descriptors are similar contextually to stigmatizing language in ICU notes and whether these relationships are stable across time and geography, we identified notes on critically ill adults admitted to the University of California, San Francisco (UCSF), from 2012 through 2022 and to Beth Israel Deaconess Hospital (BIDMC) from 2001 through 2012. Because word meaning is derived largely from context, we trained unsupervised word-embedding algorithms to measure the similarity (cosine similarity) quantitatively of the context between a racial or ethnic descriptor (eg, African-American) and a stigmatizing target word (eg, nonco-operative) or group of words (violence, passivity, noncompliance, nonadherence). RESULTS: In UCSF notes, Black descriptors were less likely to be similar contextually to violent words compared with White descriptors. Contrastingly, in BIDMC notes, Black descriptors were more likely to be similar contextually to violent words compared with White descriptors. The UCSF data set also showed that Black descriptors were more similar contextually to passivity and noncompliance words compared with Latinx descriptors. INTERPRETATION: Implicit bias is identifiable in ICU notes. Racial and ethnic group descriptors carry different contextual relationships to stigmatizing words, depending on when and where notes were written. Because NLP models seem able to transmit implicit bias from training data, use of NLP algorithms in clinical prediction could reinforce disparities. Active debiasing strategies may be necessary to achieve algorithmic fairness when using language models in clinical research.

Duke Scholars

Altmetric Attention Stats
Dimensions Citation Stats

Published In

Chest

DOI

EISSN

1931-3543

Publication Date

June 2024

Volume

165

Issue

6

Start / End Page

1481 / 1490

Location

United States

Related Subject Headings

  • Respiratory System
  • Neural Networks, Computer
  • Natural Language Processing
  • Male
  • Intensive Care Units
  • Humans
  • Female
  • Electronic Health Records
  • Critical Illness
  • Bias
 

Citation

APA
Chicago
ICMJE
MLA
NLM
Cobert, J., Mills, H., Lee, A., Gologorskaya, O., Espejo, E., Jeon, S. Y., … Lee, S. J. (2024). Measuring Implicit Bias in ICU Notes Using Word-Embedding Neural Network Models. Chest, 165(6), 1481–1490. https://doi.org/10.1016/j.chest.2023.12.031
Cobert, Julien, Hunter Mills, Albert Lee, Oksana Gologorskaya, Edie Espejo, Sun Young Jeon, W John Boscardin, et al. “Measuring Implicit Bias in ICU Notes Using Word-Embedding Neural Network Models.Chest 165, no. 6 (June 2024): 1481–90. https://doi.org/10.1016/j.chest.2023.12.031.
Cobert J, Mills H, Lee A, Gologorskaya O, Espejo E, Jeon SY, et al. Measuring Implicit Bias in ICU Notes Using Word-Embedding Neural Network Models. Chest. 2024 Jun;165(6):1481–90.
Cobert, Julien, et al. “Measuring Implicit Bias in ICU Notes Using Word-Embedding Neural Network Models.Chest, vol. 165, no. 6, June 2024, pp. 1481–90. Pubmed, doi:10.1016/j.chest.2023.12.031.
Cobert J, Mills H, Lee A, Gologorskaya O, Espejo E, Jeon SY, Boscardin WJ, Heintz TA, Kennedy CJ, Ashana DC, Chapman AC, Raghunathan K, Smith AK, Lee SJ. Measuring Implicit Bias in ICU Notes Using Word-Embedding Neural Network Models. Chest. 2024 Jun;165(6):1481–1490.

Published In

Chest

DOI

EISSN

1931-3543

Publication Date

June 2024

Volume

165

Issue

6

Start / End Page

1481 / 1490

Location

United States

Related Subject Headings

  • Respiratory System
  • Neural Networks, Computer
  • Natural Language Processing
  • Male
  • Intensive Care Units
  • Humans
  • Female
  • Electronic Health Records
  • Critical Illness
  • Bias