Evaluation and mitigation of cognitive biases in medical language models

被引:1
|
作者
Schmidgall, Samuel [1 ]
Harris, Carl [2 ]
Essien, Ime [2 ]
Olshvang, Daniel [2 ]
Rahman, Tawsifur [2 ]
Kim, Ji Woong [3 ]
Ziaei, Rojin [4 ]
Eshraghian, Jason [5 ]
Abadir, Peter [6 ]
Chellappa, Rama [1 ,2 ]
机构
[1] Johns Hopkins Univ, Dept Elect & Comp Engn, Baltimore, MD 21218 USA
[2] Johns Hopkins Univ, Dept Biomed Engn, Baltimore, MD USA
[3] Johns Hopkins Univ, Dept Mech Engn, Baltimore, MD USA
[4] Univ Maryland, Dept Comp Sci, College Pk, MD USA
[5] Univ Calif Santa Cruz, Dept Elect & Comp Engn, Santa Cruz, CA USA
[6] Johns Hopkins Univ, Sch Med, Div Geriatr Med & Gerontol, Baltimore, MD USA
来源
NPJ DIGITAL MEDICINE | 2024年 / 7卷 / 01期
基金
美国国家科学基金会; 美国国家卫生研究院;
关键词
D O I
10.1038/s41746-024-01283-6
中图分类号
R19 [保健组织与事业(卫生事业管理)];
学科分类号
摘要
Increasing interest in applying large language models (LLMs) to medicine is due in part to their impressive performance on medical exam questions. However, these exams do not capture the complexity of real patient-doctor interactions because of factors like patient compliance, experience, and cognitive bias. We hypothesized that LLMs would produce less accurate responses when faced with clinically biased questions as compared to unbiased ones. To test this, we developed the BiasMedQA dataset, which consists of 1273 USMLE questions modified to replicate common clinically relevant cognitive biases. We assessed six LLMs on BiasMedQA and found that GPT-4 stood out for its resilience to bias, in contrast to Llama 2 70B-chat and PMC Llama 13B, which showed large drops in performance. Additionally, we introduced three bias mitigation strategies, which improved but did not fully restore accuracy. Our findings highlight the need to improve LLMs' robustness to cognitive biases, in order to achieve more reliable applications of LLMs in healthcare.
引用
收藏
页数:9
相关论文
共 50 条
  • [31] Learning with Enriched Inductive Biases for Vision-Language Models
    Yang, Lingxiao
    Zhang, Ru-Yuan
    Chen, Qi
    Xie, Xiaohua
    INTERNATIONAL JOURNAL OF COMPUTER VISION, 2025,
  • [32] Performance and biases of Large Language Models in public opinion simulation
    Qu, Yao
    Wang, Jue
    HUMANITIES & SOCIAL SCIENCES COMMUNICATIONS, 2024, 11 (01):
  • [33] Confirmation and Specificity Biases in Large Language Models: An Explorative Study
    O'Leary, Daniel E.
    IEEE INTELLIGENT SYSTEMS, 2025, 40 (01) : 63 - 68
  • [34] Unmasking the Mask - Evaluating Social Biases in Masked Language Models
    Kaneko, Masahiro
    Bollegala, Danushka
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 11954 - 11962
  • [35] Artificial Intelligence in mental health and the biases of language based models
    Straw, Isabel
    Callison-Burch, Chris
    PLOS ONE, 2020, 15 (12):
  • [36] Race, Gender, and Age Biases in Biomedical Masked Language Models
    Kim, Michelle YoungJin
    Kim, Junghwan
    Johnson, Kristen Marie
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023), 2023, : 11806 - 11815
  • [37] Detecting implicit biases of large language models with Bayesian hypothesis testingDetecting Implicit Biases of Large Language Models...S. Si et al.
    Shijing Si
    Xiaoming Jiang
    Qinliang Su
    Lawrence Carin
    Scientific Reports, 15 (1)
  • [38] Injecting structural hints: Using language models to study inductive biases in language learning
    Papadimitriou, Isabel
    Jurafsky, Dan
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EMNLP 2023), 2023, : 8402 - 8413
  • [39] COGNITIVE MODELS, COGNITIVE OPERATIONS, AND FIGURATIVE USES OF LANGUAGE
    de Mendoza Ibanez, Francisco Jose Ruiz
    Galera-Masegosa, Alicia
    FORMA Y FUNCION, 2012, 25 (02): : 11 - 38
  • [40] Evaluation of Generative Language Models in Personalizing Medical Information: Instrument Validation Study
    Spina, Aidin
    Andalib, Saman
    Flores, Daniel
    Vermani, Rishi
    Halaseh, Faris F.
    Nelson, M.
    JMIR AI, 2024, 3