A toolbox for surfacing health equity harms and biases in large language models

被引:6
|
作者
Pfohl, Stephen R. [1 ]
Cole-Lewis, Heather [1 ]
Sayres, Rory [1 ]
Neal, Darlene [1 ]
Asiedu, Mercy [1 ]
Dieng, Awa [2 ]
Tomasev, Nenad [2 ]
Rashid, Qazi Mamunur [1 ]
Azizi, Shekoofeh [2 ]
Rostamzadeh, Negar [1 ]
Mccoy, Liam G. [3 ]
Celi, Leo Anthony [4 ,5 ,6 ]
Liu, Yun [1 ]
Schaekermann, Mike [1 ]
Walton, Alanna [2 ]
Parrish, Alicia [2 ]
Nagpal, Chirag [1 ]
Singh, Preeti [1 ]
Dewitt, Akeiylah [1 ]
Mansfield, Philip [2 ]
Prakash, Sushant [1 ]
Heller, Katherine [1 ]
Karthikesalingam, Alan [1 ]
Semturs, Christopher [1 ]
Barral, Joelle [2 ]
Corrado, Greg [1 ]
Matias, Yossi [1 ]
Smith-Loud, Jamila [1 ]
Horn, Ivor [1 ]
Singhal, Karan [1 ]
机构
[1] Google Res, Mountain View, CA 94043 USA
[2] Google DeepMind, Mountain View, CA USA
[3] Univ Alberta, Edmonton, AB, Canada
[4] MIT, Lab Computat Physiol, Cambridge, MA USA
[5] Beth Israel Deaconess Med Ctr, Div Pulm Crit Care & Sleep Med, Boston, MA USA
[6] Harvard TH Chan Sch Publ Hlth, Dept Biostat, Boston, MA USA
基金
美国国家科学基金会;
关键词
D O I
10.1038/s41591-024-03258-2
中图分类号
Q5 [生物化学]; Q7 [分子生物学];
学科分类号
071010 ; 081704 ;
摘要
Large language models (LLMs) hold promise to serve complex health information needs but also have the potential to introduce harm and exacerbate health disparities. Reliably evaluating equity-related model failures is a critical step toward developing systems that promote health equity. We present resources and methodologies for surfacing biases with potential to precipitate equity-related harms in long-form, LLM-generated answers to medical questions and conduct a large-scale empirical case study with the Med-PaLM 2 LLM. Our contributions include a multifactorial framework for human assessment of LLM-generated answers for biases and EquityMedQA, a collection of seven datasets enriched for adversarial queries. Both our human assessment framework and our dataset design process are grounded in an iterative participatory approach and review of Med-PaLM 2 answers. Through our empirical study, we find that our approach surfaces biases that may be missed by narrower evaluation approaches. Our experience underscores the importance of using diverse assessment methodologies and involving raters of varying backgrounds and expertise. While our approach is not sufficient to holistically assess whether the deployment of an artificial intelligence (AI) system promotes equitable health outcomes, we hope that it can be leveraged and built upon toward a shared goal of LLMs that promote accessible and equitable healthcare.
引用
收藏
页码:3590 / 3600
页数:30
相关论文
共 50 条
  • [1] A toolbox for surfacing health equity harms and biases in large language models
    Pfohl, Stephen R.
    Cole-Lewis, Heather
    Sayres, Rory
    Neal, Darlene
    Asiedu, Mercy
    Dieng, Awa
    Tomasev, Nenad
    Rashid, Qazi Mamunur
    Azizi, Shekoofeh
    Rostamzadeh, Negar
    Mccoy, Liam G.
    Celi, Leo Anthony
    Liu, Yun
    Schaekermann, Mike
    Walton, Alanna
    Parrish, Alicia
    Nagpal, Chirag
    Singh, Preeti
    Dewitt, Akeiylah
    Mansfield, Philip
    Prakash, Sushant
    Heller, Katherine
    Karthikesalingam, Alan
    Semturs, Christopher
    Barral, Joelle
    Corrado, Greg
    Matias, Yossi
    Smith-Loud, Jamila
    Horn, Ivor
    Singhal, Karan
    NATURE MEDICINE, 2024, 30 (12)
  • [2] Health Equity in the Era of Large Language Models
    Tierney, Aaron A.
    Reed, Mary E.
    Grant, Richard W.
    Doo, Florence X.
    Payan, Denise D.
    Liu, Vincent X.
    AMERICAN JOURNAL OF MANAGED CARE, 2025, 31 (03): : 112 - 117
  • [3] Economics and Equity of Large Language Models:Health CarePerspective
    Nagarajan, Radha
    Kondo, Midori
    Salas, Franz
    Sezgin, Emre
    Yao, Yuan
    Klotzman, Vanessa
    Godambe, Sandip A.
    Khan, Naqi
    Limon, Alfonso
    Stephenson, Graham
    Taraman, Sharief
    Walton, Nephi
    Ehwerhemuepha, Louis
    Pandit, Jay
    Pandita, Deepti
    Weiss, Michael
    Golden, Charles
    Gold, Adam
    Henderson, John
    Shippy, Angela
    Celi, Leo Anthony
    Hogan, William R.
    Oermann, Eric K.
    Sanger, Terence
    Martel, Steven
    JOURNAL OF MEDICAL INTERNET RESEARCH, 2024, 26
  • [4] Benchmarking Cognitive Biases in Large Language Models as Evaluators
    Koo, Ryan
    Lee, Minhwa
    Raheja, Vipul
    Park, Jongin
    Kim, Zae Myung
    Kang, Dongyeop
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 517 - 545
  • [5] Biases in Large Language Models: Origins, Inventory, and Discussion
    Navigli, Roberto
    Conia, Simone
    Ross, Bjorn
    ACM JOURNAL OF DATA AND INFORMATION QUALITY, 2023, 15 (02):
  • [6] (Ir)rationality and cognitive biases in large language models
    Macmillan-Scott, Olivia
    Musolesi, Mirco
    ROYAL SOCIETY OPEN SCIENCE, 2024, 11 (06):
  • [7] Leveraging large language models to foster equity in healthcare
    Rodriguez, Jorge A.
    Alsentzer, Emily
    Bates, David W.
    JOURNAL OF THE AMERICAN MEDICAL INFORMATICS ASSOCIATION, 2024, 31 (09)
  • [8] Performance and biases of Large Language Models in public opinion simulation
    Qu, Yao
    Wang, Jue
    HUMANITIES & SOCIAL SCIENCES COMMUNICATIONS, 2024, 11 (01):
  • [9] Confirmation and Specificity Biases in Large Language Models: An Explorative Study
    O'Leary, Daniel E.
    IEEE INTELLIGENT SYSTEMS, 2025, 40 (01) : 63 - 68
  • [10] Detecting implicit biases of large language models with Bayesian hypothesis testingDetecting Implicit Biases of Large Language Models...S. Si et al.
    Shijing Si
    Xiaoming Jiang
    Qinliang Su
    Lawrence Carin
    Scientific Reports, 15 (1)