Implicit bias in large language models: Experimental proof and implications for education

被引:2
|
作者
Warr, Melissa [1 ]
Oster, Nicole Jakubczyk [2 ]
Isaac, Roger [1 ]
机构
[1] New Mexico State Univ, POB 30001, Las Cruces, NM 88003 USA
[2] Arizona State Univ, Tempe, AZ USA
关键词
Generative AI; large language models; critical technology studies; systemic bias; systemic inequity; ACHIEVEMENT GAP; SCHOOL; IDENTITY;
D O I
10.1080/15391523.2024.2395295
中图分类号
G40 [教育学];
学科分类号
040101 ; 120403 ;
摘要
We provide experimental evidence of implicit racial bias in a large language model (specifically ChatGPT 3.5) in the context of an educational task and discuss implications for the use of these tools in educational contexts. Specifically, we presented ChatGPT with identical student writing passages alongside various descriptions of student demographics, including race, socioeconomic status, and school type. Results indicate that when directly prompted to consider race, the model produced higher overall scores than responses to a control prompt, but scores given to student descriptors of Black and White were not significantly different. However, this result belied a subtler form of prejudice that was statistically significant when racial indicators were implied rather than explicitly stated. Additionally, our investigation uncovered subtle sequence effects that suggest the model is more likely to illustrate bias when variables change within a single chat. The evidence indicates that despite the implementation of guardrails by developers, biases are profoundly embedded in ChatGPT, reflective of both the training data and societal biases at large. While overt biases can be addressed to some extent, the more ingrained implicit biases present a greater challenge for the application of these technologies in education. It is critical to develop an understanding of the bias embedded in these models and how this bias presents itself in educational contexts before using LLMs to develop personalized learning tools.
引用
收藏
页数:26
相关论文
共 50 条
  • [31] From hype to evidence: exploring large language models for inter-group bias classification in higher education
    Albuquerque, Josmario
    Rienties, Bart
    Holmes, Wayne
    Hlosta, Martin
    INTERACTIVE LEARNING ENVIRONMENTS, 2024,
  • [32] IMPLICIT BIAS IN UNDERGRADUATE MEDICAL EDUCATION: DEVELOPMENT AND VALIDATION OF THE IMPLICIT BIAS ATTITUDE SCALE
    Gonzalez, Cristina M.
    Garba, Ramya J.
    Grochowalski, Joseph
    Marantz, Paul R.
    JOURNAL OF GENERAL INTERNAL MEDICINE, 2016, 31 : S268 - S269
  • [33] Large language models: implications of rapid evolution in medicine
    Cheung, Billy H. H.
    Co, Michael T. H.
    HONG KONG MEDICAL JOURNAL, 2023, 29 (06) : 557 - 560
  • [34] Implications of large language models such as ChatGPT for dental medicine
    Eggmann, Florin
    Weiger, Roland
    Zitzmann, Nicola U.
    Blatz, Markus B.
    JOURNAL OF ESTHETIC AND RESTORATIVE DENTISTRY, 2023, 35 (07) : 1098 - 1102
  • [35] A Systemic Review of Large Language Models and Their Implications in Dermatology
    Lewandowski, Milosz
    Kropidlowska, Julia
    Kvinen, Alexandra
    Baranska-Rybak, Wioletta
    AUSTRALASIAN JOURNAL OF DERMATOLOGY, 2025,
  • [36] Large language models challenge the future of higher education
    Milano, Silvia
    McGrane, Joshua A.
    Leonelli, Sabina
    NATURE MACHINE INTELLIGENCE, 2023, 5 (04) : 333 - 334
  • [37] Large language models challenge the future of higher education
    Silvia Milano
    Joshua A. McGrane
    Sabina Leonelli
    Nature Machine Intelligence, 2023, 5 : 333 - 334
  • [38] Human bias in AI models? Anchoring effects and mitigation strategies in large language models
    Nguyen, Jeremy K.
    JOURNAL OF BEHAVIORAL AND EXPERIMENTAL FINANCE, 2024, 43
  • [39] Using Large Language Models to Investigate and Categorize Bias in Clinical Documentation
    Apakama, D.
    Klang, E.
    Richardson, L.
    Nadkarni, G.
    ANNALS OF EMERGENCY MEDICINE, 2024, 84 (04) : S96 - S97
  • [40] Likelihood-based Mitigation of Evaluation Bias in Large Language Models
    Ohi, Masanari
    Kaneko, Masahiro
    Koike, Ryuto
    Loem, Mengsay
    Okazaki, Naoaki
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 3237 - 3245