Implicit bias in large language models: Experimental proof and implications for education

被引:2
|
作者
Warr, Melissa [1 ]
Oster, Nicole Jakubczyk [2 ]
Isaac, Roger [1 ]
机构
[1] New Mexico State Univ, POB 30001, Las Cruces, NM 88003 USA
[2] Arizona State Univ, Tempe, AZ USA
关键词
Generative AI; large language models; critical technology studies; systemic bias; systemic inequity; ACHIEVEMENT GAP; SCHOOL; IDENTITY;
D O I
10.1080/15391523.2024.2395295
中图分类号
G40 [教育学];
学科分类号
040101 ; 120403 ;
摘要
We provide experimental evidence of implicit racial bias in a large language model (specifically ChatGPT 3.5) in the context of an educational task and discuss implications for the use of these tools in educational contexts. Specifically, we presented ChatGPT with identical student writing passages alongside various descriptions of student demographics, including race, socioeconomic status, and school type. Results indicate that when directly prompted to consider race, the model produced higher overall scores than responses to a control prompt, but scores given to student descriptors of Black and White were not significantly different. However, this result belied a subtler form of prejudice that was statistically significant when racial indicators were implied rather than explicitly stated. Additionally, our investigation uncovered subtle sequence effects that suggest the model is more likely to illustrate bias when variables change within a single chat. The evidence indicates that despite the implementation of guardrails by developers, biases are profoundly embedded in ChatGPT, reflective of both the training data and societal biases at large. While overt biases can be addressed to some extent, the more ingrained implicit biases present a greater challenge for the application of these technologies in education. It is critical to develop an understanding of the bias embedded in these models and how this bias presents itself in educational contexts before using LLMs to develop personalized learning tools.
引用
收藏
页数:26
相关论文
共 50 条
  • [21] Explicit and implicit justifications of experimental procedures in language education: Pedagogical implications of studying expert writers' communicative resources
    Lim, Jason Miin-Hwa
    JOURNAL OF ENGLISH FOR ACADEMIC PURPOSES, 2019, 37 : 34 - 51
  • [22] A survey on multilingual large language models: corpora, alignment, and bias
    Xu, Yuemei
    Hu, Ling
    Zhao, Jiayi
    Qiu, Zihan
    Xu, Kexin
    Ye, Yuqi
    Gu, Hanwen
    FRONTIERS OF COMPUTER SCIENCE, 2025, 19 (11)
  • [23] Persistent Anti-Muslim Bias in Large Language Models
    Abid, Abubakar
    Farooqi, Maheen
    Zou, James
    AIES '21: PROCEEDINGS OF THE 2021 AAAI/ACM CONFERENCE ON AI, ETHICS, AND SOCIETY, 2021, : 298 - 306
  • [24] ROBBIE: Robust Bias Evaluation of Large Generative Language Models
    Esiobu, David
    Tan, Xiaoqing
    Hosseini, Saghar
    Ung, Megan
    Zhang, Yuchen
    Fernandes, Jude
    Dwivedi-Yu, Jane
    Presani, Eleonora
    Williams, Adina
    Meta, Eric Michael Smith
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING, EMNLP 2023, 2023, : 3764 - 3814
  • [25] Quantifying Bias in Agentic Large Language Models: A Benchmarking Approach
    Fernando, Riya
    Norton, Isabel
    Dogra, Pranay
    Sarnaik, Rohit
    Wazir, Hasan
    Ren, Zitang
    Gunda, Niveta Sree
    Mukhopadhyay, Anushka
    Lutz, Michael
    2024 5TH INFORMATION COMMUNICATION TECHNOLOGIES CONFERENCE, ICTC 2024, 2024, : 349 - 353
  • [26] Evaluating and Mitigating Gender Bias in Generative Large Language Models
    Zhou, H.
    Inkpen, D.
    Kantarci, B.
    INTERNATIONAL JOURNAL OF COMPUTERS COMMUNICATIONS & CONTROL, 2024, 19 (06)
  • [27] LARGE LANGUAGE MODELS FOR RISK OF BIAS ASSESSMENT: A CASE STUDY
    Edwards, M.
    Bishop, E.
    Reddish, K.
    Carr, E.
    di Ruffano, L. Ferrante
    VALUE IN HEALTH, 2024, 27 (12)
  • [28] The Nature and Implications of Implicit Weight Bias
    Watts, Kaaren
    Cranney, Jacquelyn
    CURRENT PSYCHIATRY REVIEWS, 2009, 5 (02) : 110 - 126
  • [29] Integrating Implicit Bias Into Counselor Education
    Boysen, Guy A.
    COUNSELOR EDUCATION AND SUPERVISION, 2010, 49 (04) : 210 - 227
  • [30] Baldur: Whole-Proof Generation and Repair with Large Language Models
    First, Emily
    Rabe, Markus N.
    Ringer, Talia
    Brun, Yuriy
    PROCEEDINGS OF THE 31ST ACM JOINT MEETING EUROPEAN SOFTWARE ENGINEERING CONFERENCE AND SYMPOSIUM ON THE FOUNDATIONS OF SOFTWARE ENGINEERING, ESEC/FSE 2023, 2023, : 1229 - 1241