Bias Unveiled: Enhancing Fairness in German Word Embeddings with Large Language Models

被引:0
|
作者
Saeid, Yasser [1 ]
Kopinski, Thomas [1 ]
机构
[1] South Westphalia Univ Appl Sci, Meschede, Germany
来源
关键词
Stereotypical biases; Gender bias; Machine learning systems; Word embedding algorithms; Bias amplification; Embedding bias; Origins of bias; Specific training documents; Efficacy; Abating bias; Methodology; Insights; Matrix; German Wikipedia corpora; Empirical endeavor; Precision; Sources of bias; Equanimity; Impartiality; LLM;
D O I
10.1007/978-3-031-78014-1_23
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
Gender bias in word embedding algorithms has garnered significant attention due to its integration into machine learning systems and its potential to reinforce stereotypes. Despite ongoing efforts, the root causes of biases in training word embeddings, specifically for the German language, remain unclear. This research presents a novel approach to tackling this problem, paving the way for new avenues of investigation. Our methodology involves a comprehensive analysis of word embeddings, focusing on how training data manipulations impact resulting biases. By examining how biases originate within specific training documents, we identify subsets that can be removed to effectively mitigate these effects. Additionally, we explore both conventional methods and new approaches using large language models (LLMs) to ensure the generated text adheres to concepts of fairness. Using few-shot prompting, we generate gender bias-free text, employing GPT-4 as a benchmark to evaluate the fairness of this process for the German language. Our method explains the intricate origins of biases within word embeddings, validated through rigorous application to German Wikipedia corpora. Our findings robustly demonstrate the efficacy of our method, showing that removing certain document subsets significantly diminishes bias in word embeddings. This is further detailed in our analysis, "Unlocking the Limits: Document Removal with an Upper Bound," in the experimental results section. Ultimately, this research presents a practical framework to uncover and mitigate biases in word embedding algorithms during training. Our goal is to advance machine learning systems that prioritize fairness and impartiality by revealing and addressing latent sources of bias.
引用
收藏
页码:308 / 325
页数:18
相关论文
共 50 条
  • [41] Knowledge Editing of Large Language Models Unconstrained by Word Order
    Ishigaki, Ryoma
    Suzuki, Jundai
    Shuzo, Masaki
    Maeda, Eisaku
    PROCEEDINGS OF THE 62ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 4: STUDENT RESEARCH WORKSHOP, 2024, : 177 - 187
  • [42] Measuring Impacts of Poisoning on Model Parameters and Embeddings for Large Language Models of Code
    Hussain, Aftab
    Rabin, Md Rafiqul Islam
    Alipour, Mohammad Amin
    PROCEEDINGS OF THE 1ST ACM INTERNATIONAL CONFERENCE ON AI-POWERED SOFTWARE, AIWARE 2024, 2024, : 59 - 64
  • [43] From Word Embeddings to Pre-Trained Language Models: A State-of-the-Art Walkthrough
    Mars, Mourad
    APPLIED SCIENCES-BASEL, 2022, 12 (17):
  • [44] A survey on multilingual large language models: corpora, alignment, and bias
    Xu, Yuemei
    Hu, Ling
    Zhao, Jiayi
    Qiu, Zihan
    Xu, Kexin
    Ye, Yuqi
    Gu, Hanwen
    FRONTIERS OF COMPUTER SCIENCE, 2025, 19 (11)
  • [45] Measuring Impacts of Poisoning on Model Parameters and Embeddings for Large Language Models of Code
    University of Houston, Houston, United States
    AIware - Proc. ACM Int. Conf. AI-Powered Softw., Co-located: ESEC/FSE, 1600, (59-64):
  • [46] Persistent Anti-Muslim Bias in Large Language Models
    Abid, Abubakar
    Farooqi, Maheen
    Zou, James
    AIES '21: PROCEEDINGS OF THE 2021 AAAI/ACM CONFERENCE ON AI, ETHICS, AND SOCIETY, 2021, : 298 - 306
  • [47] ROBBIE: Robust Bias Evaluation of Large Generative Language Models
    Esiobu, David
    Tan, Xiaoqing
    Hosseini, Saghar
    Ung, Megan
    Zhang, Yuchen
    Fernandes, Jude
    Dwivedi-Yu, Jane
    Presani, Eleonora
    Williams, Adina
    Meta, Eric Michael Smith
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING, EMNLP 2023, 2023, : 3764 - 3814
  • [48] Quantifying Bias in Agentic Large Language Models: A Benchmarking Approach
    Fernando, Riya
    Norton, Isabel
    Dogra, Pranay
    Sarnaik, Rohit
    Wazir, Hasan
    Ren, Zitang
    Gunda, Niveta Sree
    Mukhopadhyay, Anushka
    Lutz, Michael
    2024 5TH INFORMATION COMMUNICATION TECHNOLOGIES CONFERENCE, ICTC 2024, 2024, : 349 - 353
  • [49] Evaluating and Mitigating Gender Bias in Generative Large Language Models
    Zhou, H.
    Inkpen, D.
    Kantarci, B.
    INTERNATIONAL JOURNAL OF COMPUTERS COMMUNICATIONS & CONTROL, 2024, 19 (06)
  • [50] LARGE LANGUAGE MODELS FOR RISK OF BIAS ASSESSMENT: A CASE STUDY
    Edwards, M.
    Bishop, E.
    Reddish, K.
    Carr, E.
    di Ruffano, L. Ferrante
    VALUE IN HEALTH, 2024, 27 (12)