Bias Unveiled: Enhancing Fairness in German Word Embeddings with Large Language Models

被引:0
|
作者
Saeid, Yasser [1 ]
Kopinski, Thomas [1 ]
机构
[1] South Westphalia Univ Appl Sci, Meschede, Germany
来源
关键词
Stereotypical biases; Gender bias; Machine learning systems; Word embedding algorithms; Bias amplification; Embedding bias; Origins of bias; Specific training documents; Efficacy; Abating bias; Methodology; Insights; Matrix; German Wikipedia corpora; Empirical endeavor; Precision; Sources of bias; Equanimity; Impartiality; LLM;
D O I
10.1007/978-3-031-78014-1_23
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
Gender bias in word embedding algorithms has garnered significant attention due to its integration into machine learning systems and its potential to reinforce stereotypes. Despite ongoing efforts, the root causes of biases in training word embeddings, specifically for the German language, remain unclear. This research presents a novel approach to tackling this problem, paving the way for new avenues of investigation. Our methodology involves a comprehensive analysis of word embeddings, focusing on how training data manipulations impact resulting biases. By examining how biases originate within specific training documents, we identify subsets that can be removed to effectively mitigate these effects. Additionally, we explore both conventional methods and new approaches using large language models (LLMs) to ensure the generated text adheres to concepts of fairness. Using few-shot prompting, we generate gender bias-free text, employing GPT-4 as a benchmark to evaluate the fairness of this process for the German language. Our method explains the intricate origins of biases within word embeddings, validated through rigorous application to German Wikipedia corpora. Our findings robustly demonstrate the efficacy of our method, showing that removing certain document subsets significantly diminishes bias in word embeddings. This is further detailed in our analysis, "Unlocking the Limits: Document Removal with an Upper Bound," in the experimental results section. Ultimately, this research presents a practical framework to uncover and mitigate biases in word embedding algorithms during training. Our goal is to advance machine learning systems that prioritize fairness and impartiality by revealing and addressing latent sources of bias.
引用
收藏
页码:308 / 325
页数:18
相关论文
共 50 条
  • [21] Probing into the Fairness of Large Language Models: A Case Study of ChatGPT
    Li, Yunqi
    Zhang, Lanjing
    Zhang, Yongfeng
    2024 58TH ANNUAL CONFERENCE ON INFORMATION SCIENCES AND SYSTEMS, CISS, 2024,
  • [22] Saxony-Anhalt is the Worst: Bias Towards German Federal States in Large Language Models
    Kruspe, Anna
    Stillman, Mila
    KI 2024: ADVANCES IN ARTIFICIAL INTELLIGENCE, KI 2024, 2024, 14992 : 160 - 174
  • [23] Compressed models for co-reference resolution: enhancing efficiency with debiased word embeddings
    Georgios Ioannides
    Aishwarya Jadhav
    Aditi Sharma
    Samarth Navali
    Alan W. Black
    Scientific Reports, 13
  • [24] Compressed models for co-reference resolution: enhancing efficiency with debiased word embeddings
    Ioannides, Georgios
    Jadhav, Aishwarya
    Sharma, Aditi
    Navali, Samarth
    Black, Alan W.
    SCIENTIFIC REPORTS, 2023, 13 (01)
  • [25] What's the next word in large language models?
    不详
    NATURE MACHINE INTELLIGENCE, 2023, 5 (04) : 331 - 332
  • [26] What’s the next word in large language models?
    Nature Machine Intelligence, 2023, 5 : 331 - 332
  • [27] Leveraging large language models for word sense disambiguation
    Jung H. Yae
    Nolan C. Skelly
    Neil C. Ranly
    Phillip M. LaCasse
    Neural Computing and Applications, 2025, 37 (6) : 4093 - 4110
  • [28] Pipelines for Social Bias Testing of Large Language Models
    Nozza, Debora
    Bianchi, Federico
    Hovy, Dirk
    PROCEEDINGS OF WORKSHOP ON CHALLENGES & PERSPECTIVES IN CREATING LARGE LANGUAGE MODELS (BIGSCIENCE EPISODE #5), 2022, : 68 - 74
  • [29] A Causal View of Entity Bias in (Large) Language Models
    Wang, Fei
    Mo, Wenjie
    Wang, Yiwei
    Zhou, Wenxuan
    Chen, Muhao
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EMNLP 2023), 2023, : 15173 - 15184
  • [30] Cultural bias and cultural alignment of large language models
    Tao, Yan
    Viberg, Olga
    Baker, Ryan S.
    Kizilcec, Rene F.
    PNAS NEXUS, 2024, 3 (09):