Measuring Gender Bias in Word Embeddings across Domains and Discovering New Gender Bias Word Categories

被引:0
|
作者
Chaloner, Kaytlin [1 ]
Maldonado, Alfredo [1 ]
机构
[1] Trinity Coll Dublin, ADAPT Ctr, SCSS, Dublin, Ireland
来源
GENDER BIAS IN NATURAL LANGUAGE PROCESSING (GEBNLP 2019) | 2019年
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Prior work has shown that word embeddings capture human stereotypes, including gender bias. However, there is a lack of studies testing the presence of specific gender bias categories in word embeddings across diverse domains. This paper aims to fill this gap by applying the WEAT bias detection method to four sets of word embeddings trained on corpora from four different domains: news, social networking, biomedical and a gender-balanced corpus extracted from Wikipedia (GAP). We find that some domains are definitely more prone to gender bias than others, and that the categories of gender bias present also vary for each set of word embeddings. We detect some gender bias in GAP. We also propose a simple but novel method for discovering new bias categories by clustering word embeddings. We validate this method through WEAT's hypothesis testing mechanism and find it useful for expanding the relatively small set of well-known gender bias word categories commonly used in the literature.
引用
收藏
页码:25 / 32
页数:8
相关论文
共 50 条
  • [31] Word embeddings are biased. But whose bias are they reflecting?
    Davor Petreski
    Ibrahim C. Hashim
    AI & SOCIETY, 2023, 38 : 975 - 982
  • [32] Gender-sensitive word embeddings for healthcare
    Agmon, Shunit
    Gillis, Plia
    Horvitz, Eric
    Radinsky, Kira
    JOURNAL OF THE AMERICAN MEDICAL INFORMATICS ASSOCIATION, 2022, 29 (03) : 415 - 423
  • [33] Learning Gender-Neutral Word Embeddings
    Zhao, Jieyu
    Zhou, Yichao
    Li, Zeyu
    Wang, Wei
    Chang, Kai-Wei
    2018 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2018), 2018, : 4847 - 4853
  • [34] Understanding Gender Bias in Knowledge Base Embeddings
    Du, Yupei
    Zheng, Qi
    Wu, Yuanbin
    Lan, Man
    Yang, Yan
    Ma, Meirong
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 1381 - 1395
  • [35] Reducing Gender Bias in Word-Level Language Models with a Gender-Equalizing Loss Function
    Qian, Yusu
    Muaz, Urwa
    Zhang, Ben
    Hyun, Jae Won
    57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019:): STUDENT RESEARCH WORKSHOP, 2019, : 223 - 228
  • [36] Endorsement by 'looks': measuring gender bias
    Desbordes, Michel
    INTERNATIONAL JOURNAL OF SPORTS MARKETING & SPONSORSHIP, 2015, 16 (04): : 239 - 240
  • [37] Measuring Gender Bias in News Images
    Jia, Sen
    Lansdall-Welfare, Thomas
    Cristianini, Nello
    WWW'15 COMPANION: PROCEEDINGS OF THE 24TH INTERNATIONAL CONFERENCE ON WORLD WIDE WEB, 2015, : 893 - 898
  • [38] A Transparent Framework for Evaluating Unintended Demographic Bias in Word Embeddings
    Sweeney, Chris
    Najafian, Maryam
    57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), 2019, : 1662 - 1667
  • [39] Towards Detection of Subjective Bias using Contextualized Word Embeddings
    Pant, Kartikey
    Dadu, Tanvi
    Mamidi, Radhika
    WWW'20: COMPANION PROCEEDINGS OF THE WEB CONFERENCE 2020, 2020, : 75 - 76
  • [40] Debiasing Embeddings for Reduced Gender Bias in Text Classification
    Prost, Flavien
    Thain, Nithum
    Bolukbasi, Tolga
    GENDER BIAS IN NATURAL LANGUAGE PROCESSING (GEBNLP 2019), 2019, : 69 - 75