BiasAsker: Measuring the Bias in Conversational AI System

被引:12
|
作者
Wan, Yuxuan [1 ]
Wang, Wenxuan [1 ]
He, Pinjia [2 ]
Gu, Jiazhen [1 ]
Bai, Haonan [1 ]
Lyu, Michael R. [1 ]
机构
[1] Chinese Univ Hong Kong, Hong Kong, Peoples R China
[2] Chinese Univ Hong Kong, Shenzhen CUHK Shenzhen, Sch Data Sci, Shenzhen, Peoples R China
来源
PROCEEDINGS OF THE 31ST ACM JOINT MEETING EUROPEAN SOFTWARE ENGINEERING CONFERENCE AND SYMPOSIUM ON THE FOUNDATIONS OF SOFTWARE ENGINEERING, ESEC/FSE 2023 | 2023年
基金
中国国家自然科学基金;
关键词
Software testing; conversational models; social bias;
D O I
10.1145/3611643.3616310
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Powered by advanced Artificial Intelligence (AI) techniques, conversational AI systems, such as ChatGPT, and digital assistants like Siri, have been widely deployed in daily life. However, such systems may still produce content containing biases and stereotypes, causing potential social problems. Due to modern AI techniques' data-driven, black-box nature, comprehensively identifying and measuring biases in conversational systems remains challenging. Particularly, it is hard to generate inputs that can comprehensively trigger potential bias due to the lack of data containing both social groups and biased properties. In addition, modern conversational systems can produce diverse responses (e.g., chatting and explanation), which makes existing bias detection methods based solely on sentiment and toxicity hardly being adopted. In this paper, we propose BiasAsker, an automated framework to identify and measure social bias in conversational AI systems. To obtain social groups and biased properties, we construct a comprehensive social bias dataset containing a total of 841 groups and 5,021 biased properties. Given the dataset, BiasAsker automatically generates questions and adopts a novel method based on existence measurement to identify two types of biases (i.e., absolute bias and related bias) in conversational systems. Extensive experiments on eight commercial systems and two famous research models, such as ChatGPT and GPT-3, show that 32.83% of the questions generated by BiasAsker can trigger biased behaviors in these widely deployed conversational systems. All the code, data, and experimental results have been released to facilitate future research.
引用
收藏
页码:515 / 527
页数:13
相关论文
共 50 条
  • [21] Towards interculturally adaptive conversational AI
    Brandt, Adam
    Hazel, Spencer
    APPLIED LINGUISTICS REVIEW, 2025, 16 (02) : 775 - 786
  • [22] Learning towards conversational AI: A survey
    Fu T.
    Gao S.
    Zhao X.
    Wen J.-R.
    Yan R.
    AI Open, 2022, 3 : 14 - 28
  • [23] An intent recognition pipeline for conversational AI
    Chandrakala C.B.
    Bhardwaj R.
    Pujari C.
    International Journal of Information Technology, 2024, 16 (2) : 731 - 743
  • [24] From general AI to custom AI: the effects of generative conversational AI's cognitive and emotional conversational skills on user's guidance
    Wang, Kun
    Pan, Zhao
    Lu, Yaobin
    KYBERNETES, 2024,
  • [25] The appropriation of conversational AI in the workplace: A taxonomy of AI chatbot users
    Gkinko, Lorentsa
    Elbanna, Amany
    INTERNATIONAL JOURNAL OF INFORMATION MANAGEMENT, 2023, 69
  • [26] Learning towards conversational AI: A survey
    Fu, Tingchen
    Gao, Shen
    Zhao, Xueliang
    Wen, Ji-rong
    Yan, Rui
    AI OPEN, 2022, 3 : 14 - 28
  • [27] Language Ideology Bias in Conversational Technology
    Hohn, Sviatlana
    Migge, Bettina
    Dippold, Doris
    Schneider, Britta
    Mauw, Sjouke
    CHATBOT RESEARCH AND DESIGN, CONVERSATIONS 2023, 2024, 14524 : 133 - 148
  • [28] Teaching Students About Conversational AI Using CONVO, a Conversational Programming Agent
    Zhu, Jessica
    Van Brummelen, Jessica
    2021 IEEE SYMPOSIUM ON VISUAL LANGUAGES AND HUMAN-CENTRIC COMPUTING (VL/HCC 2021), 2021,
  • [29] Mitigation of AI adoption bias through an improved autonomous AI system for diabetic retinal disease
    Abramoff, Michael D.
    Lavin, Philip T.
    Jakubowski, Julie R.
    Blodi, Barbara A.
    Keeys, Mia
    Joyce, Cara
    Folk, James C.
    NPJ DIGITAL MEDICINE, 2024, 7 (01):
  • [30] AI, Bias, and Discrimination
    El-Samad, Hana
    GEN BIOTECHNOLOGY, 2023, 2 (06): : 445 - 445