Performance and biases of Large Language Models in public opinion simulation

被引:3
|
作者
Qu, Yao [1 ]
Wang, Jue [1 ]
机构
[1] Nanyang Technol Univ, Sch Social Sci, Singapore, Singapore
来源
关键词
D O I
10.1057/s41599-024-03609-x
中图分类号
C [社会科学总论];
学科分类号
03 ; 0303 ;
摘要
The rise of Large Language Models (LLMs) like ChatGPT marks a pivotal advancement in artificial intelligence, reshaping the landscape of data analysis and processing. By simulating public opinion, ChatGPT shows promise in facilitating public policy development. However, challenges persist regarding its worldwide applicability and bias across demographics and themes. Our research employs socio-demographic data from the World Values Survey to evaluate ChatGPT's performance in diverse contexts. Findings indicate significant performance disparities, especially when comparing countries. Models perform better in Western, English-speaking, and developed nations, notably the United States, in comparison to others. Disparities also manifest across demographic groups, showing biases related to gender, ethnicity, age, education, and social class. The study further uncovers thematic biases in political and environmental simulations. These results highlight the need to enhance LLMs' representativeness and address biases, ensuring their equitable and effective integration into public opinion research alongside conventional methodologies.
引用
收藏
页数:13
相关论文
共 50 条
  • [41] Mitigating Cognitive Biases in Clinical Decision-Making Through Multi-Agent Conversations Using Large Language Models: Simulation Study
    Ke, Yuhe
    Yang, Rui
    Lie, Sui An
    Lim, Taylor Xin Yi
    Ning, Yilin
    Li, Irene
    Abdullah, Hairil Rizal
    Ting, Daniel Shu Wei
    Liu, Nan
    JOURNAL OF MEDICAL INTERNET RESEARCH, 2024, 26
  • [42] Evaluating the Performance of Large Language Models for Spanish Language in Undergraduate Admissions Exams
    Miranda, Sabino
    Pichardo-Lagunas, Obdulia
    Martinez-Seis, Bella
    Baldi, Pierre
    COMPUTACION Y SISTEMAS, 2023, 27 (04): : 1241 - 1248
  • [43] PUBLIC OPINION IN TATARSTAN ON LANGUAGE DEVELOPMENTS AND BILINGUALISM
    Nizamova, Liliya Raviliyevna
    REVISTA ENTRELINGUAS, 2021, 7
  • [44] Large language models and humans converge in judging public figures' personalities
    Cao, Xubo
    Kosinski, Michal
    PNAS NEXUS, 2024, 3 (10):
  • [45] Exciting Opportunities and Necessary Safeguards for Large Language Models in the Public Sector
    Biegelbauer, Peter
    Schindler, Alexander
    Conde-Jimenez, Rodrigo
    Weinlinger, Pia
    ERCIM NEWS, 2024, (136): : 25 - 26
  • [46] Adaptation of Large Language Models for the Public Sector: A Clustering Use Case
    Caudron, Emilien
    Ghesquiere, Nathan
    Travers, Wouter
    Balahur, Alexandra
    NATURAL LANGUAGE PROCESSING AND INFORMATION SYSTEMS, PT II, NLDB 2024, 2024, 14763 : 327 - 340
  • [47] Performance of Three Large Language Models on Dermatology Board Examinations
    Mirza, Fatima N.
    Lim, Rachel K.
    Yumeen, Sara
    Wahood, Samer
    Zaidat, Bashar
    Shah, Asghar
    Tang, Oliver Y.
    Kawaoka, John
    Seo, Su-Jean
    Dimarco, Christopher
    Muglia, Jennie
    Goldbach, Hayley S.
    Wisco, Oliver
    Qureshi, Abrar A.
    Libby, Tiffany J.
    JOURNAL OF INVESTIGATIVE DERMATOLOGY, 2024, 144 (02) : 398 - 400
  • [48] Performance of Large Language Models in a Computer Science Degree Program
    Krueger, Tim
    Gref, Michael
    ARTIFICIAL INTELLIGENCE-ECAI 2023 INTERNATIONAL WORKSHOPS, PT 2, XAI3, TACTIFUL, XI-ML, SEDAMI, RAAIT, AI4S, HYDRA, AI4AI, 2023, 2024, 1948 : 409 - 424
  • [49] Large Language Models Performance Comparison of Emotion and Sentiment Classification
    Stigall, William
    Khan, Md Abdullah Al Hafiz
    Attota, Dinesh
    Nweke, Francis
    Pei, Yong
    PROCEEDINGS OF THE 2024 ACM SOUTHEAST CONFERENCE, ACMSE 2024, 2024, : 60 - 68
  • [50] Performance of large language models in oral and maxillofacial surgery examinations
    Quah, B.
    Yong, C. W.
    Lai, C. W. M.
    Islam, I.
    INTERNATIONAL JOURNAL OF ORAL AND MAXILLOFACIAL SURGERY, 2024, 53 (10) : 881 - 886