Do Large Language Models Show Decision Heuristics Similar to Humans? A Case Study Using GPT-3.5

被引:13
|
作者
Suri, Gaurav [1 ]
Slater, Lily R. [1 ]
Ziaee, Ali [1 ]
Nguyen, Morgan [1 ]
机构
[1] San Francisco State Univ, Dept Psychol, Mind Brain & Behav, 1600 Holloway Ave, San Francisco, CA 94132 USA
关键词
natural language processing; Large Language Models; ChatGPT; heuristics; PHYSICIANS; JUDGMENT; CHOICE;
D O I
10.1037/xge0001547
中图分类号
B84 [心理学];
学科分类号
04 ; 0402 ;
摘要
A Large Language Model (LLM) is an artificial intelligence system trained on vast amounts of natural language data, enabling it to generate human-like responses to written or spoken language input. Generative Pre-Trained Transformer (GPT)-3.5 is an example of an LLM that supports a conversational agent called ChatGPT. In this work, we used a series of novel prompts to determine whether ChatGPT shows heuristics and other context-sensitive responses. We also tested the same prompts on human participants. Across four studies, we found that ChatGPT was influenced by random anchors in making estimates (anchoring, Study 1); it judged the likelihood of two events occurring together to be higher than the likelihood of either event occurring alone, and it was influenced by anecdotal information (representativeness and availability heuristic, Study 2); it found an item to be more efficacious when its features were presented positively rather than negatively-even though both presentations contained statistically equivalent information (framing effect, Study 3); and it valued an owned item more than a newly found item even though the two items were objectively identical (endowment effect, Study 4). In each study, human participants showed similar effects. Heuristics and context-sensitive responses in humans are thought to be driven by cognitive and affective processes such as loss aversion and effort reduction. The fact that an LLM-which lacks these processes-also shows such responses invites consideration of the possibility that language is sufficiently rich to carry these effects and may play a role in generating these effects in humans.
引用
收藏
页码:1066 / 1075
页数:10
相关论文
共 31 条
  • [31] Improving entity recognition using ensembles of deep learning and fine-tuned large language models: A case study on adverse event extraction from VAERS and social media
    Li, Yiming
    Viswaroopan, Deepthi
    He, William
    Li, Jianfu
    Zuo, Xu
    Xu, Hua
    Tao, Cui
    JOURNAL OF BIOMEDICAL INFORMATICS, 2025, 163