Rationality of Thought Improves Reasoning in Large Language Models

被引:0
|
作者
Gou, Tian [1 ,2 ]
Zhang, Boyao [1 ,2 ]
Sun, Zhenglie [1 ,2 ]
Wang, Jing [1 ,2 ]
Liu, Fang [1 ,2 ]
Wang, Yangang [1 ,2 ]
Wang, Jue [1 ,2 ]
机构
[1] Chinese Acad Sci, Comp Network Informat Ctr, Beijing, Peoples R China
[2] Univ Chinese Acad Sci, Beijing, Peoples R China
基金
国家重点研发计划; 北京市自然科学基金;
关键词
Large Language Models (LLMs); Zero-Shot Reasoning; Cognitive foundations of knowledge; Rationality of Thought (RoT); Cognitive Psychology; Cognitive Bias Dataset; HEURISTICS; FALLACY;
D O I
10.1007/978-981-97-5501-1_26
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
While the capabilities of large language models (LLMs) have been progressively advanced, their competence in addressing intricate reasoning tasks remains inadequate, primarily due to their insufficient cognitive capabilities. To explore the cognitive proficiency of models like GPT-4, we turn to methodologies from cognitive psychology: cognitive abilities reflect rational thinking skills, and cognitive bias tasks are often used to assess rational thinking levels. In this paper, we develop a cognitive bias dataset to measure the rational thinking and cognitive levels of LLMs. Our observations indicate that GPT-4, akin to humans, exhibits limitations in its rational thinking ability. We propose a new method, "Rationality of Thought" (RoT), to prompt LLMs into a rational thinking process during task execution. This method significantly improves the accuracy of GPT-4 on the cognitive bias task by 18.7%. Cognitive capacity is also essential for tackling complex issues, therefore, we implement RoT across various reasoning tasks. Using only a zero-shot setting, RoT outperforms inference enhancement techniques such as CoT using zero-shot, such as SVAMP(+1.8),AQUA-RAT (+6.0), ARC-c (+4.1),ARCe(+3.9) in multiple arithmetic and common sense reasoning tasks. Our empirical evaluation shows that RoT helps LLMs elevate their cognitive capabilities through rational thinking, thereby becoming more adept at navigating complex reasoning tasks.
引用
收藏
页码:343 / 358
页数:16
相关论文
共 50 条
  • [1] Chain-of-Thought Prompting Elicits Reasoning in Large Language Models
    Wei, Jason
    Wang, Xuezhi
    Schuurmans, Dale
    Bosma, Maarten
    Ichter, Brian
    Xia, Fei
    Chi, Ed H.
    Le, Quoc V.
    Zhou, Denny
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [2] Chain-of-Thought Improves Text Generation with Citations in Large Language Models
    Ji, Bin
    Liu, Huijun
    Du, Mingzhe
    Ng, See-Kiong
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 16, 2024, : 18345 - 18353
  • [3] Dissociating language and thought in large language models
    Mahowald, Kyle
    Ivanova, Anna A.
    Blank, Idan A.
    Kanwisher, Nancy
    Tenenbaum, Joshua B.
    Fedorenko, Evelina
    TRENDS IN COGNITIVE SCIENCES, 2024, 28 (06) : 517 - 540
  • [4] Large Language Models Are Reasoning Teachers
    Ho, Namgyu
    Schmid, Laura
    Yun, Se-Young
    PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023): LONG PAPERS, VOL 1, 2023, : 14852 - 14882
  • [5] Chain-of-Thought Reasoning in Tabular Language Models
    Zheng, Mingyu
    Hao, Yang
    Jiang, Wenbin
    Lin, Zheng
    Lyu, Yajuan
    She, Qiaoqiao
    Wang, Weiping
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EMNLP 2023), 2023, : 11006 - 11019
  • [6] Moderating New Waves of Online Hate with Chain-of-Thought Reasoning in Large Language Models
    Vishwamitra, Nishant
    Guo, Keyan
    Romit, Farhan Tajwar
    Ondracek, Isabelle
    Cheng, Long
    Zhao, Ziming
    Hu, Hongxin
    45TH IEEE SYMPOSIUM ON SECURITY AND PRIVACY, SP 2024, 2024, : 788 - 806
  • [7] (Ir)rationality and cognitive biases in large language models
    Macmillan-Scott, Olivia
    Musolesi, Mirco
    ROYAL SOCIETY OPEN SCIENCE, 2024, 11 (06):
  • [8] Towards Reasoning in Large Language Models: A Survey
    Huang, Jie
    Chang, Kevin Chen-Chuan
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, 2023, : 1049 - 1065
  • [9] Conversations on reasoning: Large language models in diagnosis
    Restrepo, Daniel
    Rodman, Adam
    Abdulnour, Raja-Elie
    JOURNAL OF HOSPITAL MEDICINE, 2024, 19 (08) : 731 - 735
  • [10] Emergent analogical reasoning in large language models
    Taylor Webb
    Keith J. Holyoak
    Hongjing Lu
    Nature Human Behaviour, 2023, 7 : 1526 - 1541