Large Language Models Are Neurosymbolic Reasoners

被引:0
|
作者
Fang, Meng [1 ,2 ]
Deng, Shilong [1 ]
Zhang, Yudi [2 ]
Shi, Zijing [3 ]
Chen, Ling [3 ]
Pechenizkiy, Mykola [2 ]
Wang, Jun [4 ]
机构
[1] Univ Liverpool, Liverpool, Merseyside, England
[2] Eindhoven Univ Technol, Eindhoven, Netherlands
[3] Univ Technol Sydney, Sydney, NSW, Australia
[4] UCL, London, England
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
A wide range of real-world applications is characterized by their symbolic nature, necessitating a strong capability for symbolic reasoning. This paper investigates the potential application of Large Language Models (LLMs) as symbolic reasoners. We focus on text-based games, significant benchmarks for agents with natural language capabilities, particularly in symbolic tasks like math, map reading, sorting, and applying common sense in text-based worlds. To facilitate these agents, we propose an LLM agent designed to tackle symbolic challenges and achieve in-game objectives. We begin by initializing the LLM agent and informing it of its role. The agent then receives observations and a set of valid actions from the text-based games, along with a specific symbolic module. With these inputs, the LLM agent chooses an action and interacts with the game environments. Our experimental results demonstrate that our method significantly enhances the capability of LLMs as automated agents for symbolic reasoning, and our LLM agent is effective in text-based games involving symbolic tasks, achieving an average performance of 88% across all tasks.
引用
收藏
页码:17985 / 17993
页数:9
相关论文
共 50 条
  • [1] Large Language Models Are Not Strong Abstract Reasoners
    Gendron, Gael
    Bao, Qiming
    Witbrock, Michael
    Dobbie, Gillian
    PROCEEDINGS OF THE THIRTY-THIRD INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2024, 2024, : 6270 - 6278
  • [2] Neurosymbolic AI Approach to Attribution in Large Language Models
    Tilwani, Deepa
    Venkataramanan, Revathy
    Sheth, Amit P.
    IEEE INTELLIGENT SYSTEMS, 2024, 39 (06) : 10 - 17
  • [3] Large Language Models are Zero-Shot Reasoners
    Kojima, Takeshi
    Gu, Shixiang Shane
    Reid, Machel
    Matsuo, Yutaka
    Iwasawa, Yusuke
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [4] Behavior Cloned Transformers are Neurosymbolic Reasoners
    Wang, Ruoyao
    Jansen, Peter
    Cote, Marc-Alexandre
    Ammanabrolu, Prithviraj
    17TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EACL 2023, 2023, : 2777 - 2788
  • [5] Large Language Models are few(1)-shot Table Reasoners
    Chen, Wenhu
    17TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EACL 2023, 2023, : 1120 - 1130
  • [6] Large Language Models are Better Reasoners with Self-Verification
    Weng, Yixuan
    Zhu, Minjun
    Xia, Fei
    Li, Bin
    He, Shizhu
    Liu, Shengping
    Sun, Bin
    Liu, Kang
    Zhao, Jun
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS - EMNLP 2023, 2023, : 2550 - 2575
  • [7] A framework for neurosymbolic robot action planning using large language models
    Capitanelli, Alessio
    Mastrogiovanni, Fulvio
    FRONTIERS IN NEUROROBOTICS, 2024, 18
  • [8] Large Language Models are Temporal and Causal Reasoners for Video Question Answering
    Ko, Dohwan
    Lee, Ji Soo
    Kang, Wooyoung
    Roh, Byungseok
    Kim, Hyunwoo J.
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING, EMNLP 2023, 2023, : 4300 - 4316
  • [9] Making Large Language Models Better Reasoners with Step-Aware Verifier
    Li, Yifei
    Lin, Zeqi
    Zhang, Shizhuo
    Fu, Qiang
    Chen, Bei
    Lou, Jian-Guang
    Chen, Weizhu
    PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 1, 2023, : 5315 - 5333
  • [10] Are Large Language Models Really Good Logical Reasoners? A Comprehensive Evaluation and Beyond
    Xu, Fangzhi
    Lin, Qika
    Han, Jiawei
    Zhao, Tianzhe
    Liu, Jun
    Cambria, Erik
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2025, 37 (04) : 1620 - 1634