Exploring the reversal curse and other deductive logical reasoning in BERT and GPT-based large language models

被引:0
|
作者
Wu, Da [1 ,2 ]
Yang, Jingye [1 ,2 ]
Wang, Kai [1 ,3 ]
机构
[1] Childrens Hosp Philadelphia, Raymond G Perelman Ctr Cellular & Mol Therapeut, Philadelphia, PA 19104 USA
[2] Univ Penn, Dept Math, Philadelphia, PA 19104 USA
[3] Univ Penn, Dept Pathol & Lab Med, Philadelphia, PA 19104 USA
来源
PATTERNS | 2024年 / 5卷 / 09期
关键词
BACKWARD RECALL;
D O I
10.1016/j.patter.2024.101030
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The "Reversal Curse"describes the inability of autoregressive decoder large language models (LLMs) to deduce "B is A"from "A is B,"assuming that B and A are distinct and can be uniquely identified from each other. This logical failure suggests limitations in using generative pretrained transformer (GPT) models for tasks like constructing knowledge graphs. Our study revealed that a bidirectional LLM, bidirectional encoder representations from transformers (BERT), does not suffer from this issue. To investigate further, we focused on more complex deductive reasoning by training encoder and decoder LLMs to perform union and intersection operations on sets. While both types of models managed tasks involving two sets, they struggled with operations involving three sets. Our findings underscore the differences between encoder and decoder models in handling logical reasoning. Thus, selecting BERT or GPT should depend on the task's specific needs, utilizing BERT's bidirectional context comprehension or GPT's sequence prediction strengths.
引用
收藏
页数:12
相关论文
共 49 条
  • [31] Exploring the Potential of Large Language Models for Text-Based Personality Prediction
    Molchanova, Maria
    NATURAL LANGUAGE PROCESSING AND INFORMATION SYSTEMS, PT II, NLDB 2024, 2024, 14763 : 291 - 301
  • [32] Exploring EFL university teachers' beliefs in integrating ChatGPT and other large language models in language education: a study in China
    Gao, Yang
    Wang, Qikai
    Wang, Xiaochen
    ASIA PACIFIC JOURNAL OF EDUCATION, 2024, 44 (01) : 29 - 44
  • [33] Large Language Models are Versatile Decomposers: Decomposing Evidence and Questions for Table-based Reasoning
    Ye, Yunhu
    Hui, Binyuan
    Yang, Min
    Li, Binhua
    Huang, Fei
    Li, Yongbin
    PROCEEDINGS OF THE 46TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2023, 2023, : 174 - 184
  • [34] A reasoning based explainable multimodal fake news detection for low resource language using large language models and transformers
    Lekshmiammal, Hariharan RamakrishnaIyer
    Madasamy, Anand Kumar
    JOURNAL OF BIG DATA, 2025, 12 (01)
  • [35] QUERY2TREE: a reasoning model for answering logical queries based on knowledge graph embedding and large language model
    Phan, Truong H. V.
    Do, Phuc
    KNOWLEDGE AND INFORMATION SYSTEMS, 2025,
  • [36] Reasoning about models of context: A context-oriented logical language for knowledge-based context-aware applications
    Schmidtke, Hedda R.
    Hong, Dongpyo
    Woo, Woontack
    Revue d'Intelligence Artificielle, 2008, 22 (05) : 589 - 608
  • [37] Comparative analysis of BERT-based and generative large language models for detecting suicidal ideation: a performance evaluation study
    de Oliveira, Adonias Caetano
    Bessa, Renato Freitas
    Soares, Ariel
    CADERNOS DE SAUDE PUBLICA, 2024, 40 (10):
  • [38] Performance Evaluation of Multimodal Large Language Models (LLaVA and GPT-4-based ChatGPT) in Medical Image Classification Tasks
    Guo, Yuhang
    Wan, Zhiyu
    2024 IEEE 12TH INTERNATIONAL CONFERENCE ON HEALTHCARE INFORMATICS, ICHI 2024, 2024, : 541 - 543
  • [39] Unmasking large language models by means of OpenAI GPT-4 and Google AI: A deep instruction-based analysis
    Zahid, Idrees A.
    Joudar, Shahad Sabbar
    Albahri, A. S.
    Albahri, O. S.
    Alamoodi, A. H.
    Santamaria, Jose
    Alzubaidi, Laith
    INTELLIGENT SYSTEMS WITH APPLICATIONS, 2024, 23
  • [40] SmartEdit: Exploring Complex Instruction-based Image Editing with Multimodal Large Language Models
    Huang, Yuzhou
    Xie, Liangbin
    Wang, Xintao
    Yuan, Ziyang
    Cun, Xiaodong
    Ge, Yixiao
    Zhou, Jiantao
    Dong, Chao
    Huang, Rui
    Zhang, Ruimao
    Shan, Ying
    2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2024, 2024, : 8362 - 8371