Exploring the reversal curse and other deductive logical reasoning in BERT and GPT-based large language models

被引:0
|
作者
Wu, Da [1 ,2 ]
Yang, Jingye [1 ,2 ]
Wang, Kai [1 ,3 ]
机构
[1] Childrens Hosp Philadelphia, Raymond G Perelman Ctr Cellular & Mol Therapeut, Philadelphia, PA 19104 USA
[2] Univ Penn, Dept Math, Philadelphia, PA 19104 USA
[3] Univ Penn, Dept Pathol & Lab Med, Philadelphia, PA 19104 USA
来源
PATTERNS | 2024年 / 5卷 / 09期
关键词
BACKWARD RECALL;
D O I
10.1016/j.patter.2024.101030
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The "Reversal Curse"describes the inability of autoregressive decoder large language models (LLMs) to deduce "B is A"from "A is B,"assuming that B and A are distinct and can be uniquely identified from each other. This logical failure suggests limitations in using generative pretrained transformer (GPT) models for tasks like constructing knowledge graphs. Our study revealed that a bidirectional LLM, bidirectional encoder representations from transformers (BERT), does not suffer from this issue. To investigate further, we focused on more complex deductive reasoning by training encoder and decoder LLMs to perform union and intersection operations on sets. While both types of models managed tasks involving two sets, they struggled with operations involving three sets. Our findings underscore the differences between encoder and decoder models in handling logical reasoning. Thus, selecting BERT or GPT should depend on the task's specific needs, utilizing BERT's bidirectional context comprehension or GPT's sequence prediction strengths.
引用
收藏
页数:12
相关论文
共 49 条
  • [21] Chain of Logic: Rule-Based Reasoning with Large Language Models
    Servantez, Sergio
    Barrow, Joe
    Hammond, Kristian
    Jain, Rajiv
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 2721 - 2733
  • [22] Exploring Reasoning Biases in Large Language Models Through Syllogism: Insights from the NeuBAROCO Dataset
    Ozeki, Kentaro
    Ando, Risako
    Morishita, Takanobu
    Abe, Hirohiko
    Mineshima, Koji
    Okada, Mitsuhiro
    arXiv,
  • [23] Exploring Reasoning Biases in Large Language Models Through Syllogism: Insights from the NeuBAROCO Dataset
    Ozeki, Kentaro
    Ando, Risako
    Morishita, Takanobu
    Abe, Hirohiko
    Mineshima, Koji
    Okada, Mitsuhiro
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 16063 - 16077
  • [24] MultiSurf-GPT: Facilitating Context-Aware Reasoning with Large-Scale Language Models for Multimodal Surface Sensing
    Hu, Yongquan
    Sun, Black
    An, Pengcheng
    Li, Zhuying
    Hu, Wen
    Quigley, Aaron J.
    PUBLICATION OF THE 26TH ACM INTERNATIONAL CONFERENCE ON MOBILE HUMAN-COMPUTER INTERACTION, MOBILEHCI 2024 ADJUNCT PROCEEDINGS, 2024,
  • [25] Exploring the use of large language models (LLMs) in chemical engineering education: Building core course problem models with Chat-GPT
    Tsai, Meng -Lin
    Ong, Chong Wei
    Chen, Cheng-Liang
    EDUCATION FOR CHEMICAL ENGINEERS, 2023, 44 : 71 - 95
  • [26] A Novel Approach for Machine Reading Comprehension using BERT-based Large Language Models
    Varghese, Nisha
    Shereef, Shafi
    Joy, Helen K.
    Ramasamy, Gobi
    Sridevi, R.
    Cynthia, T.
    Rajeshkanna, R.
    10TH INTERNATIONAL CONFERENCE ON ELECTRONICS, COMPUTING AND COMMUNICATION TECHNOLOGIES, CONECCT 2024, 2024,
  • [27] Exploring the potential utility of AI large language models for medical ethics: an expert panel evaluation of GPT-4
    Balas, Michael
    Wadden, Jordan Joseph
    Hebert, Philip C.
    Mathison, Eric
    Warren, Marika D.
    Seavilleklein, Victoria
    Wyzynski, Daniel
    Callahan, Alison
    Crawford, Sean A.
    Arjmand, Parnian
    Ing, Edsel B.
    JOURNAL OF MEDICAL ETHICS, 2024, 50 (02) : 90 - 96
  • [28] Triplet-based contrastive method enhances the reasoning ability of large language models
    Chen, Hongwei
    Zhu, Jiahui
    Wang, Wei
    Zhu, Yuan
    Xi, Liya
    JOURNAL OF SUPERCOMPUTING, 2025, 81 (04):
  • [29] INFORM : Information eNtropy based multi-step reasoning FOR large language Models
    Zhou, Chuyue
    You, Wangjie
    Li, Juntao
    Ye, Jing
    Chen, Kehai
    Zhang, Min
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING, EMNLP 2023, 2023, : 3565 - 3576
  • [30] A Hybrid Approach to Dimensional Aspect-Based Sentiment Analysis Using BERT and Large Language Models
    Zhang, Yice
    Xu, Hongling
    Zhang, Delong
    Xu, Ruifeng
    ELECTRONICS, 2024, 13 (18)