Exploring the reversal curse and other deductive logical reasoning in BERT and GPT-based large language models

被引:0
|
作者
Wu, Da [1 ,2 ]
Yang, Jingye [1 ,2 ]
Wang, Kai [1 ,3 ]
机构
[1] Childrens Hosp Philadelphia, Raymond G Perelman Ctr Cellular & Mol Therapeut, Philadelphia, PA 19104 USA
[2] Univ Penn, Dept Math, Philadelphia, PA 19104 USA
[3] Univ Penn, Dept Pathol & Lab Med, Philadelphia, PA 19104 USA
来源
PATTERNS | 2024年 / 5卷 / 09期
关键词
BACKWARD RECALL;
D O I
10.1016/j.patter.2024.101030
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The "Reversal Curse"describes the inability of autoregressive decoder large language models (LLMs) to deduce "B is A"from "A is B,"assuming that B and A are distinct and can be uniquely identified from each other. This logical failure suggests limitations in using generative pretrained transformer (GPT) models for tasks like constructing knowledge graphs. Our study revealed that a bidirectional LLM, bidirectional encoder representations from transformers (BERT), does not suffer from this issue. To investigate further, we focused on more complex deductive reasoning by training encoder and decoder LLMs to perform union and intersection operations on sets. While both types of models managed tasks involving two sets, they struggled with operations involving three sets. Our findings underscore the differences between encoder and decoder models in handling logical reasoning. Thus, selecting BERT or GPT should depend on the task's specific needs, utilizing BERT's bidirectional context comprehension or GPT's sequence prediction strengths.
引用
收藏
页数:12
相关论文
共 26 条
  • [1] Enhancing Neural Decoding with Large Language Models: A GPT-Based Approach
    Lee, Dong Hyeok
    Chung, Chun Kee
    2024 12TH INTERNATIONAL WINTER CONFERENCE ON BRAIN-COMPUTER INTERFACE, BCI 2024, 2024,
  • [2] Exploring a GPT-based large language model for variable autonomy in a VR-based human-robot teaming simulation
    Lakhnati, Younes
    Pascher, Max
    Gerken, Jens
    FRONTIERS IN ROBOTICS AND AI, 2024, 11
  • [3] Testing the General Deductive Reasoning Capacity of Large Language Models Using OOD Examples
    Saparov, Abulhair
    Pang, Richard Yuanzhe
    Padmakumar, Vishakh
    Joshi, Nitish
    Kazemi, Seyed Mehran
    Kim, Najoung
    He, He
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [4] ChatGPT: Where Is a Silver Lining? Exploring the realm of GPT and large language models
    Tikhonova, Elena
    Raitskaya, Lilia
    JOURNAL OF LANGUAGE AND EDUCATION, 2023, 9 (03): : 5 - 11
  • [5] Case-Based Reasoning with Language Models for Classification of Logical Fallacies
    Sourati, Zhivar
    Ilievski, Filip
    Sandlin, Hong-An
    Mermoud, Alain
    PROCEEDINGS OF THE THIRTY-SECOND INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2023, 2023, : 5188 - 5196
  • [6] A Loosely Wittgensteinian Conception of the Linguistic Understanding of Large Language Models like BERT, GPT-3, and ChatGPT
    Gubelmann, Reto
    GRAZER PHILOSOPHISCHE STUDIEN-INTERNATIONAL JOURNAL FOR ANALYTIC PHILOSOPHY, 2023, 99 (04): : 485 - 523
  • [7] From BERT to GPT-3 Codex: Harnessing the Potential of Very Large Language Models for Data Management
    Trummer, Immanuel
    PROCEEDINGS OF THE VLDB ENDOWMENT, 2022, 15 (12): : 3770 - 3773
  • [8] Exploring the capabilities of large language models for the generation of safety cases: the case of GPT-4
    Sivakumar, Mithila
    Belle, Alvine Boaye
    Shan, Jinjun
    Shahandashti, Kimya Khakzad
    32ND INTERNATIONAL REQUIREMENTS ENGINEERING CONFERENCE WORKSHOPS, REW 2024, 2024, : 35 - 45
  • [9] ChatGPT, GPT-4, and Other Large Language Models: The Next Revolution for Clinical Microbiology?
    Egli, Adrian
    CLINICAL INFECTIOUS DISEASES, 2023, 77 (09) : 1322 - 1328
  • [10] On Implementing Case-Based Reasoning with Large Language Models
    Wilkerson, Kaitlynne
    Leake, David
    CASE-BASED REASONING RESEARCH AND DEVELOPMENT, ICCBR 2024, 2024, 14775 : 404 - 417