Emergent analogical reasoning in large language models

被引:0
|
作者
Taylor Webb
Keith J. Holyoak
Hongjing Lu
机构
[1] University of California,Department of Psychology
[2] University of California,Department of Statistics
来源
Nature Human Behaviour | 2023年 / 7卷
关键词
D O I
暂无
中图分类号
学科分类号
摘要
The recent advent of large language models has reinvigorated debate over whether human cognitive capacities might emerge in such generic models given sufficient training data. Of particular interest is the ability of these models to reason about novel problems zero-shot, without any direct training. In human cognition, this capacity is closely tied to an ability to reason by analogy. Here we performed a direct comparison between human reasoners and a large language model (the text-davinci-003 variant of Generative Pre-trained Transformer (GPT)-3) on a range of analogical tasks, including a non-visual matrix reasoning task based on the rule structure of Raven’s Standard Progressive Matrices. We found that GPT-3 displayed a surprisingly strong capacity for abstract pattern induction, matching or even surpassing human capabilities in most settings; preliminary tests of GPT-4 indicated even better performance. Our results indicate that large language models such as GPT-3 have acquired an emergent ability to find zero-shot solutions to a broad range of analogy problems.
引用
收藏
页码:1526 / 1541
页数:15
相关论文
共 50 条
  • [1] Emergent analogical reasoning in large language models
    Webb, Taylor
    Holyoak, Keith J.
    Lu, Hongjing
    NATURE HUMAN BEHAVIOUR, 2023, 7 (09) : 1526 - 1541
  • [2] Assessing the Emergent Symbolic Reasoning Abilities of Llama Large Language Models
    Petruzzellis, Flavio
    Testolin, Alberto
    Sperduti, Alessandro
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING-ICANN 2024, PT V, 2024, 15020 : 266 - 276
  • [3] In-Context Analogical Reasoning with Pre-Trained Language Models
    Hu, Xiaoyang
    Storks, Shane
    Lewis, Richard L.
    Chai, Joyce
    PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 1, 2023, : 1953 - 1969
  • [4] Large Language Models Are Reasoning Teachers
    Ho, Namgyu
    Schmid, Laura
    Yun, Se-Young
    PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023): LONG PAPERS, VOL 1, 2023, : 14852 - 14882
  • [5] Emergent and Predictable Memorization in Large Language Models
    Biderman, Stella
    Prashanth, U. S. V. S. N. Sai
    Sutawika, Lintang
    Schoelkopf, Hailey
    Anthony, Quentin
    Purohit, Shivanshu
    Raff, Edward
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [6] Large-Scale Analogical Reasoning
    Chaudhri, Vinay K.
    Heymans, Stijn
    Spaulding, Aaron
    Overholtzer, Adam
    Wessel, Michael
    PROCEEDINGS OF THE TWENTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2014, : 359 - 365
  • [7] Analogical Reasoning for Natural to Formal Language Transfer
    Letard, Vincent
    Rosset, Sophie
    Illouz, Gabriel
    2015 IEEE 27TH INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE (ICTAI 2015), 2015, : 210 - 217
  • [8] Analogical reasoning in children with specific language impairment
    Leroy, Sandrine
    Parisse, Christophe
    Maillart, Christelle
    CLINICAL LINGUISTICS & PHONETICS, 2012, 26 (04) : 380 - 395
  • [9] Towards Reasoning in Large Language Models: A Survey
    Huang, Jie
    Chang, Kevin Chen-Chuan
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, 2023, : 1049 - 1065
  • [10] Conversations on reasoning: Large language models in diagnosis
    Restrepo, Daniel
    Rodman, Adam
    Abdulnour, Raja-Elie
    JOURNAL OF HOSPITAL MEDICINE, 2024, 19 (08) : 731 - 735