Emergent analogical reasoning in large language models

被引:0
|
作者
Taylor Webb
Keith J. Holyoak
Hongjing Lu
机构
[1] University of California,Department of Psychology
[2] University of California,Department of Statistics
来源
Nature Human Behaviour | 2023年 / 7卷
关键词
D O I
暂无
中图分类号
学科分类号
摘要
The recent advent of large language models has reinvigorated debate over whether human cognitive capacities might emerge in such generic models given sufficient training data. Of particular interest is the ability of these models to reason about novel problems zero-shot, without any direct training. In human cognition, this capacity is closely tied to an ability to reason by analogy. Here we performed a direct comparison between human reasoners and a large language model (the text-davinci-003 variant of Generative Pre-trained Transformer (GPT)-3) on a range of analogical tasks, including a non-visual matrix reasoning task based on the rule structure of Raven’s Standard Progressive Matrices. We found that GPT-3 displayed a surprisingly strong capacity for abstract pattern induction, matching or even surpassing human capabilities in most settings; preliminary tests of GPT-4 indicated even better performance. Our results indicate that large language models such as GPT-3 have acquired an emergent ability to find zero-shot solutions to a broad range of analogy problems.
引用
收藏
页码:1526 / 1541
页数:15
相关论文
共 50 条
  • [1] Emergent analogical reasoning in large language models
    Webb, Taylor
    Holyoak, Keith J.
    Lu, Hongjing
    [J]. NATURE HUMAN BEHAVIOUR, 2023, 7 (09) : 1526 - 1541
  • [2] In-Context Analogical Reasoning with Pre-Trained Language Models
    Hu, Xiaoyang
    Storks, Shane
    Lewis, Richard L.
    Chai, Joyce
    [J]. PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 1, 2023, : 1953 - 1969
  • [3] Large Language Models Are Reasoning Teachers
    Ho, Namgyu
    Schmid, Laura
    Yun, Se-Young
    [J]. PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023): LONG PAPERS, VOL 1, 2023, : 14852 - 14882
  • [4] Emergent and Predictable Memorization in Large Language Models
    Biderman, Stella
    Prashanth, U. S. V. S. N. Sai
    Sutawika, Lintang
    Schoelkopf, Hailey
    Anthony, Quentin
    Purohit, Shivanshu
    Raff, Edward
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [5] Large-Scale Analogical Reasoning
    Chaudhri, Vinay K.
    Heymans, Stijn
    Spaulding, Aaron
    Overholtzer, Adam
    Wessel, Michael
    [J]. PROCEEDINGS OF THE TWENTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2014, : 359 - 365
  • [6] Analogical reasoning in children with specific language impairment
    Leroy, Sandrine
    Parisse, Christophe
    Maillart, Christelle
    [J]. CLINICAL LINGUISTICS & PHONETICS, 2012, 26 (04) : 380 - 395
  • [7] Analogical Reasoning for Natural to Formal Language Transfer
    Letard, Vincent
    Rosset, Sophie
    Illouz, Gabriel
    [J]. 2015 IEEE 27TH INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE (ICTAI 2015), 2015, : 210 - 217
  • [8] Conversations on reasoning: Large language models in diagnosis
    Restrepo, Daniel
    Rodman, Adam
    Abdulnour, Raja-Elie
    [J]. JOURNAL OF HOSPITAL MEDICINE, 2024, 19 (08) : 731 - 735
  • [9] Inductive reasoning in humans and large language models
    Han, Simon Jerome
    Ransom, Keith J.
    Perfors, Andrew
    Kemp, Charles
    [J]. COGNITIVE SYSTEMS RESEARCH, 2024, 83
  • [10] Large Language Models are Visual Reasoning Coordinators
    Chen, Liangyu
    Li, Bo
    Shen, Sheng
    Yang, Jingkang
    Li, Chunyuan
    Keutzer, Kurt
    Darrell, Trevor
    Liu, Ziwei
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,