Measuring abstract reasoning in neural networks

被引:0
|
作者
Barrett, David G. T. [1 ]
Hill, Felix [1 ]
Santoro, Adam [1 ]
Morcos, Ari S. [1 ]
Lillicrap, Timothy [1 ]
机构
[1] DeepMind, London, England
关键词
IQ TESTS; INTELLIGENCE;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Whether neural networks can learn abstract reasoning or whether they merely rely on superficial statistics is a topic of recent debate. Here, we propose a dataset and challenge designed to probe abstract reasoning, inspired by a well-known human IQ test. To succeed at this challenge, models must cope with various generalisation 'regimes' in which the training and test data differ in clearly-defined ways. We show that popular models such as ResNets perform poorly, even when the training and test sets differ only minimally, and we present a novel architecture, with a structure designed to encourage reasoning, that does significantly better. When we vary the way in which the test questions and training data differ, we find that our model is notably proficient at certain forms of generalisation, but notably weak at others. We further show that the model's ability to generalise improves markedly if it is trained to predict symbolic explanations for its answers. Altogether, we introduce and explore ways to both measure and induce stronger abstract reasoning in neural networks. Our freely-available dataset should motivate further progress in this direction.
引用
收藏
页数:10
相关论文
共 50 条
  • [41] Longitudinal changes in the neural oscillatory dynamics underlying abstract reasoning in children and adolescents
    Taylor, Brittany K.
    Heinrichs-Grahama, Elizabeth
    Eastman, Jacob A.
    Frenzel, Michaela R.
    Wang, Yu-Ping
    Calhoun, Vince D.
    Stephen, Julia M.
    Wilsona, Tony W.
    NEUROIMAGE, 2022, 253
  • [42] Abstract Interpretation of Fixpoint Iterators with Applications to Neural Networks
    Mueller, Mark Niklas
    Fischer, Marc
    Staab, Robin
    Vechev, Martin
    PROCEEDINGS OF THE ACM ON PROGRAMMING LANGUAGES-PACMPL, 2023, 7 (PLDI):
  • [43] Measuring Catastrophic Forgetting in Neural Networks
    Kemker, Ronald
    McClure, Marc
    Abitino, Angelina
    Hayes, Tyler L.
    Kanan, Christopher
    THIRTY-SECOND AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTIETH INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / EIGHTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2018, : 3390 - 3398
  • [44] On Measuring Excess Capacity in Neural Networks
    Graf, Florian
    Zeng, Sebastian
    Rieck, Bastian
    Niethammer, Marc
    Kwitt, Roland
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [45] ARTIFICIAL NEURAL NETWORKS AS MEASURING DEVICES
    EIDE, A
    LINDBLAD, T
    NUCLEAR INSTRUMENTS & METHODS IN PHYSICS RESEARCH SECTION A-ACCELERATORS SPECTROMETERS DETECTORS AND ASSOCIATED EQUIPMENT, 1992, 317 (03): : 607 - 608
  • [46] Differentiable Abstract Interpretation for Provably Robust Neural Networks
    Mirman, Matthew
    Gehr, Timon
    Vechev, Martin
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 80, 2018, 80
  • [47] Morphogenic neural networks encode abstract rules by data
    Resconi, G
    van der Wal, AJ
    INFORMATION SCIENCES, 2002, 142 (1-4) : 249 - 273
  • [48] Abstract Rule Based Pattern Learning with Neural Networks
    Kopparti, Radha
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 13718 - 13719
  • [49] Deep Residual Networks’ Abstract Reasoning Performance on Raven’s Progressive Matrices
    Wang S.
    SN Computer Science, 2021, 2 (5)
  • [50] NEURAL NETWORKS DESIGNED ON APPROXIMATE REASONING ARCHITECTURE AND THEIR APPLICATIONS
    TAKAGI, H
    SUZUKI, N
    KODA, T
    KOJIMA, Y
    IEEE TRANSACTIONS ON NEURAL NETWORKS, 1992, 3 (05): : 752 - 760