The relational processing limits of classic and contemporary neural network models of language processing

被引:2
|
作者
Puebla, Guillermo [1 ,2 ]
Martin, Andrea E. [3 ,4 ]
Doumas, Leonidas A. A. [1 ]
机构
[1] Univ Edinburgh, Sch Philosophy Psychol & Language Sci, Dept Psychol, Edinburgh, Midlothian, Scotland
[2] Univ Tarapaca, Dept Psychol, Arica, Chile
[3] Max Planck Inst Psycholinguist, Language & Computat Neural Syst Grp, Nijmegen, Netherlands
[4] Radboud Univ Nijmegen, Donders Ctr Cognit Neuroimaging, Nijmegen, Netherlands
关键词
Relational reasoning; generalisation; language processing; neural networks; deep learning; REPRESENTATION; ANALOGY;
D O I
10.1080/23273798.2020.1821906
中图分类号
R36 [病理学]; R76 [耳鼻咽喉科学];
学科分类号
100104 ; 100213 ;
摘要
Whether neural networks can capture relational knowledge is a matter of long-standing controversy. Recently, some researchers have argued that (1) classic connectionist models can handle relational structure and (2) the success of deep learning approaches to natural language processing suggests that structured representations are unnecessary to model human language. We tested the Story Gestalt model, a classic connectionist model of text comprehension, and a Sequence-to-Sequence with Attention model, a modern deep learning architecture for natural language processing. Both models were trained to answer questions about stories based on abstract thematic roles. Two simulations varied the statistical structure of new stories while keeping their relational structure intact. The performance of each model fell below chance at least under one manipulation. We argue that both models fail our tests because they can't perform dynamic binding. These results cast doubts on the suitability of traditional neural networks for explaining relational reasoning and language processing phenomena.
引用
收藏
页码:240 / 254
页数:15
相关论文
共 50 条
  • [31] The Application of Graph Neural Network in Natural Language Processing and Computer Vision
    Liu, Xiaochen
    Su, Yang
    Xu, Bingjie
    [J]. 2021 3RD INTERNATIONAL CONFERENCE ON MACHINE LEARNING, BIG DATA AND BUSINESS INTELLIGENCE (MLBDBI 2021), 2021, : 708 - 714
  • [32] A hybrid neural network and virtual reality system for spatial language processing
    Martinez, GC
    Cangelosi, A
    Coventry, KR
    [J]. IJCNN'01: INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, VOLS 1-4, PROCEEDINGS, 2001, : 16 - 21
  • [33] Neural network-based error handler in natural language processing
    S. Anbukkarasi
    S. Varadhaganapathy
    [J]. Neural Computing and Applications, 2022, 34 : 20629 - 20638
  • [34] Bidirectional processing and expanded relational network notation
    Lamb, SM
    [J]. TWENTY-THIRD LACUS FORUM 1996, 1997, : 109 - 124
  • [35] Electrophysiological models of neural processing
    Nelson, Mark E.
    [J]. WILEY INTERDISCIPLINARY REVIEWS-SYSTEMS BIOLOGY AND MEDICINE, 2011, 3 (01) : 74 - 92
  • [36] Artificial neural network prediction and grey relational grade optimisation of friction stir processing
    Wable, A. D.
    Patil, S. B.
    Pardeshi, S. S.
    [J]. COGENT ENGINEERING, 2023, 10 (01):
  • [37] NEURAL NETWORK SENSOR PROCESSING
    BRAUSE, R
    [J]. MICROPROCESSING AND MICROPROGRAMMING, 1992, 35 (1-5): : 781 - 781
  • [38] Natural language processing with neural networks
    Ma, Q
    [J]. LANGUAGE ENGINEERING CONFERENCE, PROCEEDINGS, 2003, : 45 - 56
  • [39] Neural Correlates of Subliminal Language Processing
    Axelrod, Vadim
    Bar, Moshe
    Rees, Geraint
    Yovel, Galit
    [J]. CEREBRAL CORTEX, 2015, 25 (08) : 2160 - 2169
  • [40] Neural constraints and flexibility in language processing
    Huyck, Christian R.
    [J]. BEHAVIORAL AND BRAIN SCIENCES, 2016, 39