The relational processing limits of classic and contemporary neural network models of language processing

被引:2
|
作者
Puebla, Guillermo [1 ,2 ]
Martin, Andrea E. [3 ,4 ]
Doumas, Leonidas A. A. [1 ]
机构
[1] Univ Edinburgh, Sch Philosophy Psychol & Language Sci, Dept Psychol, Edinburgh, Midlothian, Scotland
[2] Univ Tarapaca, Dept Psychol, Arica, Chile
[3] Max Planck Inst Psycholinguist, Language & Computat Neural Syst Grp, Nijmegen, Netherlands
[4] Radboud Univ Nijmegen, Donders Ctr Cognit Neuroimaging, Nijmegen, Netherlands
关键词
Relational reasoning; generalisation; language processing; neural networks; deep learning; REPRESENTATION; ANALOGY;
D O I
10.1080/23273798.2020.1821906
中图分类号
R36 [病理学]; R76 [耳鼻咽喉科学];
学科分类号
100104 ; 100213 ;
摘要
Whether neural networks can capture relational knowledge is a matter of long-standing controversy. Recently, some researchers have argued that (1) classic connectionist models can handle relational structure and (2) the success of deep learning approaches to natural language processing suggests that structured representations are unnecessary to model human language. We tested the Story Gestalt model, a classic connectionist model of text comprehension, and a Sequence-to-Sequence with Attention model, a modern deep learning architecture for natural language processing. Both models were trained to answer questions about stories based on abstract thematic roles. Two simulations varied the statistical structure of new stories while keeping their relational structure intact. The performance of each model fell below chance at least under one manipulation. We argue that both models fail our tests because they can't perform dynamic binding. These results cast doubts on the suitability of traditional neural networks for explaining relational reasoning and language processing phenomena.
引用
收藏
页码:240 / 254
页数:15
相关论文
共 50 条
  • [1] A Primer on Neural Network Models for Natural Language Processing
    Goldberg, Yoav
    [J]. JOURNAL OF ARTIFICIAL INTELLIGENCE RESEARCH, 2016, 57 : 345 - 420
  • [2] Finding Fuzziness in Neural Network Models of Language Processing
    Misra, Kanishka
    Rayz, Julia Taylor
    [J]. EXPLAINABLE AI AND OTHER APPLICATIONS OF FUZZY TECHNIQUES, NAFIPS 2021, 2022, 258 : 278 - 290
  • [3] Improving Neural Network Models for Natural Language Processing in Russian with Synonyms
    Galinsky, Ruslan
    Alekseev, Anton
    Nikolenko, Sergey I.
    [J]. PROCEEDINGS OF THE 2016 IEEE ARTIFICIAL INTELLIGENCE AND NATURAL LANGUAGE CONFERENCE (AINL FRUCT 2016), 2016, : 45 - 51
  • [4] SYMMETRY PROCESSING IN NEURAL NETWORK MODELS
    MAS, J
    RAMOS, E
    [J]. JOURNAL OF PHYSICS A-MATHEMATICAL AND GENERAL, 1989, 22 (16): : 3379 - 3391
  • [5] Grey Relational Analysis and Natural Language Processing to: Grey Language Processing
    Khuman, Arjab Singh
    Yang, Yingjie
    Liu, Sifeng
    [J]. JOURNAL OF GREY SYSTEM, 2016, 28 (01): : 88 - 97
  • [6] Emergence in Neural Network Models of Cognitive Processing
    Penna, Maria Pietronilla
    Hitchcott, Paul Kenneth
    Fastame, Maria Chiara
    Pessa, Eliano
    [J]. TOWARDS A POST-BERTALANFFY SYSTEMICS, 2016, : 117 - 126
  • [7] APPLICATION OF CONVOLUTIONAL NEURAL NETWORK IN NATURAL LANGUAGE PROCESSING
    Li, Ping
    Li, Jianping
    Wang, Gongcheng
    [J]. 2018 15TH INTERNATIONAL COMPUTER CONFERENCE ON WAVELET ACTIVE MEDIA TECHNOLOGY AND INFORMATION PROCESSING (ICCWAMTIP), 2018, : 120 - 122
  • [8] A Natural Language Processing Neural Network Comprehending English
    Ke, Yuanzhi
    Hagiwara, Masafumi
    [J]. 2015 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2015,
  • [9] Natural Language Processing Neural Network for Recall and Inference
    Sagara, Tsukasa
    Hagiwara, Masafumi
    [J]. ARTIFICIAL NEURAL NETWORKS (ICANN 2010), PT III, 2010, 6354 : 286 - 289
  • [10] Natural language processing neural network for analogical inference
    Faculty of Science and Technology, Keio University, 3-14-1 Hiyoshi, Kohoku-ku, Yokohama, 223-8522, Japan
    [J]. Proc Int Jt Conf Neural Networks, 2010,