Overestimation of Syntactic Representation in Neural Language Models

被引:0
|
作者
Kodner, Jordan [1 ]
Gupta, Nitish [2 ]
机构
[1] Univ Penn, Dept Linguist, Philadelphia, PA 19104 USA
[2] Univ Penn, Dept Comp & Informat Sci, Philadelphia, PA 19104 USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
With the advent of powerful neural language models over the last few years, research attention has increasingly focused on what aspects of language they represent that make them so successful. Several testing methodologies have been developed to probe models' syntactic representations. One popular method for determining a model's ability to induce syntactic structure trains a model on strings generated according to a template then tests the model's ability to distinguish such strings from superficially similar ones with different syntax. We illustrate a fundamental problem with this approach by reproducing positive results from a recent paper with two non-syntactic baseline language models: an n-gram model and an LSTM model trained on scrambled inputs.
引用
收藏
页码:1757 / 1762
页数:6
相关论文
共 50 条
  • [21] The representation of verbs: Evidence from syntactic priming in language production
    Pickering, MJ
    Branigan, HP
    [J]. JOURNAL OF MEMORY AND LANGUAGE, 1998, 39 (04) : 633 - 651
  • [22] Language Representation Models: An Overview
    Schomacker, Thorben
    Tropmann-Frick, Marina
    [J]. ENTROPY, 2021, 23 (11)
  • [23] The neural representation of language in users of American Sign Language
    Corina, DP
    McBurney, SL
    [J]. JOURNAL OF COMMUNICATION DISORDERS, 2001, 34 (06) : 455 - 471
  • [24] Neural aspects of second language representation and language control
    Abutalebi, Jubin
    [J]. ACTA PSYCHOLOGICA, 2008, 128 (03) : 466 - 478
  • [25] Comparative Study of Parametric and Representation Uncertainty Modeling for Recurrent Neural Network Language Models
    Yu, Jianwei
    Lam, Max W. Y.
    Hu, Shoukang
    Wu, Xixin
    Li, Xu
    Gao, Yuewen
    Liu, Xunying
    Meng, Helen
    [J]. INTERSPEECH 2019, 2019, : 3510 - 3514
  • [26] Large language models implicitly learn to straighten neural sentence trajectories to construct a predictive representation of natural language
    Hosseini, Eghbal A.
    Fedorenko, Evelina
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [27] Neural correlates of semantic and syntactic processing in German Sign Language
    Stroh, Anna-Lena
    Roesler, Frank
    Dormal, Giulia
    Salden, Uta
    Skotara, Nils
    Haenel-Faulhaber, Barbara
    Roeder, Brigitte
    [J]. NEUROIMAGE, 2019, 200 : 231 - 241
  • [28] Representation of the language and models of language change: grammaticalization as perspective
    Feltgen, Quentin
    Fagard, Benjamin
    Nadal, Jean-Pierre
    [J]. TRAITEMENT AUTOMATIQUE DES LANGUES, 2014, 55 (03): : 47 - 71
  • [29] Syntactic analysis of the sentences of the Russian language based on neural networks
    Sboev, A. G.
    Rybka, R.
    Moloshnikov, I.
    Gudovskih, D.
    [J]. 4TH INTERNATIONAL YOUNG SCIENTIST CONFERENCE ON COMPUTATIONAL SCIENCE, 2015, 66 : 277 - 286
  • [30] PARAPHRASTIC LANGUAGE MODELS AND COMBINATION WITH NEURAL NETWORK LANGUAGE MODELS
    Liu, X.
    Gales, M. J. F.
    Woodland, P. C.
    [J]. 2013 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2013, : 8421 - 8425