Neural Language Models as Psycholinguistic Subjects: Representations of Syntactic State

被引:0
|
作者
Futrell, Richard [1 ]
Wilcox, Ethan [2 ]
Morita, Takashi [3 ,4 ]
Qian, Peng [5 ]
Ballesteros, Miguel [6 ]
Levy, Roger [5 ]
机构
[1] Univ Calif Irvine, Dept Language Sci, Irvine, CA 92697 USA
[2] Harvard Univ, Dept Linguist, Cambridge, MA USA
[3] Kyoto Univ, Primate Res Inst, Kyoto, Japan
[4] MIT, Dept Linguist & Philosophy, Cambridge, MA USA
[5] MIT, Dept Brain & Cognit Sci, Cambridge, MA USA
[6] MIT, IBM Watson Lab, IBM Res, Cambridge, MA USA
关键词
PREDICTION;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We investigate the extent to which the behavior of neural network language models reflects incremental representations of syntactic state. To do so, we employ experimental methodologies which were originally developed in the field of psycholinguistics to study syntactic representation in the human mind. We examine neural network model behavior on sets of artificial sentences containing a variety of syntactically complex structures. These sentences not only test whether the networks have a representation of syntactic state, they also reveal the specific lexical cues that networks use to update these states. We test four models: two publicly available LSTM sequence models of English (Jozefowicz et al., 2016; Gulordava et al., 2018) trained on large datasets; an RNN Grammar (Dyer et al., 2016) trained on a small, parsed dataset; and an LSTM trained on the same small corpus as the RNNG. We find evidence for basic syntactic state representations in all models, but only the models trained on large datasets are sensitive to subtle lexical cues signalling changes in syntactic state.
引用
收藏
页码:32 / 42
页数:11
相关论文
共 50 条
  • [31] y An Analysis of the Utility of Explicit Negative Examples to Improve the Syntactic Abilities of Neural Language Models
    Noji, Hiroshi
    Takamura, Hiroya
    58TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2020), 2020, : 3375 - 3385
  • [32] Syntactic Reanalysis in Language Models for Speech Recognition
    Twiefel, Johannes
    Hinaut, Xavier
    Wermter, Stefan
    2017 THE SEVENTH JOINT IEEE INTERNATIONAL CONFERENCE ON DEVELOPMENT AND LEARNING AND EPIGENETIC ROBOTICS (ICDL-EPIROB), 2017, : 215 - 220
  • [33] Structural Supervision Improves Few-Shot Learning and Syntactic Generalization in Neural Language Models
    Wilcox, Ethan
    Qian, Peng
    Futrell, Richard
    Kohita, Ryosuke
    Levy, Roger
    Ballesteros, Miguel
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 4640 - 4652
  • [34] Encoding syntactic representations with a neural network for sentiment collocation extraction
    Yanyan Zhao
    Bing Qin
    Ting Liu
    Science China Information Sciences, 2017, 60
  • [35] Encoding syntactic representations with a neural network for sentiment collocation extraction
    Yanyan ZHAO
    Bing QIN
    Ting LIU
    ScienceChina(InformationSciences), 2017, 60 (11) : 7 - 18
  • [36] Encoding syntactic representations with a neural network for sentiment collocation extraction
    Zhao, Yanyan
    Qin, Bing
    Liu, Ting
    SCIENCE CHINA-INFORMATION SCIENCES, 2017, 60 (11)
  • [37] Correlating neural and symbolic representations of language
    Chrupala, Grzegorz
    Alishahi, Afra
    57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), 2019, : 2952 - 2962
  • [38] Unifying Lexical, Syntactic, and Structural Representations of Written Language for Authorship Attribution
    Jafariakinabad F.
    Hua K.A.
    SN Computer Science, 2021, 2 (6)
  • [39] Language comprehension in the bilingual brain: fMRI and ERP support for psycholinguistic models
    van Heuven, Walter J. B.
    Dijkstra, Ton
    BRAIN RESEARCH REVIEWS, 2010, 64 (01) : 104 - 122
  • [40] Can large language models help augment English psycholinguistic datasets?
    Trott, Sean
    BEHAVIOR RESEARCH METHODS, 2024, 56 (06) : 6082 - 6100