Neural Language Models as Psycholinguistic Subjects: Representations of Syntactic State

被引:0
|
作者
Futrell, Richard [1 ]
Wilcox, Ethan [2 ]
Morita, Takashi [3 ,4 ]
Qian, Peng [5 ]
Ballesteros, Miguel [6 ]
Levy, Roger [5 ]
机构
[1] Univ Calif Irvine, Dept Language Sci, Irvine, CA 92697 USA
[2] Harvard Univ, Dept Linguist, Cambridge, MA USA
[3] Kyoto Univ, Primate Res Inst, Kyoto, Japan
[4] MIT, Dept Linguist & Philosophy, Cambridge, MA USA
[5] MIT, Dept Brain & Cognit Sci, Cambridge, MA USA
[6] MIT, IBM Watson Lab, IBM Res, Cambridge, MA USA
关键词
PREDICTION;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We investigate the extent to which the behavior of neural network language models reflects incremental representations of syntactic state. To do so, we employ experimental methodologies which were originally developed in the field of psycholinguistics to study syntactic representation in the human mind. We examine neural network model behavior on sets of artificial sentences containing a variety of syntactically complex structures. These sentences not only test whether the networks have a representation of syntactic state, they also reveal the specific lexical cues that networks use to update these states. We test four models: two publicly available LSTM sequence models of English (Jozefowicz et al., 2016; Gulordava et al., 2018) trained on large datasets; an RNN Grammar (Dyer et al., 2016) trained on a small, parsed dataset; and an LSTM trained on the same small corpus as the RNNG. We find evidence for basic syntactic state representations in all models, but only the models trained on large datasets are sensitive to subtle lexical cues signalling changes in syntactic state.
引用
收藏
页码:32 / 42
页数:11
相关论文
共 50 条
  • [21] From language-specific to shared syntactic representations: The influence of second language proficiency on syntactic sharing in bilinguals
    Bernolet, Sarah
    Hartsuiker, Robert J.
    Pickering, Martin J.
    COGNITION, 2013, 127 (03) : 287 - 306
  • [22] Neural Representations of Natural Language
    Feng, Haoda
    NATURAL LANGUAGE ENGINEERING, 2022, 28 (05) : 669 - 672
  • [23] Exact training of a neural syntactic language model
    Emami, A
    Jelinek, F
    2004 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, VOL I, PROCEEDINGS: SPEECH PROCESSING, 2004, : 245 - 248
  • [24] Sense representations for Portuguese: experiments with sense embeddings and deep neural language models
    Jéssica Rodrigues da Silva
    Helena de M. Caseli
    Language Resources and Evaluation, 2021, 55 : 901 - 924
  • [25] Sense representations for Portuguese: experiments with sense embeddings and deep neural language models
    Rodrigues da Silva, Jessica
    Caseli, Helena de M.
    LANGUAGE RESOURCES AND EVALUATION, 2021, 55 (04) : 901 - 924
  • [26] Enriching Rare Word Representations in Neural Language Models by Embedding Matrix Augmentation
    Khassanov, Yerbolat
    Zeng, Zhiping
    Van Tung Pham
    Xu, Haihua
    Chng, Eng Siong
    INTERSPEECH 2019, 2019, : 3505 - 3509
  • [27] Syntactic processing models and their implications for the study of language
    Ayelen Stetie, Noelia
    REVISTA DE ESTUDOS DA LINGUAGEM, 2021, 29 (03) : 2117 - 2162
  • [28] Refining Targeted Syntactic Evaluation of Language Models
    Newman, Benjamin
    Ang, Kai-Siang
    Gong, Julia
    Hewitt, John
    2021 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL-HLT 2021), 2021, : 3710 - 3723
  • [29] Effects of Syntactic Distance and Word Order on Language Processing: An Investigation Based on a Psycholinguistic Treebank of English
    Ruochen Niu
    Haitao Liu
    Journal of Psycholinguistic Research, 2022, 51 : 1043 - 1062
  • [30] Effects of Syntactic Distance and Word Order on Language Processing: An Investigation Based on a Psycholinguistic Treebank of English
    Niu, Ruochen
    Liu, Haitao
    JOURNAL OF PSYCHOLINGUISTIC RESEARCH, 2022, 51 (05) : 1043 - 1062