Quantifying the Contextualization of Word Representations with Semantic Class Probing

被引:0
|
作者
Zhao, Mengjie [1 ]
Dufter, Philipp [1 ]
Yaghoobzadeh, Yadollah [2 ]
Schutze, Hinrich [1 ]
机构
[1] Ludwig Maximilians Univ Munchen, CIS, Munich, Germany
[2] Microsoft Turing, Montreal, PQ, Canada
基金
欧洲研究理事会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Pretrained language models achieve state-ofthe-art results on many NLP tasks, but there are still many open questions about how and why they work so well. We investigate the contextualization of words in BERT. We quantify the amount of contextualization, i.e., how well words are interpreted in context, by studying the extent to which semantic classes of a word can be inferred from its contextualized embedding. Quantifying contextualization helps in understanding and utilizing pretrained language models. We show that the top layer representations support highly accurate inference of semantic classes; that the strongest contextualization effects occur in the lower layers; that local context is mostly sufficient for contextualizing words; and that top layer representations are more task-specific after finetuning while lower layer representations are more transferable. Finetuning uncovers task-related features, but pretrained knowledge about contextualization is still well preserved.
引用
收藏
页码:1219 / 1234
页数:16
相关论文
共 50 条
  • [41] Extracting semantic representations from word co-occurrence statistics: A computational study
    Bullinaria, John A.
    Levy, Joseph P.
    BEHAVIOR RESEARCH METHODS, 2007, 39 (03) : 510 - 526
  • [42] Professional Music Training and Novel Word Learning: From Faster Semantic Encoding to Longer-lasting Word Representations
    Dittinger, Eva
    Barbaroux, Mylene
    D'Imperio, Mariapaola
    Jancke, Lutz
    Elmer, Stefan
    Besson, Mireille
    JOURNAL OF COGNITIVE NEUROSCIENCE, 2016, 28 (10) : 1584 - 1602
  • [43] Tweet Contextualization Approach Using a Semantic Query Expansion
    Dhokar, Amira
    Hlaoua, Lobna
    Ben Romdhane, Lotfi
    KNOWLEDGE-BASED AND INTELLIGENT INFORMATION & ENGINEERING SYSTEMS (KSE 2021), 2021, 192 : 387 - 396
  • [44] On Quantifying Semantic Information
    D'Alfonso, Simon
    INFORMATION, 2011, 2 (01): : 61 - 101
  • [45] Better Word Representations with Word Weight
    Song, Gege
    Huang, Xianglin
    Cao, Gang
    Tao, Zhulin
    Liu, Wei
    Yang, Lifang
    2019 IEEE 21ST INTERNATIONAL WORKSHOP ON MULTIMEDIA SIGNAL PROCESSING (MMSP 2019), 2019,
  • [46] The impact of semantic reference on word class: An fMRI study of action and object naming
    Saccuman, M. Cristina
    Cappa, Stefano F.
    Bates, Elizabeth A.
    Arevalo, Analia
    Della Rosa, Pasquale
    Danna, Massimo
    Perani, Daniela
    NEUROIMAGE, 2006, 32 (04) : 1865 - 1878
  • [47] Semantic domain and grammatical class effects in the picture-word interference paradigm
    Rodriguez-Ferreiro, Javier
    Davies, Robert
    Cuetos, Fernando
    LANGUAGE COGNITION AND NEUROSCIENCE, 2014, 29 (01) : 125 - 135
  • [48] The linguistic summarization and the interpretability, scalability of fuzzy representations of multilevel semantic structures of word-domains
    Cat Ho Nguyen
    Thi Lan Pham
    Nguyen, Tu N.
    Cam Ha Ho
    Thu Anh Nguyen
    MICROPROCESSORS AND MICROSYSTEMS, 2021, 81 (81)
  • [49] ENHANCING SPARSE VOICE ANNOTATION FOR SEMANTIC RETRIEVAL OF PERSONAL PHOTOS BY CONTINUOUS SPACE WORD REPRESENTATIONS
    Liou, Yuan-ming
    Lu, Hung-tsung
    Fu, Yi-sheng
    Hsu, Winston
    Lee, Lin-shan
    2015 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING (ICASSP), 2015, : 5341 - 5345
  • [50] Semantic relation classification through low-dimensional distributed representations of partial word sequences
    Jin, Zhan
    Shibata, Chihiro
    Tago, Kazuya
    IEICE NONLINEAR THEORY AND ITS APPLICATIONS, 2019, 10 (01): : 28 - 44