Quantifying the Contextualization of Word Representations with Semantic Class Probing

被引:0
|
作者
Zhao, Mengjie [1 ]
Dufter, Philipp [1 ]
Yaghoobzadeh, Yadollah [2 ]
Schutze, Hinrich [1 ]
机构
[1] Ludwig Maximilians Univ Munchen, CIS, Munich, Germany
[2] Microsoft Turing, Montreal, PQ, Canada
基金
欧洲研究理事会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Pretrained language models achieve state-ofthe-art results on many NLP tasks, but there are still many open questions about how and why they work so well. We investigate the contextualization of words in BERT. We quantify the amount of contextualization, i.e., how well words are interpreted in context, by studying the extent to which semantic classes of a word can be inferred from its contextualized embedding. Quantifying contextualization helps in understanding and utilizing pretrained language models. We show that the top layer representations support highly accurate inference of semantic classes; that the strongest contextualization effects occur in the lower layers; that local context is mostly sufficient for contextualizing words; and that top layer representations are more task-specific after finetuning while lower layer representations are more transferable. Finetuning uncovers task-related features, but pretrained knowledge about contextualization is still well preserved.
引用
收藏
页码:1219 / 1234
页数:16
相关论文
共 50 条
  • [1] Improving the Study of Social Representations through Word Associations: Validation of Semantic Contextualization
    Piermatteo, Anthony
    Tavani, Jean-Louis
    Lo Monaco, Gregory
    FIELD METHODS, 2018, 30 (04) : 329 - 344
  • [2] Embedding Semantic Relations into Word Representations
    Bollegala, Danushka
    Maehara, Takanori
    Kawarabayashi, Ken-ichi
    PROCEEDINGS OF THE TWENTY-FOURTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE (IJCAI), 2015, : 1222 - 1228
  • [3] LINSPECTOR: Multilingual Probing Tasks for Word Representations
    Sahin, Goezde Guel
    Vania, Clara
    Kuznetsov, Ilia
    Gurevych, Iryna
    COMPUTATIONAL LINGUISTICS, 2020, 46 (02) : 335 - 385
  • [4] Enhancing Semantic Word Representations by Embedding Deep Word Relationships
    Nugaliyadde, Anupiya
    Wong, Kok Wai
    Sohel, Ferdous
    Xie, Hong
    PROCEEDINGS OF 2019 11TH INTERNATIONAL CONFERENCE ON COMPUTER AND AUTOMATION ENGINEERING (ICCAE 2019), 2019, : 82 - 87
  • [5] The Impact of Word Splitting on the Semantic Content of Contextualized Word Representations
    Soler, Aina Gari
    Labeau, Matthieu
    Clavel, Chloe
    TRANSACTIONS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, 2024, 12 : 299 - 320
  • [6] Semantic representations of word meanings by the cerebral hemispheres
    Ince, E
    Christman, SD
    BRAIN AND LANGUAGE, 2002, 80 (03) : 393 - 420
  • [7] Semantic Frame Identification with Distributed Word Representations
    Hermann, Karl Moritz
    Das, Dipanjan
    Weston, Jason
    Ganchev, Kuzman
    PROCEEDINGS OF THE 52ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1, 2014, : 1448 - 1458
  • [8] Semantic Bottlenecks: Quantifying and Improving Inspectability of Deep Representations
    Losch, Max
    Fritz, Mario
    Schiele, Bernt
    INTERNATIONAL JOURNAL OF COMPUTER VISION, 2021, 129 (11) : 3136 - 3153
  • [9] LINSPECTOR WEB: A Multilingual Probing Suite for Word Representations
    Eichler, Max
    Sahin, Gozde Gul
    Gurevych, Iryna
    2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019): PROCEEDINGS OF SYSTEM DEMONSTRATIONS, 2019, : 127 - 132
  • [10] Semantic Bottlenecks: Quantifying and Improving Inspectability of Deep Representations
    Max Losch
    Mario Fritz
    Bernt Schiele
    International Journal of Computer Vision, 2021, 129 : 3136 - 3153