Pitfalls in the Evaluation of Sentence Embeddings

被引:0
|
作者
Eger, Steffen [1 ,2 ]
Rueckle, Andreas [1 ]
Gurevych, Iryna [1 ,2 ]
机构
[1] Tech Univ Darmstadt, Dept Comp Sci, Ubiquitous Knowledge Proc Lab UKP TUDA, Darmstadt, Germany
[2] Tech Univ Darmstadt, Dept Comp Sci, Res Training Grp AIPHES, Darmstadt, Germany
关键词
D O I
暂无
中图分类号
H0 [语言学];
学科分类号
030303 ; 0501 ; 050102 ;
摘要
Deep learning models continuously break new records across different NLP tasks. At the same time, their success exposes weaknesses of model evaluation. Here, we compile several key pitfalls of evaluation of sentence embeddings, a currently very popular NLP paradigm. These pitfalls include the comparison of embeddings of different sizes, normalization of embeddings, and the low (and diverging) correlations between transfer and probing tasks. Our motivation is to challenge the current evaluation of sentence embeddings and to provide an easy-to-access reference for future research. Based on our insights, we also recommend better practices for better future evaluations of sentence embeddings.
引用
收藏
页码:55 / 60
页数:6
相关论文
共 50 条
  • [21] Carrier Sentence Selection with Word and Context Embeddings
    Yeung, Chak Yan
    Lee, John
    Tsou, Benjamin
    PROCEEDINGS OF THE 2019 INTERNATIONAL CONFERENCE ON ASIAN LANGUAGE PROCESSING (IALP), 2019, : 439 - 444
  • [22] Conceptual Sentence Embeddings Based on Attention Mechanism
    Wang Y.-S.
    Huang H.-Y.
    Feng C.
    Zhou Q.
    Zidonghua Xuebao/Acta Automatica Sinica, 2020, 46 (07): : 1390 - 1400
  • [23] DistillCSE: Distilled Contrastive Learning for Sentence Embeddings
    Xu, Jiahao
    Shao, Wei
    Chen, Lihui
    Liu, Lemao
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EMNLP 2023), 2023, : 8153 - 8165
  • [24] The Impact of Sentence Embeddings in Turkish Paraphrase Detection
    Karaoglan, Bahar
    Yorgancioglu, Hakki Engin
    Kisla, Tarik
    Kumova Metin, Senem
    2019 27TH SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS CONFERENCE (SIU), 2019,
  • [25] Sentence embeddings in NLI with iterative refinement encoders
    Talman, Aarne
    Yli-Jyra, Anssi
    Tiedemann, Joerg
    NATURAL LANGUAGE ENGINEERING, 2019, 25 (04) : 467 - 482
  • [26] Sentence-level Privacy for Document Embeddings
    Meehan, Casey
    Mrini, Khalil
    Chaudhuri, Kamalika
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 3367 - 3380
  • [27] Comparative Study of Sentence Embeddings for Contextual Paraphrasing
    Pragst, Louisa
    Minker, Wolfgang
    Ultes, Stefan
    PROCEEDINGS OF THE 12TH INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION (LREC 2020), 2020, : 6841 - 6851
  • [28] DefSent: Sentence Embeddings using Definition Sentences
    Tsukagoshi, Hayato
    Sasano, Ryohei
    Takeda, Koichi
    ACL-IJCNLP 2021: THE 59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING, VOL 2, 2021, : 411 - 418
  • [29] How doWords Contribute to Sentence Semantics? Revisiting Sentence Embeddings with a Perturbation Method
    Yao, Wenlin
    Jin, Lifeng
    Zhang, Hongming
    Pan, Xiaoman
    Song, Kaiqiang
    Yu, Dian
    Yu, Dong
    Chen, Jianshu
    17TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EACL 2023, 2023, : 3001 - 3010
  • [30] Revisiting Recurrent Networks for Paraphrastic Sentence Embeddings
    Wieting, John
    Gimpel, Kevin
    PROCEEDINGS OF THE 55TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2017), VOL 1, 2017, : 2078 - 2088