Ditto: A Simple and Efficient Approach to Improve Sentence Embeddings

被引:0
|
作者
Chen, Qian [1 ]
Wang, Wen [1 ]
Zhang, Qinglin [1 ]
Zheng, Siqi [1 ]
Deng, Chong [1 ]
Yu, Hai [1 ]
Liu, Jiaqing [1 ]
Ma, Yukun [1 ]
Zhang, Chong [1 ]
机构
[1] Alibaba Grp, Speech Lab, Hangzhou, Peoples R China
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Prior studies diagnose the anisotropy problem in sentence representations from pre-trained language models, e.g., BERT, without fine-tuning. Our analysis reveals that the sentence embeddings from BERT suffer from a bias towards uninformative words, limiting the performance in semantic textual similarity (STS) tasks. To address this bias, we propose a simple and efficient unsupervised approach, Diagonal Attention Pooling (Ditto), which weights words with model-based importance estimations and computes the weighted average of word representations from pre-trained models as sentence embeddings. Ditto can be easily applied to any pre-trained language model as a postprocessing operation. Compared to prior sentence embedding approaches, Ditto does not add parameters nor requires any learning. Empirical evaluations demonstrate that our proposed Ditto can alleviate the anisotropy problem and improve various pre-trained models on the STS benchmarks.(1)
引用
收藏
页码:5868 / 5875
页数:8
相关论文
共 50 条
  • [31] Contrastive Learning of Sentence Embeddings from Scratch
    Zhang, Junlei
    Lan, Zhenzhong
    He, Junxian
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING, EMNLP 2023, 2023, : 3916 - 3932
  • [32] BioSentVec: creating sentence embeddings for biomedical texts
    Chen, Qingyu
    Peng, Yifan
    Lu, Zhiyong
    2019 IEEE INTERNATIONAL CONFERENCE ON HEALTHCARE INFORMATICS (ICHI), 2019, : 246 - 250
  • [33] MCSE: Multimodal Contrastive Learning of Sentence Embeddings
    Zhang, Miaoran
    Mosbach, Marius
    Adelani, David Ifeoluwa
    Hedderich, Michael A.
    Klakow, Dietrich
    NAACL 2022: THE 2022 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES, 2022, : 5959 - 5969
  • [34] Carrier Sentence Selection with Word and Context Embeddings
    Yeung, Chak Yan
    Lee, John
    Tsou, Benjamin
    PROCEEDINGS OF THE 2019 INTERNATIONAL CONFERENCE ON ASIAN LANGUAGE PROCESSING (IALP), 2019, : 439 - 444
  • [35] Conceptual Sentence Embeddings Based on Attention Mechanism
    Wang Y.-S.
    Huang H.-Y.
    Feng C.
    Zhou Q.
    Zidonghua Xuebao/Acta Automatica Sinica, 2020, 46 (07): : 1390 - 1400
  • [36] DistillCSE: Distilled Contrastive Learning for Sentence Embeddings
    Xu, Jiahao
    Shao, Wei
    Chen, Lihui
    Liu, Lemao
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EMNLP 2023), 2023, : 8153 - 8165
  • [37] The Impact of Sentence Embeddings in Turkish Paraphrase Detection
    Karaoglan, Bahar
    Yorgancioglu, Hakki Engin
    Kisla, Tarik
    Kumova Metin, Senem
    2019 27TH SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS CONFERENCE (SIU), 2019,
  • [38] Sentence embeddings in NLI with iterative refinement encoders
    Talman, Aarne
    Yli-Jyra, Anssi
    Tiedemann, Joerg
    NATURAL LANGUAGE ENGINEERING, 2019, 25 (04) : 467 - 482
  • [39] Sentence-level Privacy for Document Embeddings
    Meehan, Casey
    Mrini, Khalil
    Chaudhuri, Kamalika
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 3367 - 3380
  • [40] Comparative Study of Sentence Embeddings for Contextual Paraphrasing
    Pragst, Louisa
    Minker, Wolfgang
    Ultes, Stefan
    PROCEEDINGS OF THE 12TH INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION (LREC 2020), 2020, : 6841 - 6851