Ditto: A Simple and Efficient Approach to Improve Sentence Embeddings

被引:0
|
作者
Chen, Qian [1 ]
Wang, Wen [1 ]
Zhang, Qinglin [1 ]
Zheng, Siqi [1 ]
Deng, Chong [1 ]
Yu, Hai [1 ]
Liu, Jiaqing [1 ]
Ma, Yukun [1 ]
Zhang, Chong [1 ]
机构
[1] Alibaba Grp, Speech Lab, Hangzhou, Peoples R China
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Prior studies diagnose the anisotropy problem in sentence representations from pre-trained language models, e.g., BERT, without fine-tuning. Our analysis reveals that the sentence embeddings from BERT suffer from a bias towards uninformative words, limiting the performance in semantic textual similarity (STS) tasks. To address this bias, we propose a simple and efficient unsupervised approach, Diagonal Attention Pooling (Ditto), which weights words with model-based importance estimations and computes the weighted average of word representations from pre-trained models as sentence embeddings. Ditto can be easily applied to any pre-trained language model as a postprocessing operation. Compared to prior sentence embedding approaches, Ditto does not add parameters nor requires any learning. Empirical evaluations demonstrate that our proposed Ditto can alleviate the anisotropy problem and improve various pre-trained models on the STS benchmarks.(1)
引用
收藏
页码:5868 / 5875
页数:8
相关论文
共 50 条
  • [1] Efficient comparison of sentence embeddings
    Zoupanos, Spyros
    Kolovos, Stratis
    Kanavos, Athanasios
    Papadimitriou, Orestis
    Maragoudakis, Manolis
    PROCEEDINGS OF THE 12TH HELLENIC CONFERENCE ON ARTIFICIAL INTELLIGENCE, SETN 2022, 2022,
  • [2] SimCSE: Simple Contrastive Learning of Sentence Embeddings
    Gao, Tianyu
    Yao, Xingcheng
    Chen, Danqi
    2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 6894 - 6910
  • [3] SimTDE: Simple Transformer Distillation for Sentence Embeddings
    Xie, Jian
    He, Xin
    Wang, Jiyang
    Qiu, Zimeng
    Kebarighotbi, Ali
    Ghassemi, Farhad
    PROCEEDINGS OF THE 46TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2023, 2023, : 2389 - 2393
  • [4] Simple Data Transformations for Mitigating the Syntactic Similarity to Improve Sentence Embeddings at Supervised Contrastive Learning
    Kim, Minji
    Cho, Whanhee
    Kim, Soohyeong
    Choi, Yong Suk
    ADVANCED INTELLIGENT SYSTEMS, 2024, 6 (08)
  • [5] Attention-Driven Dropout: A Simple Method to Improve Self-supervised Contrastive Sentence Embeddings
    Stermann, Fabian
    Chalkidis, Ilias
    Vahidi, Amihossein
    Bischl, Bernd
    Rezaei, Mina
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES: RESEARCH TRACK, PT I, ECML PKDD 2024, 2024, 14941 : 89 - 106
  • [6] Simple Techniques for Enhancing Sentence Embeddings in Generative Language Models
    Zhang, Bowen
    Chang, Kehua
    Li, Chunping
    ADVANCED INTELLIGENT COMPUTING TECHNOLOGY AND APPLICATIONS, PT III, ICIC 2024, 2024, 14877 : 52 - 64
  • [7] Unsupervised Random Walk Sentence Embeddings: A Strong but Simple Baseline
    Ethayarajh, Kawin
    REPRESENTATION LEARNING FOR NLP, 2018, : 91 - 100
  • [8] On the Dimensionality of Sentence Embeddings
    Wang, Hongwei
    Zhang, Hongming
    Yu, Dong
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EMNLP 2023), 2023, : 10344 - 10354
  • [9] Conceptual Sentence Embeddings
    Wang, Yashen
    Huang, Heyan
    Feng, Chong
    Zhou, Qiang
    Gu, Jiahui
    WEB-AGE INFORMATION MANAGEMENT, PT I, 2016, 9658 : 390 - 401
  • [10] A Bidirectional LSTM Approach with Word Embeddings for Sentence Boundary Detection
    Chenglin Xu
    Lei Xie
    Xiong Xiao
    Journal of Signal Processing Systems, 2018, 90 : 1063 - 1075