An Efficient Self-Supervised Cross-View Training For Sentence Embedding

被引:0
|
作者
Limkonchotiwat, Peerat [1 ]
Ponwitayarat, Wuttikorn [1 ]
Lowphansirikul, Lalita [1 ]
Udomcharoenchaikit, Can [1 ]
Chuangsuwanich, Ekapol [2 ]
Nutanong, Sarana [1 ]
机构
[1] VISTEC, Sch Informat Sci & Technol, Rayong, Thailand
[2] Chulalongkorn Univ, Dept Comp Engn, Bangkok, Thailand
关键词
Computational linguistics - Semantics;
D O I
10.1162/tacl_a_00620
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Self-supervised sentence representation learning is the task of constructing an embedding space for sentences without relying on human annotation efforts. One straightforward approach is to finetune a pretrained language model (PLM) with a representation learning method such as contrastive learning. While this approach achieves impressive performance on larger PLMs, the performance rapidly degrades as the number of parameters decreases. In this paper, we propose a framework called Self-supervised Cross-View Training (SCT) to narrow the performance gap between large and small PLMs. To evaluate the effectiveness of SCT, we compare it to 5 baseline and state-of-the-art competitors on seven Semantic Textual Similarity (STS) benchmarks using 5 PLMs with the number of parameters ranging from 4M to 340M. The experimental results show that STC outperforms the competitors for PLMs with less than 100M parameters in 18 of 21 cases.1
引用
下载
收藏
页码:1572 / 1587
页数:16
相关论文
共 50 条
  • [21] A SELF-SUPERVISED NETWORK EMBEDDING METHOD
    Wang, Dongqi
    Chen, Xin
    Chen, Dongming
    JOURNAL OF NONLINEAR AND CONVEX ANALYSIS, 2022, 23 (10) : 2319 - 2328
  • [22] Text Summary Generation Model Based on Sentence Fusion and Self-Supervised Training
    Zou A.
    Hao W.
    Jin D.
    Chen G.
    Moshi Shibie yu Rengong Zhineng/Pattern Recognition and Artificial Intelligence, 2022, 35 (05): : 401 - 411
  • [23] Graph regularized supervised cross-view hashing
    Xin Shu
    Haiyan Jiang
    Huanliang Xu
    Multimedia Tools and Applications, 2018, 77 : 28207 - 28224
  • [24] Accelerating Self-Supervised Learning via Efficient Training Strategies
    Kocyigit, Mustafa Taha
    Hospedales, Timothy M.
    Bilen, Hakan
    2023 IEEE/CVF WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV), 2023, : 5643 - 5653
  • [25] SELF-SUPERVISED ADVERSARIAL TRAINING
    Chen, Kejiang
    Chen, Yuefeng
    Zhou, Hang
    Mao, Xiaofeng
    Li, Yuhong
    He, Yuan
    Xue, Hui
    Zhang, Weiming
    Yu, Nenghai
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 2218 - 2222
  • [26] Community Detection Based on Deep Network Embedding with Dual Self-supervised Training
    Chen, Yunfang
    Mao, Haotian
    Wang, Li
    Zhang, Wei
    Communications in Computer and Information Science, 2022, 1563 CCIS : 643 - 656
  • [27] Swarm Self-supervised Hypergraph Embedding for Recommendation
    Jian, Meng
    Bai, Yulong
    Guo, Jingjing
    Wu, Lifang
    ACM TRANSACTIONS ON KNOWLEDGE DISCOVERY FROM DATA, 2024, 18 (04)
  • [28] Putting Self-Supervised Token Embedding on the Tables
    Szafraniec, Marc
    Marti, Gautier
    Donnat, Philippe
    2017 16TH IEEE INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS (ICMLA), 2017, : 707 - 710
  • [29] TirSA: A Three Stage Approach for UAV-Satellite Cross-View Geo-Localization Based on Self-Supervised Feature Enhancement
    Sun J.
    Sun H.
    Lei L.
    Ji K.
    Kuang G.
    IEEE Transactions on Circuits and Systems for Video Technology, 2024, 34 (09) : 1 - 1
  • [30] Grouped Contrastive Learning of Self-Supervised Sentence Representation
    Wang, Qian
    Zhang, Weiqi
    Lei, Tianyi
    Peng, Dezhong
    APPLIED SCIENCES-BASEL, 2023, 13 (17):