miCSE: Mutual Information Contrastive Learning for Low-shot Sentence Embeddings

被引:0
|
作者
Klein, Tassilo [1 ]
Nabi, Moin [1 ]
机构
[1] SAP AI Res, Berlin, Germany
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper presents miCSE, a mutual information-based contrastive learning framework that significantly advances the stateof-the-art in few-shot sentence embedding. The proposed approach imposes alignment between the attention pattern of different views during contrastive learning. Learning sentence embeddings with miCSE entails enforcing the structural consistency across augmented views for every sentence, making contrastive self-supervised learning more sample efficient. As a result, the proposed approach shows strong performance in the few-shot learning domain. While it achieves superior results compared to state-of-the-art methods on multiple benchmarks in few-shot learning, it is comparable in the full-shot scenario. This study opens up avenues for efficient self-supervised learning methods that are more robust than current contrastive methods for sentence embedding.1
引用
收藏
页码:6159 / 6177
页数:19
相关论文
共 50 条
  • [1] Low-shot Object Learning with Mutual Exclusivity Bias
    Thai, Anh
    Humayun, Ahmad
    Stojanov, Stefan
    Huang, Zixuan
    Boote, Bikram
    Rehg, James M.
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [2] Composition-contrastive Learning for Sentence Embeddings
    Chanchani, Sachin
    Huang, Ruihong
    [J]. PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023): LONG PAPERS, VOL 1, 2023, : 15836 - 15848
  • [3] SimCSE: Simple Contrastive Learning of Sentence Embeddings
    Gao, Tianyu
    Yao, Xingcheng
    Chen, Danqi
    [J]. 2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 6894 - 6910
  • [4] MCSE: Multimodal Contrastive Learning of Sentence Embeddings
    Zhang, Miaoran
    Mosbach, Marius
    Adelani, David Ifeoluwa
    Hedderich, Michael A.
    Klakow, Dietrich
    [J]. NAACL 2022: THE 2022 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES, 2022, : 5959 - 5969
  • [5] Stable Contrastive Learning for Self-Supervised Sentence Embeddings With Pseudo-Siamese Mutual Learning
    Xie, Yutao
    Wu, Qiyu
    Chen, Wei
    Wang, Tengjiao
    [J]. IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2022, 30 : 3046 - 3059
  • [6] Low-Shot Learning with Imprinted Weights
    Qi, Hang
    Brown, Matthew
    Lowe, David G.
    [J]. 2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 5822 - 5830
  • [7] AdCSE: An Adversarial Method for Contrastive Learning of Sentence Embeddings
    Li, Renhao
    Duan, Lei
    Xie, Guicai
    Xiao, Shan
    Jiang, Weipeng
    [J]. DATABASE SYSTEMS FOR ADVANCED APPLICATIONS, DASFAA 2022, PT III, 2022, : 165 - 180
  • [8] Low-Shot Learning from Imaginary Data
    Wang, Yu-Xiong
    Girshick, Ross
    Hebert, Martial
    Hariharan, Bharath
    [J]. 2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 7278 - 7286
  • [9] Low-shot Learning in Natural Language Processing
    Xia, Congying
    Zhang, Chenwei
    Zhang, Jiawei
    Liang, Tingting
    Peng, Hao
    Yu, Philip S.
    [J]. 2020 IEEE SECOND INTERNATIONAL CONFERENCE ON COGNITIVE MACHINE INTELLIGENCE (COGMI 2020), 2020, : 185 - 189
  • [10] Low-shot learning and class imbalance: a survey
    Billion Polak, Preston
    Prusa, Joseph D.
    Khoshgoftaar, Taghi M.
    [J]. JOURNAL OF BIG DATA, 2024, 11 (01)