SuperConText: Supervised Contrastive Learning Framework for Textual Representations

被引:1
|
作者
Moukafih, Youness [1 ,2 ]
Sbihi, Nada [1 ]
Ghogho, Mounir [1 ]
Smaili, Kamel [2 ]
机构
[1] Univ Int Rabat, Coll Engn & Architecture, TIC Lab, Sale 11103, Morocco
[2] Loria, Campus Sci, Vandoeuvre Les Nancy, France
关键词
Training; Task analysis; Benchmark testing; Representation learning; Entropy; Deep learning; Text categorization; contrastive learning; text classification; hard negative examples;
D O I
10.1109/ACCESS.2023.3241490
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In the last decade, Deep neural networks (DNNs) have been proven to outperform conventional machine learning models in supervised learning tasks. Most of these models are typically optimized by minimizing the well-known Cross-Entropy objective function. The latter, however, has a number of drawbacks, including poor margins and instability. Taking inspiration from the recent self-supervised Contrastive representation learning approaches, we introduce Supervised Contrastive learning framework for Textual representations (SuperConText) to address those issues. We pretrain a neural network by minimizing a novel fully-supervised contrastive loss. The goal is to increase both inter-class separability and intra-class compactness of the embeddings in the latent space. Examples belonging to the same class are regarded as positive pairs, while examples belonging to different classes are considered negatives. Further, we propose a simple yet effective method for selecting hard negatives during the training phase. In extensive series of experiments, we study the impact of a number of parameters on the quality of the learned representations (e.g. the batch size). Simulation results show that the proposed solution outperforms several competing approaches on various large-scale text classification benchmarks without requiring specialized architectures, data augmentations, memory banks, or additional unsupervised data. For instance, we achieved top-1 accuracy of 61.94% on the Amazon-F dataset, which is 3.54% above the best result obtained when using the cross-entropy with the same model architecture.
引用
收藏
页码:16820 / 16830
页数:11
相关论文
共 50 条
  • [1] A Supervised Contrastive Framework for Learning Disentangled Representations of Cell Perturbation Data
    Tu, Xinming
    Hutter, Jan-Christian
    JerryWang, Zitong
    Kudo, Takamasa
    Regev, Aviv
    Lopez, Romain
    MACHINE LEARNING IN COMPUTATIONAL BIOLOGY, VOL 240, 2023, 240
  • [2] Pairwise Supervised Contrastive Learning of Sentence Representations
    Zhang, Dejiao
    Li, Shang-Wen
    Xiao, Wei
    Zhu, Henghui
    Nallapati, Ramesh
    Arnold, Andrew O.
    Xiang, Bing
    2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 5786 - 5798
  • [3] DeCLUTR: Deep Contrastive Learning for Unsupervised Textual Representations
    Giorgi, John
    Nitski, Osvald
    Wang, Bo
    Bader, Gary
    59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING, VOL 1 (ACL-IJCNLP 2021), 2021, : 879 - 895
  • [4] Supervised Contrastive Learning for Text Emotion Category Representations
    Wang, Xiang-Yu
    Zong, Cheng-Qing
    Ruan Jian Xue Bao/Journal of Software, 2024, 35 (10): : 4794 - 4805
  • [5] A Simplified Framework for Contrastive Learning for Node Representations
    Hong, Ilgee
    Huy Tran
    Donnat, Claire
    FIFTY-SEVENTH ASILOMAR CONFERENCE ON SIGNALS, SYSTEMS & COMPUTERS, IEEECONF, 2023, : 573 - 577
  • [6] A Simple Framework for Contrastive Learning of Visual Representations
    Chen, Ting
    Kornblith, Simon
    Norouzi, Mohammad
    Hinton, Geoffrey
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 119, 2020, 119
  • [7] A Simple Framework for Contrastive Learning of Visual Representations
    Chen, Ting
    Kornblith, Simon
    Norouzi, Mohammad
    Hinton, Geoffrey
    25TH AMERICAS CONFERENCE ON INFORMATION SYSTEMS (AMCIS 2019), 2019,
  • [8] A Probabilistic Contrastive Framework for Semi-Supervised Learning
    Lin, Huibin
    Zhang, Chun-Yang
    Wang, Shiping
    Guo, Wenzhong
    IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 8767 - 8779
  • [9] Towards a Unified Framework of Contrastive Learning for Disentangled Representations
    Matthes, Stefan
    Han, Zhiwei
    Shen, Hao
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [10] WSCFER: Improving Facial Expression Representations by Weak Supervised Contrastive Learning
    Nie, Wei
    Chen, Bowen
    Wu, Wenhao
    Xu, Xiu
    Ren, Weihong
    Liu, Honghai
    2023 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2023, : 9816 - 9823