SupCL-Seq: Supervised Contrastive Learning for Downstream Optimized Sequence Representations

被引:0
|
作者
Sedghamiz, Hooman [1 ]
Raval, Shivam [1 ]
Santus, Enrico [1 ]
Alhanai, Tuka [2 ]
Ghassemi, Mohammad [3 ]
机构
[1] Bayer Pharmaceut, DSIG, Whippany, NJ 07981 USA
[2] New York Univ, Abu Dhabi, U Arab Emirates
[3] Michigan State Univ, E Lansing, MI 48824 USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
While contrastive learning is proven to be an effective training strategy in computer vision, Natural Language Processing (NLP) is only recently adopting it as a self-supervised alternative to Masked Language Modeling (MLM) for improving sequence representations. This paper introduces SupCL-Seq, which extends the supervised contrastive learning from computer vision to the optimization of sequence representations in NLP. By altering the dropout mask probability in standard Transformer architectures (e.g. BERTbase), for every representation (anchor), we generate augmented altered views. A supervised contrastive loss is then utilized to maximize the system's capability of pulling together similar samples (e.g., anchors and their altered views) and pushing apart the samples belonging to the other classes. Despite its simplicity, SupCLSeq leads to large gains in many sequence classification tasks on the GLUE benchmark compared to a standard BERTbase, including 6% absolute improvement on CoLA, 5:4% on MRPC, 4:7% on RTE and 2:6% on STSB. We also show consistent gains over selfsupervised contrastively learned representations, especially in non-semantic tasks. Finally we show that these gains are not solely due to augmentation, but rather to a downstream optimized sequence representation. Code: https://github.com/hooman650/SupCL-Seq
引用
收藏
页码:3398 / 3403
页数:6
相关论文
共 26 条
  • [1] Pairwise Supervised Contrastive Learning of Sentence Representations
    Zhang, Dejiao
    Li, Shang-Wen
    Xiao, Wei
    Zhu, Henghui
    Nallapati, Ramesh
    Arnold, Andrew O.
    Xiang, Bing
    2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 5786 - 5798
  • [2] Supervised Contrastive Learning for Text Emotion Category Representations
    Wang, Xiang-Yu
    Zong, Cheng-Qing
    Ruan Jian Xue Bao/Journal of Software, 2024, 35 (10): : 4794 - 4805
  • [3] SuperConText: Supervised Contrastive Learning Framework for Textual Representations
    Moukafih, Youness
    Sbihi, Nada
    Ghogho, Mounir
    Smaili, Kamel
    IEEE ACCESS, 2023, 11 : 16820 - 16830
  • [4] Multimodal Supervised Contrastive Learning in Remote Sensing Downstream Tasks
    Berg, Paul
    Uzun, Baki
    Pham, Minh-Tan
    Courty, Nicolas
    IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2024, 21 : 1 - 5
  • [5] Multi-view Contrastive Self-Supervised Learning of Accounting Data Representations for Downstream Audit Tasks
    Schreyer, Marco
    Sattarov, Timur
    Borth, Damian
    ICAIF 2021: THE SECOND ACM INTERNATIONAL CONFERENCE ON AI IN FINANCE, 2021,
  • [6] WSCFER: Improving Facial Expression Representations by Weak Supervised Contrastive Learning
    Nie, Wei
    Chen, Bowen
    Wu, Wenhao
    Xu, Xiu
    Ren, Weihong
    Liu, Honghai
    2023 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2023, : 9816 - 9823
  • [7] Self-Supervised Visual Representations Learning by Contrastive Mask Prediction
    Zhao, Yucheng
    Wang, Guangting
    Luo, Chong
    Zeng, Wenjun
    Zha, Zheng-Jun
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 10140 - 10149
  • [8] A Supervised Contrastive Framework for Learning Disentangled Representations of Cell Perturbation Data
    Tu, Xinming
    Hutter, Jan-Christian
    JerryWang, Zitong
    Kudo, Takamasa
    Regev, Aviv
    Lopez, Romain
    MACHINE LEARNING IN COMPUTATIONAL BIOLOGY, VOL 240, 2023, 240
  • [9] Optimizing Upstream Representations for Out-of-Domain Detection with Supervised Contrastive Learning
    Wang, Bo
    Mine, Tsunenori
    PROCEEDINGS OF THE 32ND ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2023, 2023, : 2585 - 2595
  • [10] Modeling Discriminative Representations for Out-of-Domain Detection with Supervised Contrastive Learning
    Zeng, Zhiyuan
    He, Keqing
    Yan, Yuanmeng
    Liu, Zijun
    Wu, Yanan
    Xu, Hong
    Jiang, Huixing
    Xu, Weiran
    ACL-IJCNLP 2021: THE 59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING, VOL 2, 2021, : 870 - 878