A Multi-Level Supervised Contrastive Learning Framework for Low-Resource Natural Language Inference

被引:4
|
作者
Li, Shu'ang [1 ]
Hu, Xuming [1 ]
Lin, Li [1 ]
Liu, Aiwei [1 ]
Wen, Lijie [1 ]
Yu, Philip S. S. [2 ]
机构
[1] Tsinghua Univ, Sch Software, Beijing 100190, Peoples R China
[2] Univ Illinois, Dept Comp Sci, Chicago, IL 60607 USA
基金
中国国家自然科学基金;
关键词
Task analysis; Semantics; Training; Annotations; Bicycles; Neural networks; Costs; Natural language inference; contrastive learning; low-resource; multi-level;
D O I
10.1109/TASLP.2023.3270771
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
Natural Language Inference (NLI) is a growingly essential task in natural language understanding, which requires inferring the relationship between the sentence pairs (premise and hypothesis). Recently, low-resource natural language inference has gained increasing attention, due to significant savings in manual annotation costs and a better fit with real-world scenarios. Existing works fail to characterize discriminative representations between different classes with limited training data, which may cause faults in label prediction. Here we propose a multi-level supervised contrastive learning framework named MultiSCL for low-resource natural language inference. MultiSCL leverages a sentence-level and pair-level contrastive learning objective to discriminate between different classes of sentence pairs by bringing those in one class together and pushing away those in different classes. MultiSCL adopts a data augmentation module that generates different views for input samples to better learn the latent representation. The pair-level representation is obtained from a cross attention module. We conduct extensive experiments on two public NLI datasets in low-resource settings, and the accuracy of MultiSCL exceeds other models by 1.8%, 3.1% and 4.1% on SNLI, MNLI and Sick with 5 instances per label respectively. Moreover, our method outperforms the previous state-of-the-art method on cross-domain tasks of text classification.
引用
收藏
页码:1771 / 1783
页数:13
相关论文
共 50 条
  • [1] PAIR-LEVEL SUPERVISED CONTRASTIVE LEARNING FOR NATURAL LANGUAGE INFERENCE
    Li, Shu'ang
    Hu, Xuming
    Lin, Li
    Wen, Lijie
    [J]. 2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 8237 - 8241
  • [2] Multi-level Contrastive Learning Framework for Sequential Recommendation
    Wang, Ziyang
    Liu, Huoyu
    Wei, Wei
    Hu, Yue
    Mao, Xian-Ling
    He, Shaojian
    Fang, Rui
    Chen, Dangyang
    [J]. PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2022, 2022, : 2098 - 2107
  • [3] Multi-level Contrastive Learning for Self-Supervised Vision Transformers
    Mo, Shentong
    Sun, Zhun
    Li, Chao
    [J]. 2023 IEEE/CVF WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV), 2023, : 2777 - 2786
  • [4] Multi-level graph contrastive learning
    Shao, Pengpeng
    Tao, Jianhua
    [J]. NEUROCOMPUTING, 2024, 570
  • [5] Contrastive Learning for Morphological Disambiguation Using Large Language Models in Low-Resource Settings
    Tolegen, Gulmira
    Toleu, Alymzhan
    Mussabayev, Rustam
    [J]. Applied Sciences (Switzerland), 2024, 14 (21):
  • [6] Contrastive fine-tuning for low-resource graph-level transfer learning
    Duan, Yutai
    Liu, Jie
    Chen, Shaowei
    Wu, Jianhua
    [J]. INFORMATION SCIENCES, 2024, 659
  • [7] MoleMCL: a multi-level contrastive learning framework for molecular pre-training
    Zhang, Xinyi
    Xu, Yanni
    Jiang, Changzhi
    Shen, Lian
    Liu, Xiangrong
    [J]. BIOINFORMATICS, 2024, 40 (04)
  • [8] M-Sim: Multi-level Semantic Inference Model for Chinese short answer scoring in low-resource scenarios
    Lai, Peichao
    Ye, Feiyang
    Fu, Yanggeng
    Chen, Zhiwei
    Wu, Yingjie
    Wang, Yilei
    [J]. COMPUTER SPEECH AND LANGUAGE, 2024, 84
  • [9] Towards Spoken Language Understanding via Multi-level Multi-grained Contrastive Learning
    Cheng, Xuxin
    Xu, Wanshi
    Zhu, Zhihong
    Li, Hongxiang
    Zou, Yuexian
    [J]. PROCEEDINGS OF THE 32ND ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2023, 2023, : 326 - 336
  • [10] Multi-level Contrastive Learning for Commonsense Question Answering
    Fang, Quntian
    Huang, Zhen
    Zhang, Ziwen
    Hu, Minghao
    Hu, Biao
    Wang, Ankun
    Li, Dongsheng
    [J]. KNOWLEDGE SCIENCE, ENGINEERING AND MANAGEMENT, PT IV, KSEM 2023, 2023, 14120 : 318 - 331