Simple Data Transformations for Mitigating the Syntactic Similarity to Improve Sentence Embeddings at Supervised Contrastive Learning

被引:0
|
作者
Kim, Minji [1 ]
Cho, Whanhee [2 ,3 ]
Kim, Soohyeong [1 ]
Choi, Yong Suk [1 ,2 ]
机构
[1] Hanyang Univ, Dept Artificial Intelligence, Seoul 04763, South Korea
[2] Hanyang Univ, Dept Comp Sci, Seoul 04763, South Korea
[3] Univ Utah, Sch Comp, Salt Lake City, UT 84112 USA
基金
新加坡国家研究基金会;
关键词
contrastive learning; sentence embedding; syntactic transformation;
D O I
10.1002/aisy.202300717
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Contrastive learning of sentence representations has achieved great improvements in several natural language processing tasks. However, the supervised contrastive learning model trained on the natural language inference (NLI) dataset is insufficient to elucidate the semantics of sentences since it is prone to make a prediction based on heuristics. Herein, by using the ParsEVAL and the word overlap metric, it is shown that sentence pairs in the NLI dataset have strong syntactic similarity and propose a framework to compensate for this problem in two aspects. 1) Apply simple syntactic transformations to the hypothesis and 2) expand the objective to SupCon Loss to leverage variants of sentences. The method is evaluated on semantic textual similarity (STS) tasks and transfer tasks. The proposed methods improve the performance of the BERT-based baseline in STS Benchmark and SICK Relatedness by 1.48% and 2.2%. Furthermore, the model achieves 82.65% on the HANS benchmark dataset, to the best of our knowledge, which is a state-of-the-art performance demonstrating that our approach is effective in grasping semantics without heuristics in the NLI dataset at supervised contrastive learning. The code is available at . Using the ParsEVAL and the word overlap metric, it shows that sentence pairs in the natural language inference dataset have strong syntactic similarity. For compensating this problem, applying simple syntactic transformations and expanding the objective to SupCon Loss to leverage variants of sentences are used. This approach is effective in grasping semantics without heuristics.image (c) 2024 WILEY-VCH GmbH
引用
收藏
页数:10
相关论文
共 50 条
  • [41] Simple Flow-Based Contrastive Learning for BERT Sentence Representations
    Tian, Ziyi
    Liu, Qun
    Liu, Maotao
    Deng, Wei
    ADVANCES IN SWARM INTELLIGENCE, ICSI 2022, PT II, 2022, : 265 - 275
  • [42] Prefix Data Augmentation for Contrastive Learning of Unsupervised Sentence Embedding
    Wang, Chunchun
    Lv, Shu
    APPLIED SCIENCES-BASEL, 2024, 14 (07):
  • [43] Importance-aware contrastive learning via semantically augmented instances for unsupervised sentence embeddings
    Xin Ma
    Hong Li
    Jiawen Shi
    Yi Zhang
    Zhigao Long
    International Journal of Machine Learning and Cybernetics, 2023, 14 : 2979 - 2990
  • [44] Supervised contrastive learning over prototype-label embeddings for network intrusion detection
    Lopez-Martin, Manuel
    Sanchez-Esguevillas, Antonio
    Arribas, Juan Ignacio
    Carro, Belen
    INFORMATION FUSION, 2022, 79 : 200 - 228
  • [45] Importance-aware contrastive learning via semantically augmented instances for unsupervised sentence embeddings
    Ma, Xin
    Li, Hong
    Shi, Jiawen
    Zhang, Yi
    Long, Zhigao
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2023, 14 (09) : 2979 - 2990
  • [46] Deep Learning Embeddings for Data Series Similarity Search
    Wang, Qitong
    Palpanas, Themis
    KDD '21: PROCEEDINGS OF THE 27TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY & DATA MINING, 2021, : 1708 - 1716
  • [47] Weakly Supervised Temporal Sentence Grounding with Gaussian-based Contrastive Proposal Learning
    Zheng, Minghang
    Huang, Yanjie
    Chen, Qingchao
    Peng, Yuxin
    Liu, Yang
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 15534 - 15543
  • [48] CleanCLIP: Mitigating Data Poisoning Attacks in Multimodal Contrastive Learning
    Bansal, Hritik
    Singhi, Nishad
    Yang, Yu
    Yin, Fan
    Grover, Aditya
    Chang, Kai-Wei
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION, ICCV, 2023, : 112 - 123
  • [49] Evaluating Semantic Textual Similarity in Clinical Sentences Using Deep Learning and Sentence Embeddings
    Antunes, Rui
    Silva, Joao Figueira
    Matos, Sergio
    PROCEEDINGS OF THE 35TH ANNUAL ACM SYMPOSIUM ON APPLIED COMPUTING (SAC'20), 2020, : 662 - 669
  • [50] Learning Sentence Embeddings based on Weighted Contexts from Unlabelled Data
    Ding, Yixin
    Xu, Liutong
    PROCEEDINGS OF 2018 IEEE 9TH INTERNATIONAL CONFERENCE ON SOFTWARE ENGINEERING AND SERVICE SCIENCE (ICSESS), 2018, : 793 - 796