Contrastive Unsupervised Representation Learning With Optimize-Selected Training Samples

被引:0
|
作者
Cheng, Yujun [1 ]
Zhang, Zhewei [1 ]
Li, Xuejing [1 ]
Wang, Shengjin [1 ]
机构
[1] Tsinghua Univ, Dept Elect Engn, Beijing 100084, Peoples R China
关键词
Contrastive learning; PAC-Bayes generalization; unsupervised representation learning;
D O I
10.1109/TNNLS.2024.3424331
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Contrastive unsupervised representation learning (CURL) is a technique that seeks to learn feature sets from unlabeled data. It has found widespread and successful application in unsupervised feature learning, with the design of positive and negative pairs serving as the type of data samples. While CURL has seen empirical successes in recent years, there is still room for improvement in terms of the pair data generation process. This includes tasks such as combining and re-filtering samples, or implementing transformations among positive/negative pairs. We refer to this as the sample selection process. In this article, we introduce an optimized pair-data sample selection method for CURL. This method efficiently ensures that the two types of sampled data (similar pair and dissimilar pair) do not belong to the same class. We provide a theoretical analysis to demonstrate why our proposed method enhances learning performance by analyzing its error probability. Furthermore, we extend our proof into PAC-Bayes generalization to illustrate how our method tightens the bounds provided in previous literature. Our numerical experiments on text/image datasets show that our method achieves competitive accuracy with good generalization bounds.
引用
收藏
页数:11
相关论文
共 50 条
  • [41] Extending Contrastive Learning to Unsupervised Coreset Selection
    Ju, Jeongwoo
    Jung, Heechul
    Oh, Yoonju
    Kim, Junmo
    IEEE ACCESS, 2022, 10 : 7704 - 7715
  • [42] Extending Contrastive Learning to Unsupervised Redundancy Identification
    Ju, Jeongwoo
    Jung, Heechul
    Kim, Junmo
    APPLIED SCIENCES-BASEL, 2022, 12 (04):
  • [43] Contrastive Learning for Unsupervised Video Highlight Detection
    Badamdorj, Taivanbat
    Rochan, Mrigank
    Wang, Yang
    Cheng, Li
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, : 14022 - 14032
  • [44] Contrastive Intrinsic Control for Unsupervised Reinforcement Learning
    Laskin, Michael
    Liu, Hao
    Peng, Xue Bin
    Yarats, Denis
    Rajeswaran, Aravind
    Abbeel, Pieter
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [45] Contrastive Learning for Unsupervised Radar Place Recognition
    Gadd, Matthew
    De Martini, Daniele
    Newman, Paul
    2021 20TH INTERNATIONAL CONFERENCE ON ADVANCED ROBOTICS (ICAR), 2021, : 344 - 349
  • [46] An interpretable unsupervised capsule network via comprehensive contrastive learning and two-stage training
    Zeng, Ru
    Song, Yan
    Zhong, Yanjiu
    Pattern Recognition, 2025, 158
  • [47] HSimCSE: Improving Contrastive Learning of Unsupervised Sentence Representation with Adversarial Hard Positives and Dual Hard Negatives
    Xu, Bo
    Wei, Shouang
    Cheng, Luyi
    Huang, Shizhou
    Song, Hui
    Du, Ming
    Wang, Hongya
    2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN, 2023,
  • [48] PARCEL: Physics-Based Unsupervised Contrastive Representation Learning for Multi-Coil MR Imaging
    Wang, Shanshan
    Wu, Ruoyou
    Li, Cheng
    Zou, Juan
    Zhang, Ziyao
    Liu, Qiegen
    Xi, Yan
    Zheng, Hairong
    IEEE-ACM TRANSACTIONS ON COMPUTATIONAL BIOLOGY AND BIOINFORMATICS, 2023, 20 (05) : 2659 - 2670
  • [49] Learning Contrastive Representation for Semantic Correspondence
    Taihong Xiao
    Sifei Liu
    Shalini De Mello
    Zhiding Yu
    Jan Kautz
    Ming-Hsuan Yang
    International Journal of Computer Vision, 2022, 130 : 1293 - 1309
  • [50] LoCo: Local Contrastive Representation Learning
    Xiong, Yuwen
    Ren, Mengye
    Urtasun, Raquel
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33