Disturbed Augmentation Invariance for Unsupervised Visual Representation Learning

被引:3
|
作者
Cheng, Haoyang [1 ]
Li, Hongliang [1 ]
Wu, Qingbo [1 ]
Qiu, Heqian [1 ]
Zhang, Xiaoliang [1 ]
Meng, Fanman [1 ]
Zhao, Taijin [1 ]
机构
[1] Univ Elect Sci & Technol China UESTC, Sch Informat & Commun Engn, Chengdu 611731, Peoples R China
基金
中国国家自然科学基金;
关键词
Unsupervised learning; self-supervised learning; representation learning; contrastive learning; convolutional neural network;
D O I
10.1109/TCSVT.2023.3272741
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Contrastive learning has gained great prominence recently, which achieves excellent performance by simple augmentation invariance. However, the simple contrastive pairs suffer from lacking of diversity due to the mechanical augmentation strategies. In this paper, we propose Disturbed Augmentation Invariance (DAI for abbreviation), which constructs disturbed contrastive pairs by generating appropriate disturbed views for each augmented view in the feature space to increase the diversity. In practice, we establish a multivariate normal distribution for each augmented view, whose mean is corresponding augmented view and covariance matrix is estimated from its nearest neighbors in the dataset. Then we sample random vectors from this distribution as the disturbed views to construct disturbed contrastive pairs. In order to avoid extra computational cost with the increase of disturbed contrastive pairs, we utilize an upper bound of the trivial disturbed augmentation invariance loss to construct the DAI loss. In addition, we propose Bottleneck version of Disturbed Augmentation Invariance (BDAI for abbreviation) inspired by the Information Bottleneck principle, which further refines the extracted information and learns a compact representation by additionally increasing the variance of the original contrastive pair. In order to make BDAI work effectively, we design a statistical strategy to control the balance between the amount of the information shared by all disturbed contrastive pairs and the compactness of the representation. Our approach gets a consistent improvement over the popular contrastive learning methods on a variety of downstream tasks, e.g. image classification, object detection and instance segmentation.
引用
收藏
页码:6924 / 6938
页数:15
相关论文
共 50 条
  • [21] MEJIGCLU: MORE EFFECTIVE JIGSAW CLUSTERING FOR UNSUPERVISED VISUAL REPRESENTATION LEARNING
    Zhang, Yongsheng
    Liu, Qing
    Zhao, Yang
    Liang, Yixiong
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 2135 - 2139
  • [22] Unsupervised Visual Representation Learning by Online Constrained K-Means
    Qian, Qi
    Xu, Yuanhong
    Hu, Juhua
    Li, Hao
    Jin, Rong
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 16619 - 16628
  • [23] Unsupervised Visual Representation Learning via Mutual Information Regularized Assignment
    Lee, Dong Hoon
    Choi, Sungik
    Kim, Hyunwoo
    Chung, Sae-Young
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [24] Unsupervised Visual Representation Learning by Graph-Based Consistent Constraints
    Li, Dong
    Hung, Wei-Chih
    Huang, Jia-Bin
    Wang, Shengjin
    Ahuja, Narendra
    Yang, Ming-Hsuan
    COMPUTER VISION - ECCV 2016, PT IV, 2016, 9908 : 678 - 694
  • [25] SeA: Semantic Adversarial Augmentation for Last Layer Features from Unsupervised Representation Learning
    Qian, Qi
    Xu, Yuanhong
    Hui, Juhua
    COMPUTER VISION - ECCV 2024, PT LXXVIII, 2025, 15136 : 1 - 17
  • [26] On invariance and selectivity in representation learning
    Anselmi, Fabio
    Rosasco, Lorenzo
    Poggio, Tomaso
    INFORMATION AND INFERENCE-A JOURNAL OF THE IMA, 2016, 5 (02) : 134 - 158
  • [27] Explicitly learning augmentation invariance for image classification by Consistent Augmentation
    Li, Xiaosong
    Wu, Yanxia
    Tang, Chuheng
    Fu, Yan
    Zhang, Lidan
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2024, 130
  • [28] Un-mix: Rethinking Image Mixtures for Unsupervised Visual Representation Learning
    Shen, Zhiqiang
    Liu, Zechun
    Liu, Zhuang
    Savvides, Marios
    Darrell, Trevor
    Xing, Eric
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / THE TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 2216 - 2224
  • [29] Invariant and consistent: Unsupervised representation learning for few-shot visual recognition
    Wu, Heng
    Zhao, Yifan
    Li, Jia
    NEUROCOMPUTING, 2023, 520 : 1 - 14
  • [30] Unsupervised Visual Representation Learning via Multi-Dimensional Relationship Alignment
    Cheng, Haoyang
    Li, Hongliang
    Qiu, Heqian
    Wu, Qingbo
    Zhang, Xiaoliang
    Meng, Fanman
    Ngan, King Ngi
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2023, 32 : 1613 - 1626