MVEB: Self-Supervised Learning With Multi-View Entropy Bottleneck

被引:0
|
作者
Wen, Liangjian [1 ,2 ]
Wang, Xiasi [3 ]
Liu, Jianzhuang [4 ]
Xu, Zenglin [5 ,6 ]
机构
[1] Southwestern Univ Finance & Econ, Sch Comp & Artificial Intelligence, Chengdu 610074, Peoples R China
[2] Southwestern Univ Finance & Econ, Res Inst Digital Econ & Interdisciplinary Sci, Chengdu 610074, Peoples R China
[3] Hong Kong Univ Sci & Technol, Hong Kong, Peoples R China
[4] Shenzhen Inst Adv Technol, Shenzhen 518055, Peoples R China
[5] Harbin Inst Technol Shenzhen, Shenzhen 150001, Peoples R China
[6] Pengcheng Lab, Shenzhen 518066, Peoples R China
关键词
Task analysis; Entropy; Mutual information; Supervised learning; Feature extraction; Minimal sufficient representation; representation learning; self-supervised learning;
D O I
10.1109/TPAMI.2024.3380065
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Self-supervised learning aims to learn representation that can be effectively generalized to downstream tasks. Many self-supervised approaches regard two views of an image as both the input and the self-supervised signals, assuming that either view contains the same task-relevant information and the shared information is (approximately) sufficient for predicting downstream tasks. Recent studies show that discarding superfluous information not shared between the views can improve generalization. Hence, the ideal representation is sufficient for downstream tasks and contains minimal superfluous information, termed minimal sufficient representation. One can learn this representation by maximizing the mutual information between the representation and the supervised view while eliminating superfluous information. Nevertheless, the computation of mutual information is notoriously intractable. In this work, we propose an objective termed multi-view entropy bottleneck (MVEB) to learn minimal sufficient representation effectively. MVEB simplifies the minimal sufficient learning to maximizing both the agreement between the embeddings of two views and the differential entropy of the embedding distribution. Our experiments confirm that MVEB significantly improves performance. For example, it achieves top-1 accuracy of 76.9% on ImageNet with a vanilla ResNet-50 backbone on linear evaluation. To the best of our knowledge, this is the new state-of-the-art result with ResNet-50.
引用
收藏
页码:6097 / 6108
页数:12
相关论文
共 50 条
  • [1] Self-supervised learning for multi-view stereo
    Ito S.
    Kaneko N.
    Sumi K.
    Seimitsu Kogaku Kaishi/Journal of the Japan Society for Precision Engineering, 2020, 86 (12): : 1042 - 1050
  • [2] Self-Supervised Information Bottleneck for Deep Multi-View Subspace Clustering
    Wang, Shiye
    Li, Changsheng
    Li, Yanming
    Yuan, Ye
    Wang, Guoren
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2023, 32 (1555-1567) : 1555 - 1567
  • [3] Self-Supervised Representations for Multi-View Reinforcement Learning
    Yang, Huanhuan
    Shi, Dianxi
    Xie, Guojun
    Peng, Yingxuan
    Zhang, Yi
    Yang, Yantai
    Yang, Shaowu
    UNCERTAINTY IN ARTIFICIAL INTELLIGENCE, VOL 180, 2022, 180 : 2203 - 2213
  • [4] Self-supervised Learning of Depth Inference for Multi-view Stereo
    Yang, Jiayu
    Alvarez, Jose M.
    Liu, Miaomiao
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 7522 - 7530
  • [5] Multi-view self-supervised learning on heterogeneous graphs for recommendation
    Zhang, Yunjia
    Zhang, Yihao
    Liao, Weiwen
    Li, Xiaokang
    Wang, Xibin
    APPLIED SOFT COMPUTING, 2025, 174
  • [6] Multi-view and multi-augmentation for self-supervised visual representation learning
    Tran, Van Nhiem
    Huang, Chi-En
    Liu, Shen-Hsuan
    Aslam, Muhammad Saqlain
    Yang, Kai-Lin
    Li, Yung-Hui
    Wang, Jia-Ching
    APPLIED INTELLIGENCE, 2024, 54 (01) : 629 - 656
  • [7] Multi-view and multi-augmentation for self-supervised visual representation learning
    Van Nhiem Tran
    Chi-En Huang
    Shen-Hsuan Liu
    Muhammad Saqlain Aslam
    Kai-Lin Yang
    Yung-Hui Li
    Jia-Ching Wang
    Applied Intelligence, 2024, 54 : 629 - 656
  • [8] Self-Supervised Discriminative Feature Learning for Deep Multi-View Clustering
    Xu, Jie
    Ren, Yazhou
    Tang, Huayi
    Yang, Zhimeng
    Pan, Lili
    Yang, Yang
    Pu, Xiaorong
    Yu, Philip S.
    He, Lifang
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2023, 35 (07) : 7470 - 7482
  • [9] Multi-View Collaborative Training and Self-Supervised Learning for Group Recommendation
    Wei, Feng
    Chen, Shuyu
    MATHEMATICS, 2025, 13 (01)
  • [10] Multi-view Self-supervised Heterogeneous Graph Embedding
    Zhao, Jianan
    Wen, Qianlong
    Sun, Shiyu
    Ye, Yanfang
    Zhang, Chuxu
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2021: RESEARCH TRACK, PT II, 2021, 12976 : 319 - 334