Self-Supervised Learning with an Information Maximization Criterion

被引:0
|
作者
Ozsoy, Serdar [1 ,2 ]
Hamdan, Shadi [1 ,3 ]
Arik, Sercan O. [4 ]
Yuret, Deniz [1 ,3 ]
Erdogan, Alper T. [1 ,2 ]
机构
[1] Koc Univ, KUIS AI Ctr, Istanbul, Turkiye
[2] Koc Univ, EEE Dept, Istanbul, Turkiye
[3] Koc Univ, CE Dept, Istanbul, Turkiye
[4] Google Cloud Res, Sunnyvale, CA USA
关键词
MATRIX FACTORIZATION; FRAMEWORK;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Self-supervised learning allows AI systems to learn effective representations from large amounts of data using tasks that do not require costly labeling. Mode collapse, i.e., the model producing identical representations for all inputs, is a central problem to many self-supervised learning approaches, making self-supervised tasks, such as matching distorted variants of the inputs, ineffective. In this article, we argue that a straightforward application of information maximization among alternative latent representations of the same input naturally solves the collapse problem and achieves competitive empirical results. We propose a self-supervised learning method, CorInfoMax, that uses a second-order statistics-based mutual information measure that reflects the level of correlation among its arguments. Maximizing this correlative information measure between alternative representations of the same input serves two purposes: (1) it avoids the collapse problem by generating feature vectors with non-degenerate covariances; (2) it establishes relevance among alternative representations by increasing the linear dependence among them. An approximation of the proposed information maximization objective simplifies to a Euclidean distance-based objective function regularized by the log-determinant of the feature covariance matrix. The regularization term acts as a natural barrier against feature space degeneracy. Consequently, beyond avoiding complete output collapse to a single point, the proposed approach also prevents dimensional collapse by encouraging the spread of information across the whole feature space. Numerical experiments demonstrate that CorInfoMax achieves better or competitive performance results relative to the state-of-the-art SSL approaches.
引用
收藏
页数:14
相关论文
共 50 条
  • [1] Self-Supervised Learning with Kernel Dependence Maximization
    Li, Yazhe
    Pogodin, Roman
    Sutherland, Danica J.
    Gretton, Arthur
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [2] Label-Efficient Self-Supervised Speaker Verification With Information Maximization and Contrastive Learning
    Lepage, Theo
    Dehak, Reda
    [J]. INTERSPEECH 2022, 2022, : 4018 - 4022
  • [3] S3-Rec: Self-Supervised Learning for Sequential Recommendation with Mutual Information Maximization
    Zhou, Kun
    Wang, Hui
    Zhao, Wayne Xin
    Zhu, Yutao
    Wang, Sirui
    Zhang, Fuzheng
    Wang, Zhongyuan
    Wen, Ji-Rong
    [J]. CIKM '20: PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT, 2020, : 1893 - 1902
  • [4] Boundary-aware information maximization for self-supervised medical image segmentation
    Peng, Jizong
    Wang, Ping
    Pedersoli, Marco
    Desrosiers, Christian
    [J]. MEDICAL IMAGE ANALYSIS, 2024, 94
  • [5] ADDING DISTANCE INFORMATION TO SELF-SUPERVISED LEARNING FOR RICH REPRESENTATIONS
    Kim, Yeji
    Kong, Bai-Sun
    [J]. 2023 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2023, : 1270 - 1274
  • [6] Self-supervised video representation learning by maximizing mutual information
    Xue, Fei
    Ji, Hongbing
    Zhang, Wenbo
    Cao, Yi
    [J]. SIGNAL PROCESSING-IMAGE COMMUNICATION, 2020, 88
  • [7] Self-Supervised Graph Representation Learning via Information Bottleneck
    Gu, Junhua
    Zheng, Zichen
    Zhou, Wenmiao
    Zhang, Yajuan
    Lu, Zhengjun
    Yang, Liang
    [J]. SYMMETRY-BASEL, 2022, 14 (04):
  • [8] Self-Supervised Dialogue Learning
    Wu, Jiawei
    Wang, Xin
    Wang, William Yang
    [J]. 57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), 2019, : 3857 - 3867
  • [9] Longitudinal self-supervised learning
    Zhao, Qingyu
    Liu, Zixuan
    Adeli, Ehsan
    Pohl, Kilian M.
    [J]. MEDICAL IMAGE ANALYSIS, 2021, 71
  • [10] Self-Supervised Learning for Electroencephalography
    Rafiei, Mohammad H.
    Gauthier, Lynne V.
    Adeli, Hojjat
    Takabi, Daniel
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (02) : 1457 - 1471