Self-Distilled Self-supervised Representation Learning

被引:2
|
作者
Jang, Jiho [1 ]
Kim, Seonhoon [2 ]
Yoo, Kiyoon [1 ]
Kong, Chaerin [1 ]
Kim, Jangho [3 ]
Kwak, Nojun [1 ]
机构
[1] Seoul Natl Univ, Seoul, South Korea
[2] Coupang, Seoul, South Korea
[3] Kookmin Univ, Seoul, South Korea
关键词
D O I
10.1109/WACV56688.2023.00285
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
State-of-the-art frameworks in self-supervised learning have recently shown that fully utilizing transformer-based models can lead to performance boost compared to conventional CNN models. Striving to maximize the mutual information of two views of an image, existing works apply a contrastive loss to the final representations. Motivated by self-distillation in the supervised regime, we further exploit this by allowing the intermediate representations to learn from the final layer via the contrastive loss. Through self-distillation, the intermediate layers are better suited for instance discrimination, making the performance of an earlyexited sub-network not much degraded from that of the full network. This renders the pretext task easier also for the final layer, leading to better representations. Our method, Self-Distilled Self-Supervised Learning (SDSSL), outperforms competitive baselines (SimCLR, BYOL and MoCo v3) using ViT on various tasks and datasets. In the linear evaluation and k-NN protocol, SDSSL not only leads to superior performance in the final layers, but also in most of the lower layers. Furthermore, qualitative and quantitative analyses show how representations are formed more effectively along the transformer layers. Code is available at https://github.com/hagiss/SDSSL.
引用
收藏
页码:2828 / 2838
页数:11
相关论文
共 50 条
  • [31] Self-Supervised Learning With Segmental Masking for Speech Representation
    Yue, Xianghu
    Lin, Jingru
    Gutierrez, Fabian Ritter
    Li, Haizhou
    [J]. IEEE Journal on Selected Topics in Signal Processing, 2022, 16 (06): : 1367 - 1379
  • [32] Randomly shuffled convolution for self-supervised representation learning
    Oh, Youngjin
    Jeon, Minkyu
    Ko, Dohwan
    Kim, Hyunwoo J.
    [J]. INFORMATION SCIENCES, 2023, 623 : 206 - 219
  • [33] Self-supervised representation learning for surgical activity recognition
    Paysan, Daniel
    Haug, Luis
    Bajka, Michael
    Oelhafen, Markus
    Buhmann, Joachim M.
    [J]. INTERNATIONAL JOURNAL OF COMPUTER ASSISTED RADIOLOGY AND SURGERY, 2021, 16 (11) : 2037 - 2044
  • [34] CCBERT: Self-Supervised Code Change Representation Learning
    Zhou, Xin
    Xu, Bowen
    Han, DongGyun
    Yang, Zhou
    He, Junda
    Lo, David
    [J]. 2023 IEEE INTERNATIONAL CONFERENCE ON SOFTWARE MAINTENANCE AND EVOLUTION, ICSME, 2023, : 182 - 193
  • [35] ViewMix: Augmentation for Robust Representation in Self-Supervised Learning
    Das, Arjon
    Zhong, Xin
    [J]. IEEE ACCESS, 2024, 12 : 8461 - 8470
  • [36] Self-supervised Consensus Representation Learning for Attributed Graph
    Liu, Changshu
    Wen, Liangjian
    Kang, Zhao
    Luo, Guangchun
    Tian, Ling
    [J]. PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2021, 2021, : 2654 - 2662
  • [37] Understanding Representation Learnability of Nonlinear Self-Supervised Learning
    Yang, Ruofeng
    Li, Xiangyuan
    Jiang, Bo
    Li, Shuai
    [J]. THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 9, 2023, : 10807 - 10815
  • [38] TRIBYOL: TRIPLET BYOL FOR SELF-SUPERVISED REPRESENTATION LEARNING
    Li, Guang
    Togo, Ren
    Ogawa, Takahiro
    Haseyama, Miki
    [J]. 2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 3458 - 3462
  • [39] Self-Supervised Fair Representation Learning without Demographics
    Chai, Junyi
    Wang, Xiaoqian
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [40] Self-Supervised Representation Learning for Video Quality Assessment
    Jiang, Shaojie
    Sang, Qingbing
    Hu, Zongyao
    Liu, Lixiong
    [J]. IEEE TRANSACTIONS ON BROADCASTING, 2023, 69 (01) : 118 - 129