Auxiliary Learning for Self-Supervised Video Representation via Similarity-based Knowledge Distillation

被引:2
|
作者
Dadashzadeh, Amirhossein [1 ]
Whone, Alan [1 ]
Mirmehdi, Majid [1 ]
机构
[1] Univ Bristol, Bristol, Avon, England
关键词
D O I
10.1109/CVPRW56347.2022.00468
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Despite the outstanding success of self-supervised pretraining methods for video representation learning, they generalise poorly when the unlabeled dataset for pretraining is small or the domain difference between unlabelled data in source task (pretraining) and labeled data in target task (finetuning) is significant. To mitigate these issues, we propose a novel approach to complement self-supervised pretraining via an auxiliary pretraining phase, based on knowledge similarity distillation, auxSKD, for better generalisation with a significantly smaller amount of video data, e.g. Kinetics-100 rather than Kinetics-400. Our method deploys a teacher network that iteratively distils its knowledge to the student model by capturing the similarity information between segments of unlabelled video data. The student model meanwhile solves a pretext task by exploiting this prior knowledge. We also introduce a novel pretext task, Video Segment Pace Prediction or VSPP, which requires our model to predict the playback speed of a randomly selected segment of the input video to provide more reliable self-supervised representations. Our experimental results show superior results to the state of the art on both UCF101 and HMDB51 datasets when pretraining on K100 in apple-to-apple comparisons. Additionally, we show that our auxiliary pretraining, auxSKD, when added as an extra pretraining phase to recent state of the art self-supervised methods (i.e. VCOP, VideoPace, and RSPNet), improves their results on UCF101 and HMDB51.
引用
收藏
页码:4230 / 4239
页数:10
相关论文
共 50 条
  • [1] Noise Suppression With Similarity-Based Self-Supervised Deep Learning
    Niu, Chuang
    Li, Mengzhou
    Fan, Fenglei
    Wu, Weiwen
    Guo, Xiaodong
    Lyu, Qing
    Wang, Ge
    [J]. IEEE TRANSACTIONS ON MEDICAL IMAGING, 2023, 42 (06) : 1590 - 1602
  • [2] Series2vec: similarity-based self-supervised representation learning for time series classification
    Foumani, Navid Mohammadi
    Tan, Chang Wei
    Webb, Geoffrey I.
    Rezatofighi, Hamid
    Salehi, Mahsa
    [J]. DATA MINING AND KNOWLEDGE DISCOVERY, 2024, 38 (04) : 2520 - 2544
  • [3] ISD: Self-Supervised Learning by Iterative Similarity Distillation
    Tejankar, Ajinkya
    Koohpayegani, Soroush Abbasi
    Pillai, Vipin
    Favaro, Paolo
    Pirsiavash, Hamed
    [J]. 2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 9589 - 9598
  • [4] Masked Video Distillation: Rethinking Masked Feature Modeling for Self-supervised Video Representation Learning
    Wang, Rui
    Chen, Dongdong
    Wu, Zuxuan
    Chen, Yinpeng
    Dai, Xiyang
    Liu, Mengchen
    Yuan, Lu
    Jiang, Yu-Gang
    [J]. 2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR, 2023, : 6312 - 6322
  • [5] SELF-SUPERVISED REPRESENTATION LEARNING FOR ULTRASOUND VIDEO
    Jiao, Jianbo
    Droste, Richard
    Drukker, Lior
    Papageorghiou, Aris T.
    Noble, J. Alison
    [J]. 2020 IEEE 17TH INTERNATIONAL SYMPOSIUM ON BIOMEDICAL IMAGING (ISBI 2020), 2020, : 1847 - 1850
  • [6] Self-supervised knowledge distillation in counterfactual learning for VQA
    Bi, Yandong
    Jiang, Huajie
    Zhang, Hanfu
    Hu, Yongli
    Yin, Baocai
    [J]. PATTERN RECOGNITION LETTERS, 2024, 177 : 33 - 39
  • [7] Image quality assessment based on self-supervised learning and knowledge distillation
    Sang, Qingbing
    Shu, Ziru
    Liu, Lixiong
    Hu, Cong
    Wu, Qin
    [J]. JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2023, 90
  • [8] Self-supervised heterogeneous graph learning with iterative similarity distillation
    Wang, Tianfeng
    Pan, Zhisong
    Hu, Guyu
    Xu, Kun
    Zhang, Yao
    [J]. KNOWLEDGE-BASED SYSTEMS, 2023, 276
  • [9] Self-supervised knowledge distillation for complementary label learning
    Liu, Jiabin
    Li, Biao
    Lei, Minglong
    Shi, Yong
    [J]. NEURAL NETWORKS, 2022, 155 : 318 - 327
  • [10] Multi-Mode Online Knowledge Distillation for Self-Supervised Visual Representation Learning
    Song, Kaiyou
    Xie, Jin
    Zhang, Shan
    Luo, Zimeng
    [J]. 2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 11848 - 11857