One-stage self-distillation guided knowledge transfer for long-tailed visual recognition

被引:1
|
作者
Xia, Yuelong [1 ,2 ,3 ]
Zhang, Shu [1 ,2 ,3 ]
Wang, Jun [2 ,3 ]
Zou, Wei [1 ,2 ,3 ]
Zhou, Juxiang [2 ,3 ]
Wen, Bin [1 ,2 ,3 ]
机构
[1] Yunnan Normal Univ, Sch Informat Sci & Technol, Kunming, Peoples R China
[2] Yunnan Normal Univ, Minist Educ, Key Lab Educ Informatizat Nationalities, Kunming, Peoples R China
[3] Yunnan Normal Univ, Yunnan Key Lab Smart Educ, Kunming, Peoples R China
基金
中国国家自然科学基金;
关键词
knowledge transfer; long-tailed recognition; one-stage training; self-distillation;
D O I
10.1002/int.23068
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep learning has achieved remarkable progress for visual recognition on balanced data sets but still performs poorly on real-world long-tailed data distribution. The existing methods mainly decouple the problem into the two-stage decoupling training, that is, representation learning and classifier training, or multistage training based on knowledge distillation, thus resulting in huge training steps and extra computation cost. In this paper, we propose a conceptually simple yet effective One-stage Long-tailed Self-Distillation framework, called OLSD, which simultaneously takes representation learning and classifier training into one-stage training. For representation learning, we take two different sampling distributions and mixup them to input them into two branches, where the collaborative consistency loss is introduced to train network consistency, and we theoretically show that the proposed mixup naturally generates a tail-majority distribution mixup. For classifier training, we introduce balanced self-distillation guided knowledge transfer to improve generalization performance, where we theoretically show that proposed knowledge transfer implicitly minimizes not only cross-entropy but also KL divergence between head-to-tail and tail-to-head. Extensive experiments on long-tailed CIFAR10/100, ImageNet-LT and multilabel long-tailed VOC-LT demonstrate the proposed method's effectiveness.
引用
收藏
页码:11893 / 11908
页数:16
相关论文
共 50 条
  • [1] Balanced self-distillation for long-tailed recognition
    Ren, Ning
    Li, Xiaosong
    Wu, Yanxia
    Fu, Yan
    KNOWLEDGE-BASED SYSTEMS, 2024, 290
  • [2] Self Supervision to Distillation for Long-Tailed Visual Recognition
    Li, Tianhao
    Wang, Limin
    Wu, Gangshan
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 610 - 619
  • [3] MDCS: More Diverse Experts with Consistency Self-distillation for Long-tailed Recognition
    Zhao, Qihao
    Jiang, Chen
    Hu, Wei
    Zhang, Fan
    Liu, Jun
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 11563 - 11574
  • [4] Virtual Student Distribution Knowledge Distillation for Long-Tailed Recognition
    Liu, Haodong
    Huang, Xinlei
    Tang, Jialiang
    Jiang, Ning
    PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2024, PT IV, 2025, 15034 : 406 - 419
  • [5] Dynamic collaborative learning with heterogeneous knowledge transfer for long-tailed visual recognition
    Zhou, Hao
    Luo, Tingjin
    He, Yongming
    INFORMATION FUSION, 2025, 115
  • [6] KDTM: Multi-Stage Knowledge Distillation Transfer Model for Long-Tailed DGA Detection
    Fan, Baoyu
    Ma, Han
    Liu, Yue
    Yuan, Xiaochen
    Ke, Wei
    MATHEMATICS, 2024, 12 (05)
  • [7] Relational Subsets Knowledge Distillation for Long-Tailed Retinal Diseases Recognition
    Ju, Lie
    Wang, Xin
    Wang, Lin
    Liu, Tongliang
    Zhao, Xin
    Drummond, Tom
    Mahapatra, Dwarikanath
    Ge, Zongyuan
    MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION - MICCAI 2021, PT VIII, 2021, 12908 : 3 - 12
  • [8] Long-Tailed Visual Recognition via Self-Heterogeneous Integration with Knowledge Excavation
    Jin, Yan
    Li, Mengke
    Lu, Yang
    Cheung, Yiu-ming
    Wang, Hanzi
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 23695 - 23704
  • [9] Balanced knowledge distillation for long-tailed learning
    Zhang, Shaoyu
    Chen, Chen
    Hu, Xiyuan
    Peng, Silong
    NEUROCOMPUTING, 2023, 527 : 36 - 46
  • [10] A Survey on Long-Tailed Visual Recognition
    Yang, Lu
    Jiang, He
    Song, Qing
    Guo, Jun
    INTERNATIONAL JOURNAL OF COMPUTER VISION, 2022, 130 (07) : 1837 - 1872