One-stage self-distillation guided knowledge transfer for long-tailed visual recognition

被引:1
|
作者
Xia, Yuelong [1 ,2 ,3 ]
Zhang, Shu [1 ,2 ,3 ]
Wang, Jun [2 ,3 ]
Zou, Wei [1 ,2 ,3 ]
Zhou, Juxiang [2 ,3 ]
Wen, Bin [1 ,2 ,3 ]
机构
[1] Yunnan Normal Univ, Sch Informat Sci & Technol, Kunming, Peoples R China
[2] Yunnan Normal Univ, Minist Educ, Key Lab Educ Informatizat Nationalities, Kunming, Peoples R China
[3] Yunnan Normal Univ, Yunnan Key Lab Smart Educ, Kunming, Peoples R China
基金
中国国家自然科学基金;
关键词
knowledge transfer; long-tailed recognition; one-stage training; self-distillation;
D O I
10.1002/int.23068
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep learning has achieved remarkable progress for visual recognition on balanced data sets but still performs poorly on real-world long-tailed data distribution. The existing methods mainly decouple the problem into the two-stage decoupling training, that is, representation learning and classifier training, or multistage training based on knowledge distillation, thus resulting in huge training steps and extra computation cost. In this paper, we propose a conceptually simple yet effective One-stage Long-tailed Self-Distillation framework, called OLSD, which simultaneously takes representation learning and classifier training into one-stage training. For representation learning, we take two different sampling distributions and mixup them to input them into two branches, where the collaborative consistency loss is introduced to train network consistency, and we theoretically show that the proposed mixup naturally generates a tail-majority distribution mixup. For classifier training, we introduce balanced self-distillation guided knowledge transfer to improve generalization performance, where we theoretically show that proposed knowledge transfer implicitly minimizes not only cross-entropy but also KL divergence between head-to-tail and tail-to-head. Extensive experiments on long-tailed CIFAR10/100, ImageNet-LT and multilabel long-tailed VOC-LT demonstrate the proposed method's effectiveness.
引用
收藏
页码:11893 / 11908
页数:16
相关论文
共 50 条
  • [21] Nested Collaborative Learning for Long-Tailed Visual Recognition
    Li, Jun
    Tan, Zichang
    Wan, Jun
    Lei, Zhen
    Guo, Guodong
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 6939 - 6948
  • [22] Probabilistic Contrastive Learning for Long-Tailed Visual Recognition
    Du, Chaoqun
    Wang, Yulin
    Song, Shiji
    Huang, Gao
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2024, 46 (09) : 5890 - 5904
  • [23] Balanced Contrastive Learning for Long-Tailed Visual Recognition
    Zhu, Jianggang
    Wang, Zheng
    Chen, Jingjing
    Chen, Yi-Ping Phoebe
    Jiang, Yu-Gang
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 6898 - 6907
  • [24] Exploring the auxiliary learning for long-tailed visual recognition
    Zhang, Junjie
    Liu, Lingqiao
    Wang, Peng
    Zhang, Jian
    NEUROCOMPUTING, 2021, 449 : 303 - 314
  • [25] Mitigating biases in long-tailed recognition via semantic-guided feature transfer
    Shi, Sheng
    Wang, Peng
    Zhang, Xinfeng
    Fan, Jianping
    NEUROCOMPUTING, 2024, 590
  • [26] Inflated Episodic Memory with Region Self-Attention for Long-Tailed Visual Recognition
    Zhu, Linchao
    Yang, Yi
    2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, : 4343 - 4352
  • [27] FoPro-KD: Fourier Prompted Effective Knowledge Distillation for Long-Tailed Medical Image Recognition
    Elbatel, Marawan
    Marti, Robert
    Li, Xiaomeng
    IEEE TRANSACTIONS ON MEDICAL IMAGING, 2024, 43 (03) : 954 - 965
  • [28] Long-Tailed Multi-label Retinal Diseases Recognition via Relational Learning and Knowledge Distillation
    Zhou, Qian
    Zou, Hua
    Wang, Zhongyuan
    MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION, MICCAI 2022, PT II, 2022, 13432 : 709 - 718
  • [29] Text-guided Fourier Augmentation for long-tailed recognition
    Wang, Weiqiu
    Chen, Zining
    Su, Fei
    Zhao, Zhicheng
    PATTERN RECOGNITION LETTERS, 2024, 179 : 38 - 44
  • [30] Balanced knowledge distillation for one-stage object detector
    Lee, Sungwook
    Lee, Seunghyun
    Song, Byung Cheol
    NEUROCOMPUTING, 2022, 500 : 394 - 404