One-stage self-distillation guided knowledge transfer for long-tailed visual recognition

被引:1
|
作者
Xia, Yuelong [1 ,2 ,3 ]
Zhang, Shu [1 ,2 ,3 ]
Wang, Jun [2 ,3 ]
Zou, Wei [1 ,2 ,3 ]
Zhou, Juxiang [2 ,3 ]
Wen, Bin [1 ,2 ,3 ]
机构
[1] Yunnan Normal Univ, Sch Informat Sci & Technol, Kunming, Peoples R China
[2] Yunnan Normal Univ, Minist Educ, Key Lab Educ Informatizat Nationalities, Kunming, Peoples R China
[3] Yunnan Normal Univ, Yunnan Key Lab Smart Educ, Kunming, Peoples R China
基金
中国国家自然科学基金;
关键词
knowledge transfer; long-tailed recognition; one-stage training; self-distillation;
D O I
10.1002/int.23068
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep learning has achieved remarkable progress for visual recognition on balanced data sets but still performs poorly on real-world long-tailed data distribution. The existing methods mainly decouple the problem into the two-stage decoupling training, that is, representation learning and classifier training, or multistage training based on knowledge distillation, thus resulting in huge training steps and extra computation cost. In this paper, we propose a conceptually simple yet effective One-stage Long-tailed Self-Distillation framework, called OLSD, which simultaneously takes representation learning and classifier training into one-stage training. For representation learning, we take two different sampling distributions and mixup them to input them into two branches, where the collaborative consistency loss is introduced to train network consistency, and we theoretically show that the proposed mixup naturally generates a tail-majority distribution mixup. For classifier training, we introduce balanced self-distillation guided knowledge transfer to improve generalization performance, where we theoretically show that proposed knowledge transfer implicitly minimizes not only cross-entropy but also KL divergence between head-to-tail and tail-to-head. Extensive experiments on long-tailed CIFAR10/100, ImageNet-LT and multilabel long-tailed VOC-LT demonstrate the proposed method's effectiveness.
引用
收藏
页码:11893 / 11908
页数:16
相关论文
共 50 条
  • [41] Semantic Transfer from Head to Tail: Enlarging Tail Margin for Long-Tailed Visual Recognition
    Zhang, Shan
    Ni, Yao
    Du, Jinhao
    Liu, Yanxia
    Koniusz, Piotr
    2024 IEEE/CVF WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION, WACV 2024, 2024, : 1339 - 1349
  • [42] GistNet: a Geometric Structure Transfer Network for Long-Tailed Recognition
    Liu, Bo
    Li, Haoxiang
    Kang, Hao
    Hua, Gang
    Vasconcelos, Nuno
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 8189 - 8198
  • [43] MEID: Mixture-of-Experts with Internal Distillation for Long-Tailed Video Recognition
    Li, Xinjie
    Xu, Huijuan
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 2, 2023, : 1451 - 1459
  • [44] Bilinear-experts network with self-adaptive sampler for long-tailed visual recognition
    Wang, Qin
    Kwong, Sam
    Wang, Xizhao
    NEUROCOMPUTING, 2025, 633
  • [45] Weight-guided class complementing for long-tailed image recognition
    Zhao, Xinqiao
    Xiao, Jimin
    Yu, Siyue
    Li, Hui
    Zhang, Bingfeng
    PATTERN RECOGNITION, 2023, 138
  • [46] Open-Vocabulary One-Stage Detection with Hierarchical Visual-Language Knowledge Distillation
    Mal, Zongyang
    Luo, Guan
    Gao, Jin
    Li, Liang
    Chen, Yuxin
    Wang, Shaoru
    Zhang, Congxuan
    Hu, Weiming
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, : 14054 - 14063
  • [47] SAR Image Classification with Knowledge Distillation and Class Balancing for Long-Tailed Distributions
    Jahan, Chowdhury Sadman
    Savakis, Andreas
    Blasch, Erik
    2022 IEEE 14TH IMAGE, VIDEO, AND MULTIDIMENSIONAL SIGNAL PROCESSING WORKSHOP (IVMSP), 2022,
  • [48] Self-distillation object segmentation via pyramid knowledge representation and transfer
    Zheng, Yunfei
    Sun, Meng
    Wang, Xiaobing
    Cao, Tieyong
    Zhang, Xiongwei
    Xing, Lixing
    Fang, Zheng
    MULTIMEDIA SYSTEMS, 2023, 29 (05) : 2615 - 2631
  • [49] Long-Tailed Recognition Based on Self-attention Mechanism
    Feng, Zekai
    Jia, Hong
    Li, Mengke
    ADVANCED INTELLIGENT COMPUTING TECHNOLOGY AND APPLICATIONS, PT II, ICIC 2024, 2024, 14876 : 380 - 391
  • [50] Self-distillation object segmentation via pyramid knowledge representation and transfer
    Yunfei Zheng
    Meng Sun
    Xiaobing Wang
    Tieyong Cao
    Xiongwei Zhang
    Lixing Xing
    Zheng Fang
    Multimedia Systems, 2023, 29 : 2615 - 2631