MEID: Mixture-of-Experts with Internal Distillation for Long-Tailed Video Recognition

被引:0
|
作者
Li, Xinjie [1 ]
Xu, Huijuan [1 ]
机构
[1] Penn State Univ, University Pk, PA 16802 USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The long-tailed video recognition problem is especially challenging, as videos tend to be long and untrimmed, and each video may contain multiple classes, causing frame-level class imbalance. The previous method tackles the long-tailed video recognition only through frame-level sampling for class rebalance without distinguishing the frame-level feature representation between head and tail classes. To improve the frame-level feature representation of tail classes, we modulate the frame-level features with an auxiliary distillation loss to reduce the distribution distance between head and tail classes. Moreover, we design a mixture-of-experts framework with two different expert designs, i.e., the first expert with an attention-based classification network handling the original long-tailed distribution, and the second expert dealing with the re-balanced distribution from class-balanced sampling. Notably, in the second expert, we specifically focus on the frames unsolved by the first expert by designing a complementary frame selection module, which inherits the attention weights from the first expert and selects frames with low attention weights, and we also enhance the motion feature representation for these selected frames. To highlight the multi-label challenge in long-tailed video recognition, we create two additional benchmarks based on Charades and CharadesEgo videos with the multi-label property, called CharadesLT and CharadesEgoLT. Extensive experiments are conducted on the existing long-tailed video benchmark VideoLT and the two new benchmarks to verify the effectiveness of our proposed method with state-of-the-art performance. The code and proposed benchmarks are released at https://github.com/VisionLanguageLab/MEID.
引用
收藏
页码:1451 / 1459
页数:9
相关论文
共 50 条
  • [41] Distilling Virtual Examples for Long-tailed Recognition
    He, Yin-Yin
    Wu, Jianxin
    Wei, Xiu-Shen
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 235 - 244
  • [42] Open Long-Tailed Recognition in a Dynamic World
    Liu, Ziwei
    Miao, Zhongqi
    Zhan, Xiaohang
    Wang, Jiayun
    Gong, Boqing
    Yu, Stella X.
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2024, 46 (03) : 1836 - 1851
  • [43] Normalizing Batch Normalization for Long-Tailed Recognition
    Bao, Yuxiang
    Kang, Guoliang
    Yang, Linlin
    Duan, Xiaoyue
    Zhao, Bo
    Zhang, Baochang
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2025, 34 : 209 - 220
  • [44] Minority-Oriented Vicinity Expansion with Attentive Aggregation for Video Long-Tailed Recognition
    Moon, WonJun
    Seong, Hyun Seok
    Heo, Jae-Pil
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 2, 2023, : 1931 - 1939
  • [45] FoPro-KD: Fourier Prompted Effective Knowledge Distillation for Long-Tailed Medical Image Recognition
    Elbatel, Marawan
    Marti, Robert
    Li, Xiaomeng
    IEEE TRANSACTIONS ON MEDICAL IMAGING, 2024, 43 (03) : 954 - 965
  • [46] One-stage self-distillation guided knowledge transfer for long-tailed visual recognition
    Xia, Yuelong
    Zhang, Shu
    Wang, Jun
    Zou, Wei
    Zhou, Juxiang
    Wen, Bin
    INTERNATIONAL JOURNAL OF INTELLIGENT SYSTEMS, 2022, 37 (12) : 11893 - 11908
  • [47] Teaching Teachers First and Then Student: Hierarchical Distillation to Improve Long-Tailed Object Recognition in Aerial Images
    Zhao, Wenda
    Liu, Jiani
    Liu, Yu
    Zhao, Fan
    He, You
    Lu, Huchuan
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2022, 60
  • [48] Long-Tailed Multi-label Retinal Diseases Recognition via Relational Learning and Knowledge Distillation
    Zhou, Qian
    Zou, Hua
    Wang, Zhongyuan
    MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION, MICCAI 2022, PT II, 2022, 13432 : 709 - 718
  • [49] Class-Balanced Regularization for Long-Tailed Recognition
    Xu, Yuge
    Lyu, Chuanlong
    NEURAL PROCESSING LETTERS, 2024, 56 (03)
  • [50] Feature fusion network for long-tailed visual recognition
    Zhou, Xuesong
    Zhai, Junhai
    Cao, Yang
    PATTERN RECOGNITION, 2023, 144