Memory-Based Jitter: Improving Visual Recognition on Long-Tailed Data with Diversity in Memory

被引:0
|
作者
Liu, Jialun [1 ,2 ]
Li, Wenhui [1 ]
Sun, Yifan [2 ]
机构
[1] Jilin Univ, Key Lab Symbol Computat & Knowledge Engn, Changchun, Jilin, Peoples R China
[2] Baidu Res, Beijing, Peoples R China
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper considers deep visual recognition on long-tailed data. To make our method general, we tackle two applied scenarios, i.e., deep classification and deep metric learning. Under the long-tailed data distribution, the most classes (i.e., tail classes) only occupy relatively few samples and are prone to lack of within-class diversity. A radical solution is to augment the tail classes with higher diversity. To this end, we introduce a simple and reliable method named Memory-based Jitter (MBJ). We observe that during training, the deep model constantly changes its parameters after every iteration, yielding the phenomenon of weight jitters. Consequentially, given a same image as the input, two historical editions of the model generate two different features in the deeply-embedded space, resulting in feature jitters. Using a memory bank, we collect these (model or feature) jitters across multiple training iterations and get the so-called Memory-based Jitter. The accumulated jitters enhance the within-class diversity for the tail classes and consequentially improves long-tailed visual recognition. With slight modifications, MBJ is applicable for two fundamental visual recognition tasks, i.e., deep image classification and deep metric learning (on long-tailed data). Extensive experiments on five long-tailed classification benchmarks and two deep metric learning benchmarks demonstrate significant improvement. Moreover, the achieved performance are on par with the state of the art on both tasks.
引用
收藏
页码:1720 / 1728
页数:9
相关论文
共 50 条
  • [1] Inflated Episodic Memory with Region Self-Attention for Long-Tailed Visual Recognition
    Zhu, Linchao
    Yang, Yi
    [J]. 2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, : 4343 - 4352
  • [2] Improving Calibration for Long-Tailed Recognition
    Zhong, Zhisheng
    Cui, Jiequan
    Liu, Shu
    Jia, Jiaya
    [J]. 2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 16484 - 16493
  • [3] A Survey on Long-Tailed Visual Recognition
    Yang, Lu
    Jiang, He
    Song, Qing
    Guo, Jun
    [J]. INTERNATIONAL JOURNAL OF COMPUTER VISION, 2022, 130 (07) : 1837 - 1872
  • [4] A Survey on Long-Tailed Visual Recognition
    Lu Yang
    He Jiang
    Qing Song
    Jun Guo
    [J]. International Journal of Computer Vision, 2022, 130 : 1837 - 1872
  • [5] Memory-Based Neighbourhood Embedding for Visual Recognition
    Li, Suichan
    Chen, Dapeng
    Liu, Bin
    Yu, Nenghai
    Zhao, Rui
    [J]. 2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, : 6101 - 6110
  • [6] Decoupled Optimisation for Long-Tailed Visual Recognition
    Cong, Cong
    Xuan, Shiyu
    Liu, Sidong
    Zhang, Shiliang
    Pagnucco, Maurice
    Song, Yang
    [J]. THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 2, 2024, : 1380 - 1388
  • [7] BSDA in Visual Recognition: Balanced Semantic Data Augmentation for Long-Tailed Data
    Wang, Yifan
    Huang, Eaven
    Wang, Runan
    Leng, Tuo
    [J]. 2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN, 2023,
  • [8] Class-Difficulty Based Methods for Long-Tailed Visual Recognition
    Saptarshi Sinha
    Hiroki Ohashi
    Katsuyuki Nakamura
    [J]. International Journal of Computer Vision, 2022, 130 : 2517 - 2531
  • [9] Prototype-based classifier learning for long-tailed visual recognition
    Xiu-Shen Wei
    Shu-Lin Xu
    Hao Chen
    Liang Xiao
    Yuxin Peng
    [J]. Science China Information Sciences, 2022, 65
  • [10] Prototype-based classifier learning for long-tailed visual recognition
    Xiu-Shen WEI
    Shu-Lin XU
    Hao CHEN
    Liang XIAO
    Yuxin PENG
    [J]. Science China(Information Sciences), 2022, (06) : 62 - 76