Meta label associated loss for fine-grained visual recognition

被引:0
|
作者
Li, Yanchao [1 ]
Xiao, Fu [1 ]
Li, Hao [2 ]
Li, Qun [1 ]
Yu, Shui [3 ]
机构
[1] Nanjing Univ Posts & Telecommun, Sch Comp Sci, Nanjing 210023, Peoples R China
[2] Zhoukou Normal Univ, Sch Network Engn, Zhoukou 466001, Peoples R China
[3] Univ Technol Sydney, Sch Comp Sci, Sydney 2007, Australia
基金
中国国家自然科学基金; 中国博士后科学基金;
关键词
label associated loss; weighting noisy samples; fine-grained visual recognition; noise-tolerant learning; meta-learning;
D O I
10.1007/s11432-023-3922-2
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Recently, intensive attempts have been made to design robust models for fine-grained visual recognition, most notably are the impressive gains for training with noisy labels by incorporating a reweighting strategy into a meta-learning framework. However, it is limited to up or downweighting the contribution of an instance for label reweighting approaches in the learning process. To solve this issue, a novel noise-tolerant method with auxiliary web data is proposed. Specifically, first, the associations made from embeddings of well-labeled data with those of web data and back at the same class are measured. Next, its association probability is employed as a weighting fusion strategy into angular margin-based loss, which makes the trained model robust to noisy datasets. To reduce the influence of the gap between the well-labeled and noisy web data, a bridge schema is proposed via the corresponding loss that encourages the learned embeddings to be coherent. Lastly, the formulation is encapsulated into the meta-learning framework, which can reduce the overfitting of models and learn the network parameters to be noise-tolerant. Extensive experiments are performed on benchmark datasets, and the results clearly show the superiority of the proposed method over existing state-of-the-art approaches.
引用
收藏
页数:18
相关论文
共 50 条
  • [1] Meta label associated loss for fine-grained visual recognition
    Yanchao LI
    Fu XIAO
    Hao LI
    Qun LI
    Shui YU
    [J]. Science China(Information Sciences), 2024, 67 (06) : 230 - 247
  • [2] Robust Fine-Grained Visual Recognition With Neighbor-Attention Label Correction
    Mao, Shunan
    Zhang, Shiliang
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2024, 33 : 2614 - 2626
  • [3] Fine-Grained Crowdsourcing for Fine-Grained Recognition
    Jia Deng
    Krause, Jonathan
    Li Fei-Fei
    [J]. 2013 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2013, : 580 - 587
  • [4] Annotation modification for fine-grained visual recognition
    Luo, Changzhi
    Meng, Zhijun
    Feng, Jiashi
    Ni, Bingbing
    Wang, Meng
    [J]. NEUROCOMPUTING, 2018, 274 : 58 - 65
  • [5] ProtoSimi: label correction for fine-grained visual categorization
    Shen, Jialiang
    Yao, Yu
    Huang, Shaoli
    Wang, Zhiyong
    Zhang, Jing
    Wang, Ruxing
    Yu, Jun
    Liu, Tongliang
    [J]. MACHINE LEARNING, 2024, 113 (04) : 1903 - 1920
  • [6] ProtoSimi: label correction for fine-grained visual categorization
    Jialiang Shen
    Yu Yao
    Shaoli Huang
    Zhiyong Wang
    Jing Zhang
    Ruxing Wang
    Jun Yu
    Tongliang Liu
    [J]. Machine Learning, 2024, 113 : 1903 - 1920
  • [7] Bilinear CNN Models for Fine-grained Visual Recognition
    Lin, Tsung-Yu
    RoyChowdhury, Aruni
    Maji, Subhransu
    [J]. 2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, : 1449 - 1457
  • [8] Hierarchical Bilinear Pooling for Fine-Grained Visual Recognition
    Yu, Chaojian
    Zhao, Xinyi
    Zheng, Qi
    Zhang, Peng
    You, Xinge
    [J]. COMPUTER VISION - ECCV 2018, PT XVI, 2018, 11220 : 595 - 610
  • [9] Fine-grained object recognition in underwater visual data
    Spampinato, C.
    Palazzo, S.
    Joalland, P. H.
    Paris, S.
    Glotin, H.
    Blanc, K.
    Lingrand, D.
    Precioso, F.
    [J]. MULTIMEDIA TOOLS AND APPLICATIONS, 2016, 75 (03) : 1701 - 1720
  • [10] Hierarchical gate network for fine-grained visual recognition
    Chen, Ying
    Song, Jie
    Song, Mingli
    [J]. NEUROCOMPUTING, 2022, 470 : 170 - 181