Exploring Category-Shared and Category-Specific Features for Fine-Grained Image Classification

被引:0
|
作者
Wang, Haoyu [1 ]
Chang, DongLiang [1 ]
Liu, Weidong [3 ]
Xiao, Bo [1 ]
Ma, Zhanyu [1 ,2 ]
Guo, Jun [1 ]
Chang, Yaning [1 ]
机构
[1] Beijing Univ Posts & Telecommun, Beijing 100876, Peoples R China
[2] Beijing Acad Artificial Intelligence, Beijing 100876, Peoples R China
[3] China Mobile Res Inst, Beijing 100876, Peoples R China
基金
国家重点研发计划; 中国国家自然科学基金; 北京市自然科学基金;
关键词
Fine-grained image classification; Semantic intra-class similarity; Channel-wise attention; Spatial-wise attention;
D O I
10.1007/978-3-030-88004-0_15
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The attention mechanism is one of the most vital branches to solve fine-grained image classification (FGIC) tasks, while most existing attention-based methods only focus on inter-class variance and barely model the intra-class similarity. They perform the classification tasks by enhancing inter-class variance, which narrows down the intra-class similarity indirectly. In this paper, we intend to utilize the intra-class similarity as assistance to improve the classification performance of the obtained attention feature maps. To obtain and utilize the intra-class information, a novel attention mechanism, named category-shared and category-specific feature extraction module (CSS-FEM) is proposed in this paper. CSS-FEM firstly extracts the category-shared features based on the intra-class semantic relationship, then focuses on the discriminative parts. CSS-FEM is assembled by two parts: 1) The category-shared feature extraction module extracts category-shared features that contain high intra-class semantic similarity, to reduce the large intra-class variances. 2) The category-specific feature extraction module performs spatial-attention mechanism in category-shared features to find the discriminative information as category-specific features to decrease the high inter-class similarity. Compared with the state-of-the-art methods, the experimental results on three commonly used FGIC datasets show that the effectiveness and competitiveness of the proposed CSS-FEM. Ablation experiments and visualizations are also provided for further demonstrations.
引用
收藏
页码:179 / 190
页数:12
相关论文
共 50 条
  • [11] Progressive Learning of Category-Consistent Multi-Granularity Features for Fine-Grained Visual Classification
    Du, Ruoyi
    Xie, Jiyang
    Ma, Zhanyu
    Chang, Dongliang
    Song, Yi-Zhe
    Guo, Jun
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2022, 44 (12) : 9521 - 9535
  • [12] Exploring Misclassification Information for Fine-Grained Image Classification
    Wang, Da-Han
    Zhou, Wei
    Li, Jianmin
    Wu, Yun
    Zhu, Shunzhi
    SENSORS, 2021, 21 (12)
  • [13] The Scope and Limits of Fine-Grained Image and Category Information in the Ventral Visual Pathway
    Badwal, Markus W.
    Bergmann, Johanna
    Roth, Johannes H. R.
    Doeller, Christian F.
    Hebart, Martin N.
    JOURNAL OF NEUROSCIENCE, 2025, 45 (03):
  • [14] Category-Specific Object Reconstruction from a Single Image
    Kar, Abhishek
    Tulsiani, Shubham
    Carreira, Joao
    Malik, Jitendra
    2015 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2015, : 1966 - 1974
  • [15] Exploiting Category Similarity-Based Distributed Labeling for Fine-Grained Visual Classification
    Du, Pengzhen
    Sun, Zeren
    Yao, Yazhou
    Tang, Zhenmin
    IEEE ACCESS, 2020, 8 : 186679 - 186690
  • [16] DNA: Denoised Neighborhood Aggregation for Fine-grained Category Discovery
    An, Wenbin
    Tian, Feng
    Shi, Wenkai
    Chen, Yan
    Zheng, Qinghua
    Wang, QianYing
    Chen, Ping
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2023), 2023, : 12292 - 12302
  • [17] Category attention transfer for efficient fine-grained visual categorization
    Liao, Qiyu
    Wang, Dadong
    Xu, Min
    PATTERN RECOGNITION LETTERS, 2022, 153 : 10 - 15
  • [18] Category-Contrastive Fine-Grained Crowd Counting and Beyond
    Zhang, Meijing
    Chen, Mengxue
    Li, Qi
    Chen, Yanchen
    Lin, Rui
    Li, Xiaolian
    He, Shengfeng
    Liu, Wenxi
    IEEE TRANSACTIONS ON MULTIMEDIA, 2025, 27 : 477 - 488
  • [19] Fine-Grained Features for Image Captioning
    Shao, Mengyue
    Feng, Jie
    Wu, Jie
    Zhang, Haixiang
    Zheng, Yayu
    CMC-COMPUTERS MATERIALS & CONTINUA, 2023, 75 (03): : 4697 - 4712
  • [20] Learning Category-Specific Mesh Reconstruction from Image Collections
    Kanazawa, Angjoo
    Tulsiani, Shubham
    Efros, Alexei A.
    Malik, Jitendra
    COMPUTER VISION - ECCV 2018, PT 15, 2018, 11219 : 386 - 402