A fine-grained image classification method based on information interaction

被引:0
|
作者
Zhu, Shuo [1 ]
Zhang, Xukang [2 ]
Wang, Yu [2 ]
Wang, Zongyang [3 ]
Sun, Jiahao [1 ]
机构
[1] Wuxi Univ, Jiangsu Prov Engn Res Ctr Photon Devices & Syst In, Wuxi 214105, Peoples R China
[2] Nanjing Univ Informat Sci & Technol, Sch Elect & Informat Engn, Nanjing, Peoples R China
[3] Wuxi Xiyuan Technol Co Ltd, Wuxi, Jiangsu, Peoples R China
关键词
image classification; image processing; learning (artificial intelligence);
D O I
10.1049/ipr2.13295
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
To enhance the accuracy of fine-grained image classification and address challenges such as excessive interference factors within the dataset, inadequate extraction of local key features, and insufficient channel semantic association, a dual-branch information interaction model that integrates convolutional neural networks (CNN) with Vision Transformers is proposed. This model leverages the Vision Transformer branch to extract global features, which are subsequently combined with the CNN branch to further augment the model's capability for local information extraction. In order to enhance the ability of the CNN branch to extract global information and reduce the loss of feature information, a feature enhancement module is added to the CNN branch. Since the Vision Transformer branch directly convolves with the convolution kernel will result in the inability to learn the underlying features of the image, a shallow feature extraction module is proposed, and the CNN and Vision Transformer branches interact with the information of the dual branches through the down-sampling Down module and the up-sampling UP module. The accuracy of the improved method on CUB-200-2011, Stanford Cars and FGVC-Aircraft fine-grained image classification datasets are 95.2%, 97.1% and 96.9%, respectively. The experimental results show that the method has good generalization on different datasets.
引用
收藏
页码:4852 / 4861
页数:10
相关论文
共 50 条
  • [1] Exploring Misclassification Information for Fine-Grained Image Classification
    Wang, Da-Han
    Zhou, Wei
    Li, Jianmin
    Wu, Yun
    Zhu, Shunzhi
    SENSORS, 2021, 21 (12)
  • [2] Improving Fine-Grained Image Classification With Multimodal Information
    Xu, Jie
    Zhang, Xiaoqian
    Zhao, Changming
    Geng, Zili
    Feng, Yuren
    Miao, Ke
    Li, Yunji
    IEEE TRANSACTIONS ON MULTIMEDIA, 2024, 26 : 2082 - 2095
  • [3] Fine-grained image classification method based on hybrid attention module
    Lu, Weixiang
    Yang, Ying
    Yang, Lei
    FRONTIERS IN NEUROROBOTICS, 2024, 18
  • [4] A Fine-Grained Image Classification Method Built on MobileViT
    Lu, Zhengqiu
    Wang, Haiying
    INTERNATIONAL JOURNAL OF PATTERN RECOGNITION AND ARTIFICIAL INTELLIGENCE, 2024, 38 (06)
  • [5] Fine-grained image classification method with noisy labels based on retrieval augmentation
    Bao, Heng
    Deng, Lirui
    Zhang, Liang
    Chen, Xunxun
    Beijing Hangkong Hangtian Daxue Xuebao/Journal of Beijing University of Aeronautics and Astronautics, 2024, 50 (07): : 2284 - 2292
  • [6] An Interactive Deep Learning Method For Fine-grained Image Classification
    Luo, Liumin
    Wang, Mingxia
    Liu, Xiaoqing
    JOURNAL OF APPLIED SCIENCE AND ENGINEERING, 2025, 28 (04): : 701 - 708
  • [7] Image local structure information learning for fine-grained visual classification
    Lu, Jin
    Zhang, Weichuan
    Zhao, Yali
    Sun, Changming
    SCIENTIFIC REPORTS, 2022, 12 (01)
  • [8] Image local structure information learning for fine-grained visual classification
    Jin Lu
    Weichuan Zhang
    Yali Zhao
    Changming Sun
    Scientific Reports, 12
  • [9] Fine-Grained Image Classification With Global Information and Adaptive Compensation Loss
    Wu, Qin
    Miao, Shuting
    Chai, Zhilei
    Guo, Guodong
    IEEE SIGNAL PROCESSING LETTERS, 2022, 29 : 36 - 40
  • [10] Pixel Saliency Based Encoding for Fine-Grained Image Classification
    Yin, Chao
    Zhang, Lei
    Liu, Ji
    PATTERN RECOGNITION AND COMPUTER VISION (PRCV 2018), PT I, 2018, 11256 : 274 - 285