Adopting Attention and Cross-Layer Features for Fine-Grained Representation

被引:3
|
作者
Sun Fayou [1 ]
Ngo, Hea Choon [1 ]
Sek, Yong Wee [1 ]
机构
[1] Univ Teknikal Malaysia Melaka, Fac Informat & Commun Technol, Ctr Adv Comp Technol, Durian Tunggal 76100, Malacca, Malaysia
关键词
Feature extraction; Representation learning; Semantics; Transformers; Sun; Convolution; Task analysis; Associating cross-layer features; attention-based operations; self-attention; CLNET;
D O I
10.1109/ACCESS.2022.3195907
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Fine-grained visual classification (FGVC) is challenging task due to discriminative feature representations. The attention-based methods show great potential for FGVC, which neglect that the deeply digging inter-layer feature relations have an impact on refining feature learning. Similarly, the associating cross-layer features methods achieve significant feature enhancement, which lost the long-distance dependencies between elements. However, most of the previous researches neglect that these two methods are mutually correlated to reinforce feature learning, which are independent of each other in related models. Thus, we adopt the respective advantages of the two methods to promote fine-gained feature representations. In this paper, we propose a novel CLNET network, which effectively applies attention mechanism and cross-layer features to obtain feature representations. Specifically, CL-NET consists of 1) adopting self-attention to capture long-rang dependencies for each element, 2) associating cross-layer features to reinforce feature learning,and 3) to cover more feature regions,we integrate attention-based operations between output and input. Experiments verify that CLNET yields new state-of-the-art performance on three widely used fine-grained benchmarks, including CUB-200-2011, Stanford Cars and FGVC-Aircraft. The url of our code is https://github.com/dlearing/CLNET.git.
引用
收藏
页码:82376 / 82383
页数:8
相关论文
共 50 条
  • [41] Self-Attention based fine-grained cross-media hybrid network
    Shan, Wei
    Huang, Dan
    Wang, Jiangtao
    Zou, Feng
    Li, Suwen
    PATTERN RECOGNITION, 2022, 130
  • [42] Learning Cascade Attention for fine-grained image classification
    Zhu, Youxiang
    Li, Ruochen
    Yang, Yin
    Ye, Ning
    NEURAL NETWORKS, 2020, 122 : 174 - 182
  • [43] Adversarial erasing attention for fine-grained image classification
    Jinsheng Ji
    Linfeng Jiang
    Tao Zhang
    Weilin Zhong
    Huilin Xiong
    Multimedia Tools and Applications, 2021, 80 : 22867 - 22889
  • [44] The Pairs Network of Attention model for Fine-grained Classification
    Wang, Gaihua
    Han, Jingwei
    Zhang, Chuanlei
    Yao, Jingxuan
    Zhu, Bolun
    PROCEEDINGS OF THE 2024 6TH INTERNATIONAL CONFERENCE ON BIG DATA ENGINEERING, BDE 2024, 2024, : 39 - 47
  • [45] Aggregate attention module for fine-grained image classification
    Xingmei Wang
    Jiahao Shi
    Hamido Fujita
    Yilin Zhao
    Journal of Ambient Intelligence and Humanized Computing, 2023, 14 : 8335 - 8345
  • [46] Adversarial erasing attention for fine-grained image classification
    Ji, Jinsheng
    Jiang, Linfeng
    Zhang, Tao
    Zhong, Weilin
    Xiong, Huilin
    MULTIMEDIA TOOLS AND APPLICATIONS, 2021, 80 (15) : 22867 - 22889
  • [47] Decoding the neural representation of fine-grained conceptual categories
    Ghio, Marta
    Vaghi, Matilde Maria Serena
    Perani, Daniela
    Tettamanti, Marco
    NEUROIMAGE, 2016, 132 : 93 - 103
  • [48] Fine-Grained Visual-Textual Representation Learning
    He, Xiangteng
    Peng, Yuxin
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2020, 30 (02) : 520 - 531
  • [49] Fast Attention CNN for Fine-Grained Crack Segmentation
    Lee, Hyunnam
    Yoo, Juhan
    SENSORS, 2023, 23 (04)
  • [50] Fine-Grained Video Categorization with Redundancy Reduction Attention
    Zhu, Chen
    Tan, Xiao
    Zhou, Feng
    Liu, Xiao
    Yue, Kaiyu
    Ding, Errui
    Ma, Yi
    COMPUTER VISION - ECCV 2018, PT V, 2018, 11209 : 139 - 155