Adopting Attention and Cross-Layer Features for Fine-Grained Representation

被引:3
|
作者
Sun Fayou [1 ]
Ngo, Hea Choon [1 ]
Sek, Yong Wee [1 ]
机构
[1] Univ Teknikal Malaysia Melaka, Fac Informat & Commun Technol, Ctr Adv Comp Technol, Durian Tunggal 76100, Malacca, Malaysia
关键词
Feature extraction; Representation learning; Semantics; Transformers; Sun; Convolution; Task analysis; Associating cross-layer features; attention-based operations; self-attention; CLNET;
D O I
10.1109/ACCESS.2022.3195907
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Fine-grained visual classification (FGVC) is challenging task due to discriminative feature representations. The attention-based methods show great potential for FGVC, which neglect that the deeply digging inter-layer feature relations have an impact on refining feature learning. Similarly, the associating cross-layer features methods achieve significant feature enhancement, which lost the long-distance dependencies between elements. However, most of the previous researches neglect that these two methods are mutually correlated to reinforce feature learning, which are independent of each other in related models. Thus, we adopt the respective advantages of the two methods to promote fine-gained feature representations. In this paper, we propose a novel CLNET network, which effectively applies attention mechanism and cross-layer features to obtain feature representations. Specifically, CL-NET consists of 1) adopting self-attention to capture long-rang dependencies for each element, 2) associating cross-layer features to reinforce feature learning,and 3) to cover more feature regions,we integrate attention-based operations between output and input. Experiments verify that CLNET yields new state-of-the-art performance on three widely used fine-grained benchmarks, including CUB-200-2011, Stanford Cars and FGVC-Aircraft. The url of our code is https://github.com/dlearing/CLNET.git.
引用
收藏
页码:82376 / 82383
页数:8
相关论文
共 50 条
  • [21] Cross-Layer Framework for Fine-Grained Channel Access in Next Generation High-Density Wi Fi Networks
    ZHAO Haitao
    ZHANG Shaojie
    Emiliano GarciaPalacios
    中国通信, 2016, 13 (02) : 55 - 67
  • [22] Cross-Layer Framework for Fine-Grained Channel Access in Next Generation High-Density Wi Fi Networks
    ZHAO Haitao
    ZHANG Shaojie
    Emiliano Garcia-Palacios
    China Communications, 2016, (02) : 55 - 67
  • [23] Fine-Grained Features for Image Captioning
    Shao, Mengyue
    Feng, Jie
    Wu, Jie
    Zhang, Haixiang
    Zheng, Yayu
    CMC-COMPUTERS MATERIALS & CONTINUA, 2023, 75 (03): : 4697 - 4712
  • [24] Beyond the Attention: Distinguish the Discriminative and Confusable Features For Fine-grained Image Classification
    Shi, Xiruo
    Xu, Liutong
    Wang, Pengfei
    Gao, Yuanyuan
    Jian, Haifang
    Liu, Wu
    MM '20: PROCEEDINGS OF THE 28TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, 2020, : 601 - 609
  • [25] Aggregating Object Features Based on Attention Weights for Fine-Grained Image Retrieval
    Lin, HongLi
    Song, Yongqi
    Zeng, Zixuan
    Wang, Weisheng
    Wang, Jiayi
    2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, : 2838 - 2844
  • [26] Multimodal Stacked Cross Attention Network for Fine-Grained Fake News Detection
    Huang, Zhongqiang
    Hu, Yuxue
    Zeng, Zhi
    Li, Xiang
    Sha, Ying
    2023 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO, ICME, 2023, : 2837 - 2842
  • [27] GRAPH FINE-GRAINED CONTRASTIVE REPRESENTATION LEARNING
    Tang, Hui
    Liang, Xun
    Guo, Yuhui
    Zheng, Xiangping
    Wu, Bo
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 3478 - 3482
  • [28] Fine-grained representation learning in convolutional autoencoders
    Luo, Chang
    Wang, Jie
    JOURNAL OF ELECTRONIC IMAGING, 2016, 25 (02)
  • [29] Representation Learning for Fine-Grained Change Detection
    O'Mahony, Niall
    Campbell, Sean
    Krpalkova, Lenka
    Carvalho, Anderson
    Walsh, Joseph
    Riordan, Daniel
    SENSORS, 2021, 21 (13)
  • [30] Fine-grained attention for image caption generation
    Chang, Yan-Shuo
    MULTIMEDIA TOOLS AND APPLICATIONS, 2018, 77 (03) : 2959 - 2971