FDAM: full-dimension attention module for deep convolutional neural networks

被引:3
|
作者
Cai, Silin [1 ]
Wang, Changping [1 ]
Ding, Jiajun [2 ,3 ]
Yu, Jun [2 ]
Fan, Jianping [2 ]
机构
[1] Hangzhou Dianzi Univ, Zhuoyue Honor Coll, 2 Main St, Hangzhou 310018, Zhejiang, Peoples R China
[2] Hangzhou Dianzi Univ, Sch Comp Sci & Technol, 2 Main St, Hangzhou 310018, Zhejiang, Peoples R China
[3] Hangzhou Dianzi Univ, Shangyu Inst Sci & Engn, Shangyu 312300, Zhejiang, Peoples R China
关键词
Attention mechanism; Convolutional neural network; Image classification; Object recognition; Elo rating mechanism;
D O I
10.1007/s13735-022-00248-3
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The attention mechanism is an important component of cross-modal research. It can improve the performance of convolutional neural networks by distinguishing the informative parts of the feature map from the useless ones. Various kinds of attention are proposed by recent studies. Different attentions use distinct division method to weight each part of the feature map. In this paper, we propose a full-dimension attention module, which is a lightweight, fully interactive 3-D attention mechanism. FDAM generates 3-D attention maps for both spatial and channel dimensions in parallel and then multiplies them to the feature map. It is difficult to obtain discriminative attention map cell under channel interaction at a low computational cost. Therefore, we adapt a generalized Elo rating mechanism to generate cell-level attention maps. We store historical information with a slight amount of non-training parameters to spread the computation over each training iteration. The proposed module can be seamlessly integrated into the end-to-end training of the CNN framework. Experiments demonstrate that it outperforms many existing attention mechanisms on different network structures and datasets for computer vision tasks, such as image classification and object detection.
引用
收藏
页码:599 / 610
页数:12
相关论文
共 50 条
  • [21] ULSAM: Ultra-Lightweight Subspace Attention Module for Compact Convolutional Neural Networks
    Saini, Rajat
    Jha, Nandan Kumar
    Das, Bedanta
    Mittal, Sparsh
    Mohan, C. Krishna
    2020 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV), 2020, : 1616 - 1625
  • [22] A Symmetric Efficient Spatial and Channel Attention (ESCA) Module Based on Convolutional Neural Networks
    Liu, Huaiyu
    Zhang, Yueyuan
    Chen, Yiyang
    SYMMETRY-BASEL, 2024, 16 (08):
  • [23] An efficient attention module for 3d convolutional neural networks in action recognition
    Guanghao Jiang
    Xiaoyan Jiang
    Zhijun Fang
    Shanshan Chen
    Applied Intelligence, 2021, 51 : 7043 - 7057
  • [24] Underground Target Classification From Full-Polarimetric GPR Data Using Deep Convolutional Neural Network With Channel Attention Module
    Li, Jingxia
    Li, Jiasu
    Huang, Zheng
    Qu, Yanlin
    Liu, Li
    Xu, Hang
    Wang, Bingjie
    IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2024, 21 : 1 - 5
  • [25] Deep Convolutional Neural Networks
    Gonzalez, Rafael C.
    IEEE SIGNAL PROCESSING MAGAZINE, 2018, 35 (06) : 79 - 87
  • [26] Reparameterized attention for convolutional neural networks
    Wu, Yiming
    Li, Ruixiang
    Yu, Yunlong
    Li, Xi
    PATTERN RECOGNITION LETTERS, 2022, 164 : 89 - 95
  • [27] Deep Convolutional Neural Networks with Layer-wise Context Expansion and Attention
    Yu, Dong
    Xiong, Wayne
    Droppo, Jasha
    Stolcke, Andreas
    Ye, Guoli
    Li, Jinyu
    Zweig, Geoffrey
    17TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2016), VOLS 1-5: UNDERSTANDING SPEECH PROCESSING IN HUMANS AND MACHINES, 2016, : 17 - 21
  • [28] MCA: Multidimensional collaborative attention in deep convolutional neural networks for image recognition
    Yu, Yang
    Zhang, Yi
    Cheng, Zeyu
    Song, Zhe
    Tang, Chengkai
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2023, 126
  • [29] Attention-guided deep convolutional neural networks for skin cancer classification
    Aggarwal, Arshiya
    Das, Nisheet
    Sreedevi, Indu
    2019 NINTH INTERNATIONAL CONFERENCE ON IMAGE PROCESSING THEORY, TOOLS AND APPLICATIONS (IPTA), 2019,
  • [30] The Costs and Benefits of Goal-Directed Attention in Deep Convolutional Neural Networks
    Luo X.
    Roads B.D.
    Love B.C.
    Computational Brain & Behavior, 2021, 4 (2) : 213 - 230