Second-order Attention Guided Convolutional Activations for Visual Recognition

被引:0
|
作者
Chen, Shannan [1 ]
Wang, Qian [1 ]
Sun, Qiule [2 ]
Liu, Bin [3 ]
Zhang, Jianxin [1 ,4 ]
Zhang, Qiang [1 ,5 ]
机构
[1] Dalian Univ, Key Lab Adv Design & Intelligent Comp, Minist Educ, Dalian, Peoples R China
[2] Dalian Univ Technol, Sch Informat & Commun Engn, Dalian, Peoples R China
[3] Dalian Univ Technol, Int Sch Informat Sci & Engn DUT RUISE, Dalian, Peoples R China
[4] Dalian Univ Technol, Sch Comp Sci & Engn, Dalian, Peoples R China
[5] Dalian Univ Technol, Sch Comp Sci & Technol, Dalian, Peoples R China
基金
国家重点研发计划; 中国国家自然科学基金;
关键词
second-order statistics; channel attention; deep convolutional networks; visual recognition;
D O I
10.1109/ICPR48806.2021.9412350
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recently, modeling deep convolutional activations by the global second-order pooling has shown great advance on visual recognition tasks. However, most of the existing deep second-order statistical models mainly compute second-order statistics of activations of the last convolutional layer as image representations, and they seldom introduce second-order statistics into earlier layers to better fit network topology, thus limiting the representational ability to a certain extent. Motivated by the flexibility of attention blocks that are commonly plugged into intermediate layers of deep convolutional networks (ConvNets), this work makes an attempt to combine deep second-order statistics with attention mechanisms in ConvNets, and further proposes a novel Second-order Attention Guided Network (SoAG-Net) for visual recognition. More specifically, SoAG-Net involves several SoAG modules seemingly inserted into intermediate layers of the network, in which SoAG collects second-order statistics of convolutional activations by polynomial kernel approximation to predict channel-wise attention maps utilized for guiding the learning of convolutional activations through tensor scaling along channel dimension. SoAG improves the nonlinearity of ConvNets and enables ConvNets to fit more complicated distribution of convolutional activations. Experiment results on three commonly used datasets illuminate that SoAG-Net outperforms its counterparts and achieves competitive performance with state-of-the-art models under the same backbone.
引用
下载
收藏
页码:3071 / 3076
页数:6
相关论文
共 50 条
  • [31] Second-order transformer network for video recognition
    Zhang, Bingbing
    Dong, Wei
    Wang, Zhenwei
    Zhang, Jianxin
    Sun, Qiule
    Alexandria Engineering Journal, 2025, 114 : 82 - 94
  • [32] Second-order Temporal Pooling for Action Recognition
    Cherian, Anoop
    Gould, Stephen
    INTERNATIONAL JOURNAL OF COMPUTER VISION, 2019, 127 (04) : 340 - 362
  • [33] Second-order Temporal Pooling for Action Recognition
    Anoop Cherian
    Stephen Gould
    International Journal of Computer Vision, 2019, 127 : 340 - 362
  • [34] RECOGNITION IN MULTICULTURAL SOCIETIES Intergroup relations as second-order recognition
    Herzog, Benno
    REVISTA INTERNACIONAL DE SOCIOLOGIA, 2015, 73 (02):
  • [35] Contextualizer: Connecting the Dots of Context with Second-Order Attention
    Maupome, Diego
    Meurs, Marie-Jean
    INFORMATION, 2022, 13 (06)
  • [36] Attention in first- and second-order motion perception
    Liu, CQ
    Lu, ZL
    Dosher, BA
    INVESTIGATIVE OPHTHALMOLOGY & VISUAL SCIENCE, 1999, 40 (04) : S973 - S973
  • [37] A model of attention-guided visual perception and recognition
    Rybak, IA
    Gusakova, VI
    Golovan, AV
    Podladchikova, LN
    Shevtsova, NA
    VISION RESEARCH, 1998, 38 (15-16) : 2387 - 2400
  • [38] 3Global Second-order Pooling Convolutional Networks
    Gao, Zilin
    Xie, Jiangtao
    Wang, Qilong
    Li, Peihua
    2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 3019 - 3028
  • [39] Visual gait recognition based on convolutional block attention network
    Xiuhui Wang
    Shaohui Hu
    Multimedia Tools and Applications, 2022, 81 : 29459 - 29476
  • [40] Visual gait recognition based on convolutional block attention network
    Wang, Xiuhui
    Hu, Shaohui
    MULTIMEDIA TOOLS AND APPLICATIONS, 2022, 81 (20) : 29459 - 29476