A Neural Autoregressive Approach to Attention-based Recognition

被引:16
|
作者
Zheng, Yin [1 ]
Zemel, Richard S. [2 ]
Zhang, Yu-Jin [1 ]
Larochelle, Hugo [3 ]
机构
[1] Tsinghua Univ, Dept Elect Engn, Beijing 10084, Peoples R China
[2] Univ Toronto, Dept Comp Sci, Toronto, ON M5S 3G4, Canada
[3] Univ Sherbrooke, Dept Informat, Sherbrooke, PQ J1K 2R1, Canada
基金
加拿大自然科学与工程研究理事会;
关键词
Deep learning; Attention-based recognition; Neural networks; Neural autoregressive distribution estimator;
D O I
10.1007/s11263-014-0765-x
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Tasks that require the synchronization of perception and action are incredibly hard and pose a fundamental challenge to the fields of machine learning and computer vision. One important example of such a task is the problem of performing visual recognition through a sequence of controllable fixations; this requires jointly deciding what inference to perform from fixations and where to perform these fixations. While these two problems are challenging when addressed separately, they become even more formidable if solved jointly. Recently, a restricted Boltzmann machine (RBM) model was proposed that could learn meaningful fixation policies and achieve good recognition performance. In this paper, we propose an alternative approach based on a feed-forward, auto-regressive architecture, which permits exact calculation of training gradients (given the fixation sequence), unlike for the RBM model. On a problem of facial expression recognition, we demonstrate the improvement gained by this alternative approach. Additionally, we investigate several variations of the model in order to shed some light on successful strategies for fixation-based recognition.
引用
收藏
页码:67 / 79
页数:13
相关论文
共 50 条
  • [21] Attention-based recurrent neural network for automatic behavior laying hen recognition
    Laleye, Frejus A. A.
    Mousse, Mikael A.
    [J]. MULTIMEDIA TOOLS AND APPLICATIONS, 2024, 83 (22) : 62443 - 62458
  • [22] EEG emotion recognition using attention-based convolutional transformer neural network
    Gong, Linlin
    Li, Mingyang
    Zhang, Tao
    Chen, Wanzhong
    [J]. BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2023, 84
  • [23] Attention-Based Deep Neural Network and Its Application to Scene Text Recognition
    He, Haizhen
    Li, Jiehan
    [J]. 2019 IEEE 11TH INTERNATIONAL CONFERENCE ON COMMUNICATION SOFTWARE AND NETWORKS (ICCSN 2019), 2019, : 672 - 677
  • [24] 4D attention-based neural network for EEG emotion recognition
    Guowen Xiao
    Meng Shi
    Mengwen Ye
    Bowen Xu
    Zhendi Chen
    Quansheng Ren
    [J]. Cognitive Neurodynamics, 2022, 16 : 805 - 818
  • [25] Attention-Based Neural Text Segmentation
    Badjatiya, Pinkesh
    Kurisinkel, Litton J.
    Gupta, Manish
    Varma, Vasudeva
    [J]. ADVANCES IN INFORMATION RETRIEVAL (ECIR 2018), 2018, 10772 : 180 - 193
  • [26] Attention-based Neural Load Forecasting: A Dynamic Feature Selection Approach
    Xiong, Jing
    Zhou, Pengyang
    Chen, Alan
    Zhang, Yu
    [J]. 2021 IEEE POWER & ENERGY SOCIETY GENERAL MEETING (PESGM), 2021,
  • [27] Underwater acoustic target recognition using attention-based deep neural network
    Xiao, Xu
    Wang, Wenbo
    Ren, Qunyan
    Gerstoft, Peter
    Ma, Li
    [J]. JASA EXPRESS LETTERS, 2021, 1 (10):
  • [28] 4D attention-based neural network for EEG emotion recognition
    Xiao, Guowen
    Shi, Meng
    Ye, Mengwen
    Xu, Bowen
    Chen, Zhendi
    Ren, Quansheng
    [J]. COGNITIVE NEURODYNAMICS, 2022, 16 (04) : 805 - 818
  • [29] Attention-Based Neural Tag Recommendation
    Yuan, Jiahao
    Jin, Yuanyuan
    Liu, Wenyan
    Wang, Xiaoling
    [J]. DATABASE SYSTEMS FOR ADVANCED APPLICATIONS (DASFAA 2019), PT II, 2019, 11447 : 350 - 365
  • [30] Neutron: an attention-based neural decompiler
    Liang, Ruigang
    Cao, Ying
    Hu, Peiwei
    Chen, Kai
    [J]. CYBERSECURITY, 2021, 4 (01)