A Neural Autoregressive Approach to Attention-based Recognition

被引:16
|
作者
Zheng, Yin [1 ]
Zemel, Richard S. [2 ]
Zhang, Yu-Jin [1 ]
Larochelle, Hugo [3 ]
机构
[1] Tsinghua Univ, Dept Elect Engn, Beijing 10084, Peoples R China
[2] Univ Toronto, Dept Comp Sci, Toronto, ON M5S 3G4, Canada
[3] Univ Sherbrooke, Dept Informat, Sherbrooke, PQ J1K 2R1, Canada
基金
加拿大自然科学与工程研究理事会;
关键词
Deep learning; Attention-based recognition; Neural networks; Neural autoregressive distribution estimator;
D O I
10.1007/s11263-014-0765-x
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Tasks that require the synchronization of perception and action are incredibly hard and pose a fundamental challenge to the fields of machine learning and computer vision. One important example of such a task is the problem of performing visual recognition through a sequence of controllable fixations; this requires jointly deciding what inference to perform from fixations and where to perform these fixations. While these two problems are challenging when addressed separately, they become even more formidable if solved jointly. Recently, a restricted Boltzmann machine (RBM) model was proposed that could learn meaningful fixation policies and achieve good recognition performance. In this paper, we propose an alternative approach based on a feed-forward, auto-regressive architecture, which permits exact calculation of training gradients (given the fixation sequence), unlike for the RBM model. On a problem of facial expression recognition, we demonstrate the improvement gained by this alternative approach. Additionally, we investigate several variations of the model in order to shed some light on successful strategies for fixation-based recognition.
引用
收藏
页码:67 / 79
页数:13
相关论文
共 50 条
  • [41] A Neural Attention-Based Encoder-Decoder Approach for English to Bangla Translation
    Al Shiam, Abdullah
    Redwan, Sadi Md.
    Kabir, Humaun
    Shin, Jungpil
    [J]. COMPUTER SCIENCE JOURNAL OF MOLDOVA, 2023, 31 (01) : 70 - 85
  • [42] Group Task Recommendation in Mobile Crowdsensing: An Attention-Based Neural Collaborative Approach
    Wei, Kaimin
    Qi, Guozi
    Li, Zhetao
    Guo, Song
    Chen, Jinpeng
    [J]. IEEE TRANSACTIONS ON MOBILE COMPUTING, 2024, 23 (08) : 8066 - 8076
  • [43] A Hierarchical Neural Attention-based Text Classifier
    Sinha, Koustuv
    Dong, Yue
    Cheung, Jackie C. K.
    Ruths, Derek
    [J]. 2018 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2018), 2018, : 817 - 823
  • [44] Attention-based graph neural networks: a survey
    Chengcheng Sun
    Chenhao Li
    Xiang Lin
    Tianji Zheng
    Fanrong Meng
    Xiaobin Rui
    Zhixiao Wang
    [J]. Artificial Intelligence Review, 2023, 56 : 2263 - 2310
  • [45] Attention-based graph neural networks: a survey
    Sun, Chengcheng
    Li, Chenhao
    Lin, Xiang
    Zheng, Tianji
    Meng, Fanrong
    Rui, Xiaobin
    Wang, Zhixiao
    [J]. ARTIFICIAL INTELLIGENCE REVIEW, 2023, 56 (SUPPL 2) : 2263 - 2310
  • [46] Attention-based Hierarchical Neural Query Suggestion
    Chen, Wanyu
    Cai, Fei
    Chen, Honghui
    de Rijke, Maarten
    [J]. ACM/SIGIR PROCEEDINGS 2018, 2018, : 1093 - 1096
  • [47] End-to-End Neural Speaker Diarization with an Iterative Refinement of Non-Autoregressive Attention-based Attractors
    Rybicka, Magdalena
    Villalba, Jesus
    Dehak, Najim
    Kowalczyk, Konrad
    [J]. INTERSPEECH 2022, 2022, : 5090 - 5094
  • [48] Attention-Based SeriesNet: An Attention-Based Hybrid Neural Network Model for Conditional Time Series Forecasting
    Cheng, Yepeng
    Liu, Zuren
    Morimoto, Yasuhiko
    [J]. INFORMATION, 2020, 11 (06)
  • [49] Dynamic Korean Sign Language Recognition Using Pose Estimation Based and Attention-Based Neural Network
    Shin, Jungpil
    Miah, Abu Saleh Musa
    Suzuki, Kota
    Hirooka, Koki
    Hasan, Md. Al Mehedi
    [J]. IEEE ACCESS, 2023, 11 : 143501 - 143513
  • [50] Attention-Based Multi-Learning Approach for Speech Emotion Recognition With Dilated Convolution
    Kakuba, Samuel
    Poulose, Alwin
    Han, Dong Seog
    [J]. IEEE ACCESS, 2022, 10 : 122302 - 122313