Towards efficient video-based action recognition: context-aware memory attention network

被引:2
|
作者
Koh, Thean Chun [1 ]
Yeo, Chai Kiat [1 ]
Jing, Xuan [1 ,2 ]
Sivadas, Sunil [2 ]
机构
[1] Nanyang Technol Univ, Sch Comp Sci & Engn, 50 Nanyang Ave, Singapore 639798, Singapore
[2] NCS Pte Ltd, Ang Mo Kio St 62, Singapore 569141, Singapore
来源
SN APPLIED SCIENCES | 2023年 / 5卷 / 12期
关键词
Action recognition; Deep learning; Convolutional neural network; Attention; BIDIRECTIONAL LSTM; CLASSIFICATION;
D O I
10.1007/s42452-023-05568-5
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
Given the prevalence of surveillance cameras in our daily lives, human action recognition from videos holds significant practical applications. A persistent challenge in this field is to develop more efficient models capable of real-time recognition with high accuracy for widespread implementation. In this research paper, we introduce a novel human action recognition model named Context-Aware Memory Attention Network (CAMA-Net), which eliminates the need for optical flow extraction and 3D convolution which are computationally intensive. By removing these components, CAMA-Net achieves superior efficiency compared to many existing approaches in terms of computation efficiency. A pivotal component of CAMA-Net is the Context-Aware Memory Attention Module, an attention module that computes the relevance score between key-value pairs obtained from the 2D ResNet backbone. This process establishes correspondences between video frames. To validate our method, we conduct experiments on four well-known action recognition datasets: ActivityNet, Diving48, HMDB51 and UCF101. The experimental results convincingly demonstrate the effectiveness of our proposed model, surpassing the performance of existing 2D-CNN based baseline models.Article HighlightsRecent human action recognition models are not yet ready for practical applications due to high computation needs.We propose a 2D CNN-based human action recognition method to reduce the computation load.The proposed method achieves competitive performance compared to most SOTA 2D CNN-based methods on public datasets.
引用
收藏
页数:12
相关论文
共 50 条
  • [31] A Discriminative Convolutional Neural Network with Context-aware Attention
    Zhou, Yuxiang
    Liao, Lejian
    Gao, Yang
    Huang, Heyan
    Wei, Xiaochi
    ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY, 2020, 11 (05)
  • [32] Graph Attention Network for Context-Aware Visual Tracking
    Shao, Yanyan
    Guo, Dongyan
    Cui, Ying
    Wang, Zhenhua
    Zhang, Liyan
    Zhang, Jianhua
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024,
  • [33] Efficient context-aware computing: a systematic model for dynamic working memory updates in context-aware computing
    Ali M.
    Arshad M.
    Uddin I.
    Binsawad M.
    Sawad A.B.
    Sohaib O.
    PeerJ Computer Science, 2024, 10 : 1 - 19
  • [34] Efficient context-aware computing: a systematic model for dynamic working memory updates in context-aware computing
    Ali, Mumtaz
    Arshad, Muhammad
    Uddin, Ijaz
    Binsawad, Muhammad
    Bin Sawad, Abdullah
    Sohaib, Osama
    PEERJ COMPUTER SCIENCE, 2024, 10
  • [35] Attention based hierarchical LSTM network for context-aware microblog sentiment classification
    Shi Feng
    Yang Wang
    Liran Liu
    Daling Wang
    Ge Yu
    World Wide Web, 2019, 22 : 59 - 81
  • [36] Hierarchical attention-based context-aware network for red tide forecasting
    He, Xiaoyu
    Shi, Suixiang
    Geng, Xiulin
    Xu, Lingyu
    APPLIED SOFT COMPUTING, 2022, 127
  • [37] Attention based hierarchical LSTM network for context-aware microblog sentiment classification
    Feng, Shi
    Wang, Yang
    Liu, Liran
    Wang, Daling
    Yu, Ge
    WORLD WIDE WEB-INTERNET AND WEB INFORMATION SYSTEMS, 2019, 22 (01): : 59 - 81
  • [38] Context-aware and co-attention network based image captioning model
    Sharma, Himanshu
    Srivastava, Swati
    IMAGING SCIENCE JOURNAL, 2023, 71 (03): : 244 - 256
  • [39] CAtNIPP: Context-Aware Attention-based Network for Informative Path Planning
    Cao, Yuhong
    Wang, Yizhuo
    Vashisth, Apoorva
    Fan, Haolin
    Sartoretti, Guillaume
    CONFERENCE ON ROBOT LEARNING, VOL 205, 2022, 205 : 1928 - 1937
  • [40] Towards Global Video Scene Segmentation with Context-Aware Transformer
    Yang, Yang
    Huang, Yurui
    Guo, Weili
    Xu, Baohua
    Xia, Dingyin
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 3, 2023, : 3206 - 3213