Towards efficient video-based action recognition: context-aware memory attention network

被引:2
|
作者
Koh, Thean Chun [1 ]
Yeo, Chai Kiat [1 ]
Jing, Xuan [1 ,2 ]
Sivadas, Sunil [2 ]
机构
[1] Nanyang Technol Univ, Sch Comp Sci & Engn, 50 Nanyang Ave, Singapore 639798, Singapore
[2] NCS Pte Ltd, Ang Mo Kio St 62, Singapore 569141, Singapore
来源
SN APPLIED SCIENCES | 2023年 / 5卷 / 12期
关键词
Action recognition; Deep learning; Convolutional neural network; Attention; BIDIRECTIONAL LSTM; CLASSIFICATION;
D O I
10.1007/s42452-023-05568-5
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
Given the prevalence of surveillance cameras in our daily lives, human action recognition from videos holds significant practical applications. A persistent challenge in this field is to develop more efficient models capable of real-time recognition with high accuracy for widespread implementation. In this research paper, we introduce a novel human action recognition model named Context-Aware Memory Attention Network (CAMA-Net), which eliminates the need for optical flow extraction and 3D convolution which are computationally intensive. By removing these components, CAMA-Net achieves superior efficiency compared to many existing approaches in terms of computation efficiency. A pivotal component of CAMA-Net is the Context-Aware Memory Attention Module, an attention module that computes the relevance score between key-value pairs obtained from the 2D ResNet backbone. This process establishes correspondences between video frames. To validate our method, we conduct experiments on four well-known action recognition datasets: ActivityNet, Diving48, HMDB51 and UCF101. The experimental results convincingly demonstrate the effectiveness of our proposed model, surpassing the performance of existing 2D-CNN based baseline models.Article HighlightsRecent human action recognition models are not yet ready for practical applications due to high computation needs.We propose a 2D CNN-based human action recognition method to reduce the computation load.The proposed method achieves competitive performance compared to most SOTA 2D CNN-based methods on public datasets.
引用
收藏
页数:12
相关论文
共 50 条
  • [1] Towards efficient video-based action recognition: context-aware memory attention network
    Thean Chun Koh
    Chai Kiat Yeo
    Xuan Jing
    Sunil Sivadas
    SN Applied Sciences, 2023, 5
  • [2] Context-Aware Memory Attention Network for Video-Based Action Recognition
    Koh, Thean Chun
    Yeo, Chai Kiat
    Vaitesswar, U. S.
    Jing, Xuan
    2022 IEEE 14TH IMAGE, VIDEO, AND MULTIDIMENSIONAL SIGNAL PROCESSING WORKSHOP (IVMSP), 2022,
  • [3] Context-Aware Attention Network for Human Emotion Recognition in Video
    Liu, Xiaodong
    Wang, Miao
    ADVANCES IN MULTIMEDIA, 2020, 2020
  • [4] Context-aware attention network for image recognition
    Jiaxu Leng
    Ying Liu
    Shang Chen
    Neural Computing and Applications, 2019, 31 : 9295 - 9305
  • [5] Context-aware attention network for image recognition
    Leng, Jiaxu
    Liu, Ying
    Chen, Shang
    NEURAL COMPUTING & APPLICATIONS, 2019, 31 (12): : 9295 - 9305
  • [6] CANet: Comprehensive Attention Network for video-based action recognition
    Gao, Xiong
    Chang, Zhaobin
    Ran, Xingcheng
    Lu, Yonggang
    KNOWLEDGE-BASED SYSTEMS, 2024, 296
  • [7] Context-aware Cascade Attention-based RNN for Video Emotion Recognition
    Sun, Man-Chin
    Hsu, Shih-Huan
    Yang, Min-Chun
    Chien, Jen-Hsien
    2018 FIRST ASIAN CONFERENCE ON AFFECTIVE COMPUTING AND INTELLIGENT INTERACTION (ACII ASIA), 2018,
  • [8] Stacked Multimodal Attention Network for Context-Aware Video Captioning
    Zheng, Yi
    Zhang, Yuejie
    Feng, Rui
    Zhang, Tao
    Fan, Weiguo
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (01) : 31 - 42
  • [9] Video-Based Eye Tracking to Detect the Attention Shift: A Computer Classroom Context-Aware System
    Kuo, Yung-Lung
    Lee, Jiann-Shu
    Hsieh, Min-Chai
    INTERNATIONAL JOURNAL OF DISTANCE EDUCATION TECHNOLOGIES, 2014, 12 (04) : 66 - 81
  • [10] Context-Aware Cross-Attention for Skeleton-Based Human Action Recognition
    Fan, Yanbo
    Weng, Shuchen
    Zhang, Yong
    Shi, Boxin
    Zhang, Yi
    IEEE ACCESS, 2020, 8 (08): : 15280 - 15290