Spatial–temporal injection network: exploiting auxiliary losses for action recognition with apparent difference and self-attention

被引:0
|
作者
Haiwen Cao
Chunlei Wu
Jing Lu
Jie Wu
Leiquan Wang
机构
[1] China University of Petroleum,College of Computer Science and Technology
来源
关键词
Action recognition; Apparent difference module; Self-attention mechanism; Spatiotemporal Features;
D O I
暂无
中图分类号
学科分类号
摘要
Two-stream convolutional networks have shown strong performance in action recognition. However, both spatial and temporal features in two-stream are learned separately. There has been almost no consideration for the different characteristics of the spatial and temporal streams, which are performed on the same operations. In this paper, we build upon two-stream convolutional networks and propose a novel spatial–temporal injection network (STIN) with two different auxiliary losses. To build spatial–temporal features as the video representation, the apparent difference module is designed to model the auxiliary temporal constraints on spatial features in spatial injection network. The self-attention mechanism is used to attend to the interested areas in the temporal injection stream, which reduces the optical flow noise influence of uninterested region. Then, these auxiliary losses enable efficient training of two complementary streams which can capture interactions between the spatial and temporal information from different perspectives. Experiments conducted on the two well-known datasets—UCF101 and HMDB51—demonstrate the effectiveness of the proposed STIN.
引用
收藏
页码:1173 / 1180
页数:7
相关论文
共 50 条
  • [21] Exploiting Attention-Consistency Loss For Spatial-Temporal Stream Action Recognition
    Xu, Haotian
    Jin, Xiaobo
    Wang, Qiufeng
    Hussain, Amir
    Huang, Kaizhu
    ACM TRANSACTIONS ON MULTIMEDIA COMPUTING COMMUNICATIONS AND APPLICATIONS, 2022, 18 (02)
  • [22] Recurrent Spatial-Temporal Attention Network for Action Recognition in Videos
    Du, Wenbin
    Wang, Yali
    Qiao, Yu
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2018, 27 (03) : 1347 - 1360
  • [23] Extreme Low Resolution Action Recognition with Spatial-Temporal Multi-Head Self-Attention and Knowledge Distillation
    Purwanto, Didik
    Pramono, Rizard Renanda Adhi
    Chen, Yie-Tarng
    Fang, Wen-Hsien
    2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS (ICCVW), 2019, : 961 - 969
  • [24] Dual Stream Spatio-Temporal Motion Fusion With Self-Attention For Action Recognition
    Jalal, Md Asif
    Aftab, Waqas
    Moore, Roger K.
    Mihaylova, Lyudmila
    2019 22ND INTERNATIONAL CONFERENCE ON INFORMATION FUSION (FUSION 2019), 2019,
  • [25] Spatio-Temporal Action Detector with Self-Attention
    Ma, Xurui
    Luo, Zhigang
    Zhang, Xiang
    Liao, Qing
    Shen, Xingyu
    Wang, Mengzhu
    2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [26] SELF-ATTENTION GUIDED DEEP FEATURES FOR ACTION RECOGNITION
    Xiao, Renyi
    Hou, Yonghong
    Guo, Zihui
    Li, Chuankun
    Wang, Pichao
    Li, Wanqing
    2019 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO (ICME), 2019, : 1060 - 1065
  • [27] Spatial-Temporal Attention for Action Recognition
    Sun, Dengdi
    Wu, Hanqing
    Ding, Zhuanlian
    Luo, Bin
    Tang, Jin
    ADVANCES IN MULTIMEDIA INFORMATION PROCESSING, PT I, 2018, 11164 : 854 - 864
  • [28] SelfGCN: Graph Convolution Network With Self-Attention for Skeleton-Based Action Recognition
    Wu, Zhize
    Sun, Pengpeng
    Chen, Xin
    Tang, Keke
    Xu, Tong
    Zou, Le
    Wang, Xiaofeng
    Tan, Ming
    Cheng, Fan
    Weise, Thomas
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2024, 33 : 4391 - 4403
  • [29] MGSAN: multimodal graph self-attention network for skeleton-based action recognition
    Wang, Junyi
    Li, Ziao
    Liu, Bangli
    Cai, Haibin
    Saada, Mohamad
    Meng, Qinggang
    MULTIMEDIA SYSTEMS, 2024, 30 (06)
  • [30] Temporal Self-Attention Network for Medical Concept Embedding
    Peng, Xueping
    Long, Guodong
    Shen, Tao
    Wang, Sen
    Jiang, Jing
    Blumenstein, Michael
    2019 19TH IEEE INTERNATIONAL CONFERENCE ON DATA MINING (ICDM 2019), 2019, : 498 - 507