FEXNet: Foreground Extraction Network for Human Action Recognition

被引:29
|
作者
Shen, Zhongwei [1 ]
Wu, Xiao-Jun [1 ]
Xu, Tianyang [1 ,2 ]
机构
[1] Jiangnan Univ, Sch Artificial Intelligence & Comp Sci, Wuxi 214122, Jiangsu, Peoples R China
[2] Univ Surrey, Ctr Vis Speech & Signal Proc, Guildford GU2 7XH, Surrey, England
基金
中国国家自然科学基金;
关键词
Convolutional neural networks; Spatiotemporal phenomena; Feature extraction; Three-dimensional displays; Solid modeling; Iron; Image recognition; Foreground-related features; spatiotemporal modeling; action recognition;
D O I
10.1109/TCSVT.2021.3103677
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
As most human actions in video sequences embody the continuous interactions between foregrounds rather than the background scene, it is significant to disentangle these foregrounds from the background for advanced action recognition systems. In this paper, therefore, we propose a Foreground EXtraction (FEX) block to explicitly model the foreground clues to achieve effective management of action subjects. In particular, the designed FEX block contains two components. The first part is a Foreground Enhancement (FE) module, which highlights the potential feature channels related to the action attributes, providing channel-level refinement for the following spatiotemporal modeling. The second phase is a Scene Segregation (SS) module, which splits feature maps into foreground and background. Specifically, a temporal model with dynamic enhancement is constructed for the foreground part, reflecting the essential nature of the action category. While the background is modeled using simple spatial convolutions, mapping the inputs to the consistent feature space. The FEX blocks can be inserted into existing 2D CNNs (denoted as FEXNet) for spatiotemporal modeling, concentrating on the foreground clues for effective action inference. Our experiments performed on Something-Something V1, V2 and Kinetics400 verify the effectiveness of the proposed method.
引用
收藏
页码:3141 / 3151
页数:11
相关论文
共 50 条
  • [31] Hybrid deep neural network model for human action recognition
    Ijjina, Earnest Paul
    Mohan, Chalavadi Krishna
    APPLIED SOFT COMPUTING, 2016, 46 : 936 - 952
  • [32] Deep network for human action recognition using Weber motion
    Chaudhary, Sachin
    Murala, Subrahmanyam
    NEUROCOMPUTING, 2019, 367 : 207 - 216
  • [33] Multi-mode neural network for human action recognition
    Zhao, Haohua
    Xue, Weichen
    Li, Xiaobo
    Gu, Zhangxuan
    Niu, Li
    Zhang, Liqing
    IET COMPUTER VISION, 2020, 14 (08) : 587 - 596
  • [34] A heterogeneous two-stream network for human action recognition
    Liao, Shengbin
    Wang, Xiaofeng
    Yang, ZongKai
    AI COMMUNICATIONS, 2023, 36 (03) : 219 - 233
  • [35] Human Action Recognition Based on Skeleton and Convolutional Neural Network
    Yang, Yusi
    Cai, Zhuohao
    Yu, Yingdong
    Wu, Tong
    Lin, Lan
    2019 PHOTONICS & ELECTROMAGNETICS RESEARCH SYMPOSIUM - FALL (PIERS - FALL), 2019, : 1109 - 1112
  • [36] Secure human action recognition by encrypted neural network inference
    Miran Kim
    Xiaoqian Jiang
    Kristin Lauter
    Elkhan Ismayilzada
    Shayan Shams
    Nature Communications, 13
  • [37] Research on Human Action Recognition Based on Convolutional Neural Network
    Wang, Peng
    Yang, Yuliang
    Li, Wanchong
    Zhang, Linhao
    Wang, Mengyuan
    Zhang, Xiaobo
    Zhu, Mengyu
    2019 28TH WIRELESS AND OPTICAL COMMUNICATIONS CONFERENCE (WOCC), 2019, : 28 - 32
  • [38] Footprint Extraction and Sports Training Action Recognition Based on Wireless Network Communication
    Jiang, Lu
    WIRELESS COMMUNICATIONS & MOBILE COMPUTING, 2022, 2022
  • [39] Collaborative Foreground, Background, and Action Modeling Network for Weakly Supervised Temporal Action Localization
    Moniruzzaman, Md.
    Yin, Zhaozheng
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2023, 33 (11) : 6939 - 6951
  • [40] Human Action Invarianceness for Human Action Recognition
    Sjarif, Nilam Nur Amir
    Shamsuddin, Siti Mariyam
    2015 9TH INTERNATIONAL CONFERENCE ON SOFTWARE, KNOWLEDGE, INFORMATION MANAGEMENT AND APPLICATIONS (SKIMA), 2015,