FEXNet: Foreground Extraction Network for Human Action Recognition

被引:29
|
作者
Shen, Zhongwei [1 ]
Wu, Xiao-Jun [1 ]
Xu, Tianyang [1 ,2 ]
机构
[1] Jiangnan Univ, Sch Artificial Intelligence & Comp Sci, Wuxi 214122, Jiangsu, Peoples R China
[2] Univ Surrey, Ctr Vis Speech & Signal Proc, Guildford GU2 7XH, Surrey, England
基金
中国国家自然科学基金;
关键词
Convolutional neural networks; Spatiotemporal phenomena; Feature extraction; Three-dimensional displays; Solid modeling; Iron; Image recognition; Foreground-related features; spatiotemporal modeling; action recognition;
D O I
10.1109/TCSVT.2021.3103677
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
As most human actions in video sequences embody the continuous interactions between foregrounds rather than the background scene, it is significant to disentangle these foregrounds from the background for advanced action recognition systems. In this paper, therefore, we propose a Foreground EXtraction (FEX) block to explicitly model the foreground clues to achieve effective management of action subjects. In particular, the designed FEX block contains two components. The first part is a Foreground Enhancement (FE) module, which highlights the potential feature channels related to the action attributes, providing channel-level refinement for the following spatiotemporal modeling. The second phase is a Scene Segregation (SS) module, which splits feature maps into foreground and background. Specifically, a temporal model with dynamic enhancement is constructed for the foreground part, reflecting the essential nature of the action category. While the background is modeled using simple spatial convolutions, mapping the inputs to the consistent feature space. The FEX blocks can be inserted into existing 2D CNNs (denoted as FEXNet) for spatiotemporal modeling, concentrating on the foreground clues for effective action inference. Our experiments performed on Something-Something V1, V2 and Kinetics400 verify the effectiveness of the proposed method.
引用
收藏
页码:3141 / 3151
页数:11
相关论文
共 50 条
  • [1] Human Action Recognition Based on Foreground Trajectory and Motion Difference Descriptors
    Dong, Suge
    Hu, Daidi
    Li, Ruijun
    Ge, Mingtao
    APPLIED SCIENCES-BASEL, 2019, 9 (10):
  • [2] Human Action Recognition across Datasets by Foreground-weighted Histogram Decomposition
    Sultani, Waqas
    Saleemi, Imran
    2014 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2014, : 764 - 771
  • [3] Saliency-based foreground trajectory extraction using multiscale hybrid masks for action recognition
    Zhang, Guoliang
    Jia, Songmin
    Zhang, Xiangyin
    Li, Xiuzhi
    JOURNAL OF ELECTRONIC IMAGING, 2018, 27 (05)
  • [4] Realistic action recognition with salient foreground trajectories
    Yi, Yang
    Zheng, Zhenxian
    Lin, Maoqing
    EXPERT SYSTEMS WITH APPLICATIONS, 2017, 75 : 44 - 55
  • [5] A spatiotemporal and motion information extraction network for action recognition
    Wang, Wei
    Wang, Xianmin
    Zhou, Mingliang
    Wei, Xuekai
    Li, Jing
    Ren, Xiaojun
    Zong, Xuemei
    WIRELESS NETWORKS, 2024, 30 (06) : 5389 - 5405
  • [6] Complex Network-based features extraction in RGB-D human action recognition
    Barkoky, Alaa
    Charkari, Nasrollah Moghaddam
    JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2022, 82
  • [7] Refined Spatial Network for Human Action Recognition
    Wu, Chunlei
    Cao, Haiwen
    Zhang, Weishan
    Wang, Leiquan
    Wei, Yiwei
    Peng, Zexin
    IEEE ACCESS, 2019, 7 : 111043 - 111052
  • [8] An Improved Action Recognition Network With Temporal Extraction and Feature Enhancement
    Jiang, Jie
    Zhang, Yi
    IEEE ACCESS, 2022, 10 : 13926 - 13935
  • [9] Moving Foreground-Aware Visual Attention and Key Volume Mining for Human Action Recognition
    Zhang, Junxuan
    Hu, Haifeng
    Lu, Xinlong
    ACM TRANSACTIONS ON MULTIMEDIA COMPUTING COMMUNICATIONS AND APPLICATIONS, 2019, 15 (03)
  • [10] Integrating Human Parsing and Pose Network for Human Action Recognition
    Ding, Runwei
    Wen, Yuhang
    Liu, Jinfu
    Dai, Nan
    Meng, Fanyang
    Liu, Mengyuan
    ARTIFICIAL INTELLIGENCE, CICAI 2023, PT I, 2024, 14473 : 182 - 194