Deep multiple aggregation networks for action recognition

被引:0
|
作者
Ahmed Mazari
Hichem Sahbi
机构
[1] Sorbonne University,CNRS, LIP6
关键词
Multiple aggregation design; 2-Stream networks; Action recognition;
D O I
暂无
中图分类号
学科分类号
摘要
Most of the current action recognition algorithms are based on deep networks which stack multiple convolutional, pooling and fully connected layers. While convolutional and fully connected operations have been widely studied in the literature, the design of pooling operations that handle action recognition, with different sources of temporal granularity in action categories, has comparatively received less attention, and existing solutions rely mainly on max or averaging operations. The latter are clearly powerless to fully exhibit the actual temporal granularity of action categories and thereby constitute a bottleneck in classification performances. In this paper, we introduce a novel hierarchical pooling design that captures different levels of temporal granularity in action recognition. Our design principle is coarse-to-fine and achieved using a tree-structured network; as we traverse this network top-down, pooling operations are getting less invariant but timely more resolute and well localized. Learning the combination of operations in this network—which best fits a given ground-truth—is obtained by solving a constrained minimization problem whose solution corresponds to the distribution of weights that capture the contribution of each level (and thereby temporal granularity) in the global hierarchical pooling process. Besides being principled and well grounded, the proposed hierarchical pooling is also video-length and resolution agnostic. Extensive experiments conducted on the challenging UCF-101, HMDB-51 and JHMDB-21 databases corroborate all these statements.
引用
收藏
相关论文
共 50 条
  • [21] Continuous Action Recognition in Manufacturing Contexts by Deep Graph Convolutional Networks
    Maselli, M. V.
    Marani, R.
    Cicirelli, G.
    D'Orazio, T.
    [J]. INTELLIGENT SYSTEMS AND APPLICATIONS, VOL 4, INTELLISYS 2023, 2024, 825 : 156 - 173
  • [22] Human Action Recognition in Video Sequences Using Deep Belief Networks
    Abdellaoui, Mehrez
    Douik, Ali
    [J]. TRAITEMENT DU SIGNAL, 2020, 37 (01) : 37 - 44
  • [23] Human Action Recognition Using Hybrid Deep Evolving Neural Networks
    Dasari, Pavan
    Zhang, Li
    Yu, Yonghong
    Huang, Haoqian
    Gao, Rong
    [J]. 2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,
  • [24] Deep Image-to-Video Adaptation and Fusion Networks for Action Recognition
    Liu, Yang
    Lu, Zhaoyang
    Li, Jing
    Yang, Tao
    Yao, Chao
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2020, 29 : 3168 - 3182
  • [25] Human action recognition in videos with articulated pose information by deep networks
    M. Farrajota
    João M. F. Rodrigues
    J. M. H. du Buf
    [J]. Pattern Analysis and Applications, 2019, 22 : 1307 - 1318
  • [26] Temporal Segment Networks: Towards Good Practices for Deep Action Recognition
    Wang, Limin
    Xiong, Yuanjun
    Wang, Zhe
    Qiao, Yu
    Lin, Dahua
    Tang, Xiaoou
    Van Gool, Luc
    [J]. COMPUTER VISION - ECCV 2016, PT VIII, 2016, 9912 : 20 - 36
  • [27] Deep Manifold Learning Combined With Convolutional Neural Networks for Action Recognition
    Chen, Xin
    Weng, Jian
    Lu, Wei
    Xu, Jiaming
    Weng, Jiasi
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2018, 29 (09) : 3938 - 3952
  • [28] Using Deep Multiple Instance Learning for Action Recognition in Still Images
    Bas, Cagdas
    Zalluhoglu, Cemil
    Ikizler-Cinbis, Nazli
    [J]. 2017 25TH SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS CONFERENCE (SIU), 2017,
  • [29] Deep Neural Networks Using Capsule Networks and Skeleton-Based Attentions for Action Recognition
    Ha, Manh-Hung
    Chen, Oscal Tzyh-Chiang
    [J]. IEEE ACCESS, 2021, 9 : 6164 - 6178
  • [30] Modelling Human Body Pose for Action Recognition Using Deep Neural Networks
    Li, Chengyang
    Tong, Ruofeng
    Tang, Min
    [J]. ARABIAN JOURNAL FOR SCIENCE AND ENGINEERING, 2018, 43 (12) : 7777 - 7788