Multi-level Sequence GAN for Group Activity Recognition

被引:10
|
作者
Gammulle, Harshala [1 ]
Denman, Simon [1 ]
Sridharan, Sridha [1 ]
Fookes, Clinton [1 ]
机构
[1] Queensland Univ Technol, SAIVT, Image & Video Res Lab, Brisbane, Qld, Australia
来源
关键词
Group activity recognition; Generative adversarial networks; Long short term memory networks;
D O I
10.1007/978-3-030-20887-5_21
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We propose a novel semi supervised, Multi Level Sequential Generative Adversarial Network (MLS-GAN) architecture for group activity recognition. In contrast to previous works which utilise manually annotated individual human action predictions, we allow the models to learn it's own internal representations to discover pertinent subactivities that aid the final group activity recognition task. The generator is fed with person-level and scene-level features that are mapped temporally through LSTM networks. Action-based feature fusion is performed through novel gated fusion units that are able to consider long-term dependancies, exploring the relationships among all individual actions, to learn an intermediate representation or `action code' for the current group activity. The network achieves it's semi-supervised behaviour by allowing it to perform group action classification together with the adversarial real/fake validation. We perform extensive evaluations on different architectural variants to demonstrate the importance of the proposed architecture. Furthermore, we show that utilising both person-level and scene-level features facilitates the group activity prediction better than using only person-level features. Our proposed architecture outperforms current state-of-the-art results for sports and pedestrian based classification tasks on Volleyball and Collective Activity datasets, showing it's flexible nature for effective learning of group activities (This research was supported by the Australian Research Council's Linkage Project LP140100282 "Improving Productivity and Efficiency of Australian Airports").
引用
收藏
页码:331 / 346
页数:16
相关论文
共 50 条
  • [21] Multi-level Contrast Network for Wearables-based Joint Activity Segmentation and Recognition
    Xia, Songpengcheng
    Chu, Lei
    Pei, Ling
    Yu, Wenxian
    Qiu, Robert C.
    2022 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM 2022), 2022, : 566 - 572
  • [22] A Multi-level Methodology for Developing UML Sequence Diagrams
    Song, Il-Yeol
    Khare, Ritu
    An, Yuan
    Hilsbos, Margaret
    CONCEPTUAL MODELING - ER 2008, PROCEEDINGS, 2008, 5231 : 114 - 127
  • [23] A Multi-Level Task Framework for Event Sequence Analysis
    Zinat, Kazi Tasnim
    Sakhamuri, Saimadhav Naga
    Chen, Aaron Sun
    Liu, Zhicheng
    IEEE Transactions on Visualization and Computer Graphics, 2025, 31 (01) : 842 - 852
  • [24] Multi-level Crowding and the Paradox of Object Recognition in Clutter
    Manassi, Mauro
    Whitney, David
    CURRENT BIOLOGY, 2018, 28 (03) : R127 - R133
  • [25] An Adaptive Multi-Level Framework for Forest Species Recognition
    Cavalin, Paulo R.
    Kapp, Marcelo N.
    Oliveira, Luiz E. S.
    2015 BRAZILIAN CONFERENCE ON INTELLIGENT SYSTEMS (BRACIS 2015), 2015, : 92 - 97
  • [26] Multi-level regulation of cellular recognition of viral dsRNA
    Peisley, Alys
    Hur, Sun
    CELLULAR AND MOLECULAR LIFE SCIENCES, 2013, 70 (11) : 1949 - 1963
  • [27] MULTI-LEVEL FEATURE ANALYSIS FOR SEMANTIC CATEGORY RECOGNITION
    Sridharan, Harini
    Cheriyadat, Anil
    2013 IEEE INTERNATIONAL GEOSCIENCE AND REMOTE SENSING SYMPOSIUM (IGARSS), 2013, : 4371 - 4374
  • [28] Multi-level contour combination features for shape recognition
    Yang, Chengzhuan
    Fang, Lincong
    Fei, Benjie
    Yu, Qian
    Wei, Hui
    COMPUTER VISION AND IMAGE UNDERSTANDING, 2023, 229
  • [29] An Ensemble Model for Multi-Level Speech Emotion Recognition
    Zheng, Chunjun
    Wang, Chunli
    Jia, Ning
    APPLIED SCIENCES-BASEL, 2020, 10 (01):
  • [30] Learning multi-level representations for affective image recognition
    Hao Zhang
    Dan Xu
    Gaifang Luo
    Kangjian He
    Neural Computing and Applications, 2022, 34 : 14107 - 14120