Unsupervised video-based action recognition using two-stream generative adversarial network

被引:0
|
作者
Wei Lin
Huanqiang Zeng
Jianqing Zhu
Chih-Hsien Hsia
Junhui Hou
Kai-Kuang Ma
机构
[1] Huaqiao University,School of Engineering and School of Information Science and Engineering
[2] Huaqiao University,School of Engineering
[3] Ilan University,Department of Computer Science and Information Engineering
[4] The City University of Hong Kong,Department of Computer Science
[5] Nanyang Technological University,School of Electrical and Electronic Engineering
来源
关键词
Action recognition; Two-stream generative adversarial network; Unsupervised learning;
D O I
暂无
中图分类号
学科分类号
摘要
Video-based action recognition faces many challenges, such as complex and varied dynamic motion, spatio-temporal similar action factors, and manual labeling of archived videos over large datasets. How to extract discriminative spatio-temporal action features in videos with resisting the effect of similar factors in an unsupervised manner is pivotal. For that, this paper proposes an unsupervised video-based action recognition method, called two-stream generative adversarial network (TS-GAN), which comprehensively learns the static texture and dynamic motion information inherited in videos with taking the detail information and global information into account. Specifically, the extraction of the spatio-temporal information in videos is achieved by a two-stream GAN. Considering that proper attention to detail is capable of alleviating the influence of spatio-temporal similar factors to the network, a global-detailed layer is proposed to resist similar factors via fusing intermediate features (i.e., detailed action information) and high-level semantic features (i.e., global action information). It is worthwhile of mentioning that the proposed TS-GAN does not require complex pretext tasks or the construction of positive and negative sample pairs, compared with recent unsupervised video-based action recognition methods. Extensive experiments conducted on the UCF101 and HMDB51 datasets have demonstrated that the proposed TS-GAN is superior to multiple classical and state-of-the-art unsupervised action recognition methods.
引用
收藏
页码:5077 / 5091
页数:14
相关论文
共 50 条
  • [21] TSGAN: A Two-Stream Generative Adversarial Network for Bit-Depth Expansion
    Yang, Zhen
    Song, Li
    Xie, Rong
    Zhang, Wenjun
    Li, Lin
    Feng, Yanan
    2020 IEEE INTERNATIONAL SYMPOSIUM ON BROADBAND MULTIMEDIA SYSTEMS AND BROADCASTING (BMSB), 2020,
  • [22] Efficient Two-Stream Network for Online Video Action Segmentation
    Kang, Min-Seok
    Park, Rae-Hong
    Park, Hyung-Min
    IEEE ACCESS, 2022, 10 : 90635 - 90646
  • [23] Human Action Recognition based on Two-Stream Ind Recurrent Neural Network
    Ge Penghua
    Zhi Min
    TENTH INTERNATIONAL CONFERENCE ON GRAPHICS AND IMAGE PROCESSING (ICGIP 2018), 2019, 11069
  • [24] A two-stream heterogeneous network for action recognition based on skeleton and RGB modalities
    Liu, Kai
    Gao, Lei
    Khan, Naimul Mefraz
    Qi, Lin
    Guan, Ling
    23RD IEEE INTERNATIONAL SYMPOSIUM ON MULTIMEDIA (ISM 2021), 2021, : 87 - 91
  • [25] Enhanced Spatial Stream of Two-Stream Network Using Optical Flow for Human Action Recognition
    Khan, Shahbaz
    Hassan, Ali
    Hussain, Farhan
    Perwaiz, Aqib
    Riaz, Farhan
    Alsabaan, Maazen
    Abdul, Wadood
    APPLIED SCIENCES-BASEL, 2023, 13 (14):
  • [26] Hidden Two-Stream Collaborative Learning Network for Action Recognition
    Zhou, Shuren
    Chen, Le
    Sugumaran, Vijayan
    CMC-COMPUTERS MATERIALS & CONTINUA, 2020, 63 (03): : 1545 - 1561
  • [27] Video-Based Fall Detection Using Human Pose and Constrained Generative Adversarial Network
    Wu, Lian
    Huang, Chao
    Fei, Lunke
    Zhao, Shuping
    Zhao, Jianchuan
    Cui, Zhongwei
    Xu, Yong
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2024, 34 (04) : 2179 - 2194
  • [28] Going deeper with two-stream ConvNets for action recognition in video surveillance
    Han, Yamin
    Zhang, Peng
    Zhuo, Tao
    Huang, Wei
    Zhang, Yanning
    PATTERN RECOGNITION LETTERS, 2018, 107 : 83 - 90
  • [29] Unsupervised Transfer Learning For Video Prediction Based on Generative Adversarial Network
    Shi, Jiwen
    Zhu, Qiuguo
    Wu, Jun
    2021 27TH INTERNATIONAL CONFERENCE ON MECHATRONICS AND MACHINE VISION IN PRACTICE (M2VIP), 2021,
  • [30] Skeleton-based Action Recognition Using Two-stream Graph Convolutional Network with Pose Refinement
    Zheng, Biao
    Chen, Luefeng
    Wu, Min
    Pedrycz, Witold
    Hirota, Kaoru
    2022 41ST CHINESE CONTROL CONFERENCE (CCC), 2022, : 6353 - 6356