ASNet: Auto-Augmented Siamese Neural Network for Action Recognition

被引:3
|
作者
Zhang, Yujia [1 ]
Po, Lai-Man [1 ]
Xiong, Jingjing [1 ]
Rehman, Yasar Abbas Ur [2 ]
Cheung, Kwok-Wai [3 ]
机构
[1] City Univ Hong Kong, Dept Elect Engn, Hong Kong, Peoples R China
[2] TCL Corp Res Co Ltd, Hong Kong, Peoples R China
[3] Hang Seng Univ Hong Kong, Sch Commun, Hong Kong, Peoples R China
关键词
action recognition; 3D-CNN; deep reinforcement learning; data augmentation; ATTENTION NETWORK; MASK;
D O I
10.3390/s21144720
中图分类号
O65 [分析化学];
学科分类号
070302 ; 081704 ;
摘要
Human action recognition methods in videos based on deep convolutional neural networks usually use random cropping or its variants for data augmentation. However, this traditional data augmentation approach may generate many non-informative samples (video patches covering only a small part of the foreground or only the background) that are not related to a specific action. These samples can be regarded as noisy samples with incorrect labels, which reduces the overall action recognition performance. In this paper, we attempt to mitigate the impact of noisy samples by proposing an Auto-augmented Siamese Neural Network (ASNet). In this framework, we propose backpropagating salient patches and randomly cropped samples in the same iteration to perform gradient compensation to alleviate the adverse gradient effects of non-informative samples. Salient patches refer to the samples containing critical information for human action recognition. The generation of salient patches is formulated as a Markov decision process, and a reinforcement learning agent called SPA (Salient Patch Agent) is introduced to extract patches in a weakly supervised manner without extra labels. Extensive experiments were conducted on two well-known datasets UCF-101 and HMDB-51 to verify the effectiveness of the proposed SPA and ASNet.
引用
收藏
页数:20
相关论文
共 50 条
  • [31] Intelligent open-set MIMO recognition in OWC using a Siamese neural network
    Zhao, Yinan
    Chen, Chen
    Cao, Hailin
    Zeng, Zhihong
    Liu, Min
    Haas, Harald
    OPTICS LETTERS, 2024, 49 (24) : 7060 - 7063
  • [32] Signature Recognition using Siamese Neural Networks
    Krishna, Voruganti Ajay
    Reddy, AtthapuramAkshay
    Nagajyothi, D.
    2021 IEEE INTERNATIONAL CONFERENCE ON MOBILE NETWORKS AND WIRELESS COMMUNICATIONS (ICMNWC), 2021,
  • [33] Symbolized flight action recognition based on neural network
    Fang W.
    Wang Y.
    Yan W.
    Lin C.
    Xi Tong Gong Cheng Yu Dian Zi Ji Shu/Systems Engineering and Electronics, 2022, 44 (03): : 737 - 745
  • [34] A Jeap-BiLSTM Neural Network for Action Recognition
    Tan, Lunzheng
    Liu, Yanfei
    Xia, Limin
    Chen, Shangsheng
    Zhou, Zhanben
    INTERNATIONAL JOURNAL OF IMAGE AND GRAPHICS, 2025, 25 (02)
  • [35] A Strict Pyramidal Deep Neural Network for Action Recognition
    Ullah, Ihsan
    Petrosino, Alfredo
    IMAGE ANALYSIS AND PROCESSING - ICIAP 2015, PT I, 2015, 9279 : 236 - 245
  • [36] Skeleton Based Action Recognition with Convolutional Neural Network
    Du, Yong
    Fu, Yun
    Wang, Liang
    PROCEEDINGS 3RD IAPR ASIAN CONFERENCE ON PATTERN RECOGNITION ACPR 2015, 2015, : 579 - 583
  • [37] Temporal Spiking Recurrent Neural Network for Action Recognition
    Wang, Wei
    Hao, Siyuan
    Wei, Yunchao
    Xia, Shengtao
    Feng, Jiashi
    Sebe, Nicu
    IEEE ACCESS, 2019, 7 : 117165 - 117175
  • [38] AutoGesNet: Auto Gesture Recognition Network Based on Neural Architecture Search
    Li, Yinqi
    Xu, Lu
    Shu, Weihua
    Tao, Ji'an
    Mei, Kuizhi
    2020 12TH INTERNATIONAL CONFERENCE ON ADVANCED COMPUTATIONAL INTELLIGENCE (ICACI), 2020, : 257 - 262
  • [39] Graphic Symbol Recognition using Auto Associative Neural Network Model
    Gellaboina, Mahesh Kumar
    Venkoparao, Vijendran G.
    ICAPR 2009: SEVENTH INTERNATIONAL CONFERENCE ON ADVANCES IN PATTERN RECOGNITION, PROCEEDINGS, 2009, : 297 - 301
  • [40] Human face recognition based on convolutional neural network and augmented dataset
    Lu, Peng
    Song, Baoye
    Xu, Lin
    SYSTEMS SCIENCE & CONTROL ENGINEERING, 2021, 9 (S2) : 29 - 37