Adversarial Attention Networks for Early Action Recognition

被引:0
|
作者
Zhang, Hong-Bo [1 ]
Pan, Wei-Xiang [1 ]
Du, Ji-Xiang [2 ]
Lei, Qing [3 ,4 ]
Chen, Yan [2 ]
Liu, Jing-Hua [3 ,4 ]
机构
[1] Huaqiao Univ, Dept Comp Sci & Technol, Xiamen 361000, Peoples R China
[2] Huaqiao Univ, Fujian Key Lab Big Data Intelligence & Secur, Xiamen 361000, Peoples R China
[3] Huaqiao Univ, Xiamen Key Lab Comp Vis & Pattern Recognit, Xiamen 361000, Peoples R China
[4] Huaqiao Univ, Fujian Prov Univ, Key Lab Comp Vis & Machine Learning, Xiamen 361000, Peoples R China
基金
中国国家自然科学基金;
关键词
Early action recognition; adversarial attention network; cross attention generator; self attention discriminator; feature fusion module;
D O I
10.1109/TETCI.2024.3437240
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Early action recognition endeavors to deduce the ongoing action by observing partial video, presenting a formidable challenge due to limited information available in the initial stages. To tackle this challenge, we introduce an innovative adversarial attention network based on generative adversarial networks. This network leverages the characteristics of both the generator and discriminator to generate unobserved action information from partial video input. The proposed method comprises a cross attention generator, self Attention discriminator, and feature fusion module. The cross attention generator captures temporal relationships in input action sequences, generating discriminative unobserved action information. The self attention discriminator adds global attention to the input sequence, capturing global context information for accurate evaluation of consistency in generated unobserved feature from cross attention generator. Finally, the feature fusion module helps the model obtain richer and more comprehensive feature representations. The proposed method is evaluated through experiments on the HMDB51, UCF101 and Something-Something v2 datasets. Experimental results demonstrate that the proposed approach outperforms existing methods across different observation ratios. Detailed ablation studies confirm the effectiveness of each component in the proposed method.
引用
下载
收藏
页数:14
相关论文
共 50 条
  • [41] An efficient attention module for 3d convolutional neural networks in action recognition
    Guanghao Jiang
    Xiaoyan Jiang
    Zhijun Fang
    Shanshan Chen
    Applied Intelligence, 2021, 51 : 7043 - 7057
  • [42] Human action recognition using two-stream attention based LSTM networks
    Dai, Cheng
    Liu, Xingang
    Lai, Jinfeng
    APPLIED SOFT COMPUTING, 2020, 86
  • [43] Two-stream Flow-guided Convolutional Attention Networks for Action Recognition
    Tran, An
    Cheong, Loong-Fah
    2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS (ICCVW 2017), 2017, : 3110 - 3119
  • [44] An efficient attention module for 3d convolutional neural networks in action recognition
    Jiang, Guanghao
    Jiang, Xiaoyan
    Fang, Zhijun
    Chen, Shanshan
    APPLIED INTELLIGENCE, 2021, 51 (10) : 7043 - 7057
  • [45] Spatial-temporal graph attention networks for skeleton-based action recognition
    Huang, Qingqing
    Zhou, Fengyu
    He, Jiakai
    Zhao, Yang
    Qin, Runze
    JOURNAL OF ELECTRONIC IMAGING, 2020, 29 (05)
  • [46] Adaptive Attention Memory Graph Convolutional Networks for Skeleton-Based Action Recognition
    Liu, Di
    Xu, Hui
    Wang, Jianzhong
    Lu, Yinghua
    Kong, Jun
    Qi, Miao
    SENSORS, 2021, 21 (20)
  • [47] Continual Activity Recognition with Generative Adversarial Networks
    Ye, Juan
    Nakwijit, Pakawat
    Schiemer, Martin
    Jha, Saurav
    Zambonelli, Franco
    ACM TRANSACTIONS ON INTERNET OF THINGS, 2021, 2 (02):
  • [48] Visual Attention Adversarial Networks for Chinese Font Translation
    Li, Te
    Yang, Fang
    Song, Yao
    ELECTRONICS, 2023, 12 (06)
  • [49] Asymmetric Generative Adversarial Networks with a New Attention Mechanism
    Chen, Jian
    Liu, Gang
    Ke, Aihua
    2022 ASIA CONFERENCE ON ALGORITHMS, COMPUTING AND MACHINE LEARNING (CACML 2022), 2022, : 186 - 192