Action-aware Masking Network with Group-based Attention for Temporal Action Localization

被引:3
|
作者
Kang, Tae-Kyung [1 ]
Lee, Gun-Hee [2 ]
Jin, Kyung-Min [1 ]
Lee, Seong-Whan [1 ]
机构
[1] Korea Univ, Dept Artificial Intelligence, Seoul, South Korea
[2] Korea Univ, Dept Comp Sci & Engn, Seoul, South Korea
关键词
D O I
10.1109/WACV56688.2023.00600
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Temporal Action Localization (TAL) is a significant and challenging task that searches for subtle human activities in an untrimmed video. To extract snippet-level video features, existing TAL methods commonly use video encoders pre-trained on short-video classification datasets. However, the snippet-level features can incur ambiguity between consecutive frames due to short and poor temporal information, disrupting the precise prediction of action instances. Several methods incorporating temporal relations have been proposed to mitigate this problem; however, they still suffer from poor video features. To address this issue, we propose a novel temporal action localization framework called an Action-aware Masking Network (AMNet). Our method simultaneously refines video features using action-aware attention and considers inherent temporal relations using self-attention and cross-attention mechanisms. First, we present an Action Masking Encoder (AME) that generates an action-aware mask to represent positive characteristics, which is then used to refine snippet-level features to be more salient around actions. Second, we design a Group Attention Module (GAM), which models relations of temporal information and exchanges mutual information by dividing the features into two groups, i.e., long and short-groups. Extensive experiments and ablation studies on two primary benchmark datasets demonstrate the effectiveness of AMNet, and our method achieves state-of-the-art performances on THUMOS-14 and ActivityNet1.3.
引用
收藏
页码:6047 / 6056
页数:10
相关论文
共 50 条
  • [41] Action Completeness Modeling with Background Aware Networks for Weakly-Supervised Temporal Action Localization
    Moniruzzaman, Md
    Yin, Zhaozheng
    He, Zhihai
    Qin, Ruwen
    Leu, Ming C.
    MM '20: PROCEEDINGS OF THE 28TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, 2020, : 2166 - 2174
  • [42] HAAN: Human Action Aware Network for Multi-label Temporal Action Detection
    Gao, Zikai
    Qiao, Peng
    Dou, Yong
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 5059 - 5069
  • [43] An attention-based bidirectional GRU network for temporal action proposals generation
    Xiaoxin Liao
    Jingyi Yuan
    Zemin Cai
    Jian-huang Lai
    The Journal of Supercomputing, 2023, 79 : 8322 - 8339
  • [44] Advancing Temporal Action Localization with a Boundary Awareness Network
    Gu, Jialiang
    Yi, Yang
    Wang, Min
    ELECTRONICS, 2024, 13 (06)
  • [45] Temporal Action Localization With Coarse-to-Fine Network
    Zhejiang Industry Polytechnic College, Department of Design and Art, Shaoxing
    312000, China
    不详
    310018, China
    IEEE Access, 2022, (96378-96387)
  • [46] Attention-Based Temporal Weighted Convolutional Neural Network for Action Recognition
    Zang, Jinliang
    Wang, Le
    Liu, Ziyi
    Zhang, Qilin
    Niu, Zhenxing
    Hua, Gang
    Zheng, Nanning
    ARTIFICIAL INTELLIGENCE APPLICATIONS AND INNOVATIONS, AIAI 2018, 2018, 519 : 97 - 108
  • [47] CRNet: Centroid Radiation Network for Temporal Action Localization
    Ding, Xinpeng
    Wang, Nannan
    Li, Jie
    Gao, Xinbo
    PATTERN RECOGNITION AND COMPUTER VISION, PT I, 2021, 13019 : 29 - 41
  • [48] Temporal Action Localization With Coarse-to-Fine Network
    Zhang, Min
    Hu, Haiyang
    Li, Zhongjin
    IEEE ACCESS, 2022, 10 : 96378 - 96387
  • [49] An attention-based bidirectional GRU network for temporal action proposals generation
    Liao, Xiaoxin
    Yuan, Jingyi
    Cai, Zemin
    Lai, Jian-huang
    JOURNAL OF SUPERCOMPUTING, 2023, 79 (08): : 8322 - 8339
  • [50] A Spatio-Temporal Motion Network for Action Recognition Based on Spatial Attention
    Yang, Qi
    Lu, Tongwei
    Zhou, Huabing
    ENTROPY, 2022, 24 (03)