Learning discriminative motion feature for enhancing multi-modal action recognition

被引:0
|
作者
Yang, Jianyu [1 ]
Huang, Yao [1 ]
Shao, Zhanpeng [2 ]
Liu, Chunping [3 ]
机构
[1] School of Rail Transportation, Soochow University, Suzhou,215000, China
[2] School of Computer Science and Technology, Zhejiang University of Technology, Hangzhou,310023, China
[3] School of Computer Science and Technology, Soochow University, Suzhou,215000, China
关键词
D O I
暂无
中图分类号
学科分类号
摘要
Video action recognition is an important topic in computer vision tasks. Most of the existing methods use CNN-based models, and multiple modalities of image features are captured from the videos, such as static frames, dynamic images, and optical flow features. However, these mainstream features contain much static information including object and background information, where the motion information of the action itself is not distinguished and strengthened. In this work, a new kind of motion feature is proposed without static information for video action recognition. We propose a quantization of motion network based on the bag-of-feature method to learn significant and discriminative motion features. In the learned feature map, the object and background information is filtered out, even if the background is moving in the video. Therefore, the motion feature is complementary to the static image feature and the static information in the dynamic image and optical flow. A multi-stream classifier is built with the proposed motion feature and other features, and the performance of action recognition is enhanced comparing to other state-of-the-art methods. © 2021 Elsevier Inc.
引用
收藏
相关论文
共 50 条
  • [41] Learning a discriminative mid-level feature for action recognition
    Liu CuiWei
    Pei MingTao
    Wu XinXiao
    Kong Yu
    Jia YunDe
    SCIENCE CHINA-INFORMATION SCIENCES, 2014, 57 (05) : 1 - 13
  • [42] Learning a discriminative mid-level feature for action recognition
    CuiWei Liu
    MingTao Pei
    XinXiao Wu
    Yu Kong
    YunDe Jia
    Science China Information Sciences, 2014, 57 : 1 - 13
  • [43] Learning a discriminative mid-level feature for action recognition
    LIU CuiWei
    PEI MingTao
    WU XinXiao
    KONG Yu
    JIA YunDe
    Science China(Information Sciences), 2014, 57 (05) : 195 - 207
  • [44] Learning Discriminative Feature Representation for Open Set Action Recognition
    Zhang, Hongjie
    Liu, Yi
    Wang, Yali
    Wang, Limin
    Qiao, Yu
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 7696 - 7705
  • [45] Cross-modal learning with multi-modal model for video action recognition based on adaptive weight training
    Zhou, Qingguo
    Hou, Yufeng
    Zhou, Rui
    Li, Yan
    Wang, Jinqiang
    Wu, Zhen
    Li, Hung-Wei
    Weng, Tien-Hsiung
    CONNECTION SCIENCE, 2024, 36 (01)
  • [46] Rethinking Fusion Baselines for Multi-modal Human Action Recognition
    Jiang, Hongda
    Li, Yanghao
    Song, Sijie
    Liu, Jiaying
    ADVANCES IN MULTIMEDIA INFORMATION PROCESSING, PT III, 2018, 11166 : 178 - 187
  • [47] MULTI-MODAL FUSION WITH OBSERVATION POINTS FOR SKELETON ACTION RECOGNITION
    Singh, Iqbal
    Zhu, Xiaodan
    Greenspan, Michael
    2020 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2020, : 1781 - 1785
  • [48] Multi-Modal Three-Stream Network for Action Recognition
    Khalid, Muhammad Usman
    Yu, Jie
    2018 24TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2018, : 3210 - 3215
  • [49] Vision-Based Multi-Modal Framework for Action Recognition
    Romaissa, Beddiar Djamila
    Mourad, Oussalah
    Brahim, Nini
    2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, : 5859 - 5866
  • [50] Discriminative multi-modal deep generative models
    Du, Fang
    Zhang, Jiangshe
    Hu, Junying
    Fei, Rongrong
    KNOWLEDGE-BASED SYSTEMS, 2019, 173 : 74 - 82