Learning discriminative motion feature for enhancing multi-modal action recognition

被引:0
|
作者
Yang, Jianyu [1 ]
Huang, Yao [1 ]
Shao, Zhanpeng [2 ]
Liu, Chunping [3 ]
机构
[1] School of Rail Transportation, Soochow University, Suzhou,215000, China
[2] School of Computer Science and Technology, Zhejiang University of Technology, Hangzhou,310023, China
[3] School of Computer Science and Technology, Soochow University, Suzhou,215000, China
关键词
D O I
暂无
中图分类号
学科分类号
摘要
Video action recognition is an important topic in computer vision tasks. Most of the existing methods use CNN-based models, and multiple modalities of image features are captured from the videos, such as static frames, dynamic images, and optical flow features. However, these mainstream features contain much static information including object and background information, where the motion information of the action itself is not distinguished and strengthened. In this work, a new kind of motion feature is proposed without static information for video action recognition. We propose a quantization of motion network based on the bag-of-feature method to learn significant and discriminative motion features. In the learned feature map, the object and background information is filtered out, even if the background is moving in the video. Therefore, the motion feature is complementary to the static image feature and the static information in the dynamic image and optical flow. A multi-stream classifier is built with the proposed motion feature and other features, and the performance of action recognition is enhanced comparing to other state-of-the-art methods. © 2021 Elsevier Inc.
引用
收藏
相关论文
共 50 条
  • [11] Multi-Modal Multi-Action Video Recognition
    Shi, Zhensheng
    Liang, Ju
    Li, Qianqian
    Zheng, Haiyong
    Gu, Zhaorui
    Dong, Junyu
    Zheng, Bing
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 13658 - 13667
  • [12] Modality Mixer for Multi-modal Action Recognition
    Lee, Sumin
    Woo, Sangmin
    Park, Yeonju
    Nugroho, Muhammad Adi
    Kim, Changick
    2023 IEEE/CVF WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV), 2023, : 3297 - 3306
  • [13] Electromagnetic signal feature fusion and recognition based on multi-modal deep learning
    Hou C.
    Zhang X.
    Chen X.
    International Journal of Performability Engineering, 2020, 16 (06): : 941 - 949
  • [14] Gesture recognition based on multi-modal feature weight
    Duan, Haojie
    Sun, Ying
    Cheng, Wentao
    Jiang, Du
    Yun, Juntong
    Liu, Ying
    Liu, Yibo
    Zhou, Dalin
    Concurrency and Computation: Practice and Experience, 2021, 33 (05)
  • [15] Fusional Recognition for Depressive Tendency With Multi-Modal Feature
    Wang, Hong
    Zhou, Ying
    Yu, Fengping
    Zhao, Lili
    Wang, Caiyu
    Ren, Yanju
    IEEE ACCESS, 2019, 7 : 38702 - 38713
  • [16] Gesture recognition based on multi-modal feature weight
    Duan, Haojie
    Sun, Ying
    Cheng, Wentao
    Jiang, Du
    Yun, Juntong
    Liu, Ying
    Liu, Yibo
    Zhou, Dalin
    CONCURRENCY AND COMPUTATION-PRACTICE & EXPERIENCE, 2021, 33 (05):
  • [17] Multi-Modal Deep Learning-Based Violin Bowing Action Recognition
    Liu, Bao-Yun
    Jen, Yi-Hsin
    Sun, Shih-Wei
    Su, Li
    Chang, Pao-Chi
    2020 IEEE INTERNATIONAL CONFERENCE ON CONSUMER ELECTRONICS - TAIWAN (ICCE-TAIWAN), 2020,
  • [18] DISCRIMINATIVE HALLUCINATION FOR MULTI-MODAL FEW-SHOT LEARNING
    Pahde, Frederik
    Nabi, Moin
    Klein, Tassilo
    Jaehnichen, Patrick
    2018 25TH IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2018, : 156 - 160
  • [19] Interactive Learning of a Dual Convolution Neural Network for Multi-Modal Action Recognition
    Li, Qingxia
    Gao, Dali
    Zhang, Qieshi
    Wei, Wenhong
    Ren, Ziliang
    MATHEMATICS, 2022, 10 (21)
  • [20] Multi-modal broad learning for material recognition
    Wang, Zhaoxin
    Liu, Huaping
    Xu, Xinying
    Sun, Fuchun
    COGNITIVE COMPUTATION AND SYSTEMS, 2021, 3 (02) : 123 - 130