Learning discriminative motion feature for enhancing multi-modal action recognition

被引:0
|
作者
Yang, Jianyu [1 ]
Huang, Yao [1 ]
Shao, Zhanpeng [2 ]
Liu, Chunping [3 ]
机构
[1] School of Rail Transportation, Soochow University, Suzhou,215000, China
[2] School of Computer Science and Technology, Zhejiang University of Technology, Hangzhou,310023, China
[3] School of Computer Science and Technology, Soochow University, Suzhou,215000, China
关键词
D O I
暂无
中图分类号
学科分类号
摘要
Video action recognition is an important topic in computer vision tasks. Most of the existing methods use CNN-based models, and multiple modalities of image features are captured from the videos, such as static frames, dynamic images, and optical flow features. However, these mainstream features contain much static information including object and background information, where the motion information of the action itself is not distinguished and strengthened. In this work, a new kind of motion feature is proposed without static information for video action recognition. We propose a quantization of motion network based on the bag-of-feature method to learn significant and discriminative motion features. In the learned feature map, the object and background information is filtered out, even if the background is moving in the video. Therefore, the motion feature is complementary to the static image feature and the static information in the dynamic image and optical flow. A multi-stream classifier is built with the proposed motion feature and other features, and the performance of action recognition is enhanced comparing to other state-of-the-art methods. © 2021 Elsevier Inc.
引用
收藏
相关论文
共 50 条
  • [31] Geological Body Recognition Based on Multi-Modal Feature Fusion
    Fu S.
    Li C.
    Zhang H.
    Liu C.
    Li F.
    Diqiu Kexue - Zhongguo Dizhi Daxue Xuebao/Earth Science - Journal of China University of Geosciences, 2023, 48 (10): : 3743 - 3752
  • [32] Learning Multi-modal Attentional Consensus in Action Recognition for Elderly-Care Robots
    Kim, Hyungmin
    Kim, Dohyung
    Kim, Jaehong
    2021 18TH INTERNATIONAL CONFERENCE ON UBIQUITOUS ROBOTS (UR), 2021, : 308 - 313
  • [33] Anticipative Feature Fusion Transformer for Multi-Modal Action Anticipation
    Zhong, Zeyun
    Schneider, David
    Voit, Michael
    Stiefelhagen, Rainer
    Beyerer, Juergen
    2023 IEEE/CVF WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV), 2023, : 6057 - 6066
  • [34] Multi-modal Action Segmentation in the Kitchen with a Feature Fusion Approach
    Kogure, Shunsuke
    Aoki, Yoshimitsu
    FIFTEENTH INTERNATIONAL CONFERENCE ON QUALITY CONTROL BY ARTIFICIAL VISION, 2021, 11794
  • [35] Multi-Frequency RF Sensor Data Adaptation for Motion Recognition with Multi-Modal Deep Learning
    Rahman, M. Mahbubur
    Gurbuz, Sevgi Z.
    2021 IEEE RADAR CONFERENCE (RADARCONF21): RADAR ON THE MOVE, 2021,
  • [36] Multi-View and Multi-Modal Action Recognition with Learned Fusion
    Ardianto, Sandy
    Hang, Hsueh-Ming
    2018 ASIA-PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE (APSIPA ASC), 2018, : 1601 - 1604
  • [37] A Multi-Modal Deep Learning Approach for Emotion Recognition
    Shahzad, H. M.
    Bhatti, Sohail Masood
    Jaffar, Arfan
    Rashid, Muhammad
    INTELLIGENT AUTOMATION AND SOFT COMPUTING, 2023, 36 (02): : 1561 - 1570
  • [38] Multi-modal human motion recognition based on behaviour tree
    Yang, Qin
    Zhou, Zhenhua
    INTERNATIONAL JOURNAL OF BIOMETRICS, 2024, 16 (3-4) : 381 - 398
  • [39] MMSS: Multi-modal Sharable and Specific Feature Learning for RGB-D Object Recognition
    Wang, Anran
    Cai, Jianfei
    Lu, Jiwen
    Cham, Tat-Jen
    2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, : 1125 - 1133
  • [40] Learning a Discriminative Feature Descriptor with Sparse Coding for Action Recognition
    Li, Lingqiao
    Zhang, Tao
    Pan, Xipeng
    Yang, Huihua
    Liu, Zhenbing
    2018 17TH INTERNATIONAL SYMPOSIUM ON DISTRIBUTED COMPUTING AND APPLICATIONS FOR BUSINESS ENGINEERING AND SCIENCE (DCABES), 2018, : 80 - 83