Contrastive self-supervised representation learning without negative samples for multimodal human action recognition

被引:1
|
作者
Yang, Huaigang [1 ]
Ren, Ziliang [1 ,2 ]
Yuan, Huaqiang [1 ]
Xu, Zhenyu [2 ]
Zhou, Jun [1 ]
机构
[1] Dongguan Univ Technol, Sch Comp Sci & Technol, Dongguan, Peoples R China
[2] Chinese Acad Sci, Shenzhen Inst Adv Technol, CAS Key Lab Human Machine Intelligence Synergy Sys, Shenzhen, Peoples R China
基金
中国国家自然科学基金;
关键词
human action recognition; multimodal representation; feature encoder; contrastive self-supervised learning; Transformer;
D O I
10.3389/fnins.2023.1225312
中图分类号
Q189 [神经科学];
学科分类号
071006 ;
摘要
Action recognition is an important component of human-computer interaction, and multimodal feature representation and learning methods can be used to improve recognition performance due to the interrelation and complementarity between different modalities. However, due to the lack of large-scale labeled samples, the performance of existing ConvNets-based methods are severely constrained. In this paper, a novel and effective multi-modal feature representation and contrastive self-supervised learning framework is proposed to improve the action recognition performance of models and the generalization ability of application scenarios. The proposed recognition framework employs weight sharing between two branches and does not require negative samples, which could effectively learn useful feature representations by using multimodal unlabeled data, e.g., skeleton sequence and inertial measurement unit signal (IMU). The extensive experiments are conducted on two benchmarks: UTD-MHAD and MMAct, and the results show that our proposed recognition framework outperforms both unimodal and multimodal baselines in action retrieval, semi-supervised learning, and zero-shot learning scenarios.
引用
收藏
页数:14
相关论文
共 50 条
  • [41] Supervised and Self-Supervised Learning for Assembly Line Action Recognition
    Indris, Christopher
    Ibrahim, Fady
    Ibrahem, Hatem
    Bramesfeld, Gotz
    Huo, Jie
    Ahmad, Hafiz Mughees
    Hayat, Syed Khizer
    Wang, Guanghui
    JOURNAL OF IMAGING, 2025, 11 (01)
  • [42] Self-supervised representation learning for surgical activity recognition
    Paysan, Daniel
    Haug, Luis
    Bajka, Michael
    Oelhafen, Markus
    Buhmann, Joachim M.
    INTERNATIONAL JOURNAL OF COMPUTER ASSISTED RADIOLOGY AND SURGERY, 2021, 16 (11) : 2037 - 2044
  • [43] Self-Supervised ECG Representation Learning for Emotion Recognition
    Sarkar, Pritam
    Etemad, Ali
    IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2022, 13 (03) : 1541 - 1554
  • [44] Self-supervised representation learning for surgical activity recognition
    Daniel Paysan
    Luis Haug
    Michael Bajka
    Markus Oelhafen
    Joachim M. Buhmann
    International Journal of Computer Assisted Radiology and Surgery, 2021, 16 : 2037 - 2044
  • [45] Contrastive Self-Supervised Learning for Sensor-Based Human Activity Recognition: A Review
    Chen, Hui
    Gouin-Vallerand, Charles
    Bouchard, Kevin
    Gaboury, Sebastien
    Couture, Melanie
    Bier, Nathalie
    Giroux, Sylvain
    IEEE ACCESS, 2024, 12 : 152511 - 152531
  • [46] SG-CLR: Semantic representation-guided contrastive learning for self-supervised skeleton-based action recognition
    Liu, Ruyi
    Liu, Yi
    Wu, Mengyao
    Xin, Wentian
    Miao, Qiguang
    Liu, Xiangzeng
    Lie, Long
    PATTERN RECOGNITION, 2025, 162
  • [47] Boost Supervised Pretraining for Visual Transfer Learning: Implications of Self-Supervised Contrastive Representation Learning
    Sun, Jinghan
    Wei, Dong
    Ma, Kai
    Wang, Liansheng
    Zheng, Yefeng
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / THE TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 2307 - 2315
  • [48] Cross-stream contrastive learning for self-supervised skeleton-based action recognition
    Li, Ding
    Tang, Yongqiang
    Zhang, Zhizhong
    Zhang, Wensheng
    IMAGE AND VISION COMPUTING, 2023, 135
  • [49] Self-Supervised Fair Representation Learning without Demographics
    Chai, Junyi
    Wang, Xiaoqian
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [50] Mixing up contrastive learning: Self-supervised representation learning for time series
    Wickstrom, Kristoffer
    Kampffmeyer, Michael
    Mikalsen, Karl Oyvind
    Jenssen, Robert
    PATTERN RECOGNITION LETTERS, 2022, 155 : 54 - 61