Multi-modal fusion in ergonomic health: bridging visual and pressure for sitting posture detection

被引:1
|
作者
Quan, Qinxiao [1 ]
Gao, Yang [2 ]
Bai, Yang [1 ]
Jin, Zhanpeng [1 ]
机构
[1] South China Univ Technol, Sch Future Technol, Guangzhou, Peoples R China
[2] East China Normal Univ, Sch Comp Sci, Shanghai, Peoples R China
关键词
Pressure sensing; Computer vision; Sitting posture recognition; Feature fusion; Multi-label classification; RECOGNITION;
D O I
10.1007/s42486-024-00164-x
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
As the contradiction between the pursuit of health and the increasing duration of sedentary office work intensifies, there has been a growing focus on maintaining correct sitting posture while working in recent years. Scientific studies have shown that sitting posture correction plays a positive role in alleviating physical pain. With the rapid development of artificial intelligence technology, a significant amount of research has shifted towards implementing sitting posture detection and recognition systems using machine learning approaches. In this paper, we introduce an innovative sitting posture recognition system that integrates visual and pressure modalities. The system employs a differentiated pre-training strategy for training the bimodal models and features a feature fusion module designed based on feed-forward networks. Our system utilizes commonly available built-in cameras in laptops for collecting visual data and thin-film pressure sensor mats for pressure data in office scenarios. It achieved an F1-Macro score of 95.43% on a dataset with complex composite actions, marking an improvement of 7.13% and 10.79% over systems that rely solely on pressure or visual modalities, respectively, and a 7.07% improvement over systems using a uniform pre-training strategy.
引用
收藏
页码:380 / 393
页数:14
相关论文
共 50 条
  • [31] Multi-Modal Fake News Detection via Bridging the Gap between Modals
    Liu, Peng
    Qian, Wenhua
    Xu, Dan
    Ren, Bingling
    Cao, Jinde
    ENTROPY, 2023, 25 (04)
  • [32] Multi-Modal Posture Recognition System for Healthcare Applications
    Sreeni, Siddarth
    Hari, S. R.
    Harikrishnan, R.
    Sreejith, V
    PROCEEDINGS OF TENCON 2018 - 2018 IEEE REGION 10 CONFERENCE, 2018, : 0373 - 0376
  • [33] Bridging the View Disparity Between Radar and Camera Features for Multi-Modal Fusion 3D Object Detection
    Zhou, Taohua
    Chen, Junjie
    Shi, Yining
    Jiang, Kun
    Yang, Mengmeng
    Yang, Diange
    IEEE TRANSACTIONS ON INTELLIGENT VEHICLES, 2023, 8 (02): : 1523 - 1535
  • [34] Visual Prompt Multi-Modal Tracking
    Zhu, Jiawen
    Lai, Simiao
    Chen, Xin
    Wang, Dong
    Lu, Huchuan
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 9516 - 9526
  • [35] Multi-modal affine fusion network for social media rumor detection
    Fu, Boyang
    Sui, Jie
    PEERJ COMPUTER SCIENCE, 2022, 8
  • [36] VISUAL AS MULTI-MODAL ARGUMENTATION IN LAW
    Novak, Marko
    BRATISLAVA LAW REVIEW, 2021, 5 (01): : 91 - 110
  • [37] Citrus Huanglongbing Detection Based on Multi-Modal Feature Fusion Learning
    Yang, Dongzi
    Wang, Fengcheng
    Hu, Yuqi
    Lan, Yubin
    Deng, Xiaoling
    FRONTIERS IN PLANT SCIENCE, 2021, 12
  • [38] Multi-modal deep fusion based fake news detection method
    Jing Q.
    Fan X.
    Wang B.
    Bi J.
    Tan H.
    High Technology Letters, 2022, 32 (04) : 392 - 403
  • [39] MULTI-MODAL FEATURE FUSION NETWORK FOR GHOST IMAGING OBJECT DETECTION
    Hu, Nan
    Ma, Huimin
    Le, Chao
    Shao, Xuehui
    2018 25TH IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2018, : 351 - 355
  • [40] Vehicle Detection of Multi-Modal Attention Fusion Under Different Illumination
    Wang, Jiaqi
    Zhang, Qi
    Huang, Wei
    Computer Engineering and Applications, 2024, 60 (16) : 116 - 123