Multimodal Transformer for Nursing Activity Recognition

被引:12
|
作者
Ijaz, Momal [1 ]
Diaz, Renato [1 ]
Chen, Chen [1 ,2 ]
机构
[1] Univ Cent Florida, Dept Comp Sci, Orlando, FL 32816 USA
[2] Univ Cent Florida, Ctr Res Comp Vis, Orlando, FL 32816 USA
关键词
D O I
10.1109/CVPRW56347.2022.00224
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
In an aging population, elderly patient safety is a primary concern at hospitals and nursing homes, which demands for increased nurse care. By performing nurse activity recognition, we can not only make sure that all patients get an equal desired care, but it can also free nurses from manual documentation of activities they perform, leading to a fair and safe place of care for the elderly. In this work, we present a multimodal transformer-based network, which extracts features from skeletal joints and acceleration data, and fuses them to perform nurse activity recognition. Our method achieves state-of-the-art performance of 81.8% accuracy on the benchmark dataset available for nurse activity recognition from the Nurse Care Activity Recognition Challenge. We perform ablation studies to show that our fusion model is better than single modality transformer variants (using only acceleration or skeleton joints data). Our solution also outperforms state-of-the-art ST-GCN, GRU and other classical hand-crafted-feature-based classifier solutions by a margin of 1.6%, on the NCRC dataset. Code is available at https://github.com/Momilijaz96/MMT_for_NCRC.
引用
收藏
页码:2064 / 2073
页数:10
相关论文
共 50 条
  • [1] Multimodal Transformer for Nursing Activity Recognition
    Ijaz, Momal
    Diaz, Renato
    Chen, Chen
    [J]. arXiv, 2022,
  • [2] Self-supervised multimodal fusion transformer for passive activity recognition
    Koupai, Armand K.
    Bocus, Mohammud J.
    Santos-Rodriguez, Raul
    Piechocki, Robert J.
    McConville, Ryan
    [J]. IET WIRELESS SENSOR SYSTEMS, 2022, 12 (5-6) : 149 - 160
  • [3] Multi-agent Transformer Networks for Multimodal Human Activity Recognition
    Li, Jingcheng
    Yao, Lina
    Li, Binghao
    Wang, Xianzhi
    Sammut, Claude
    [J]. PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2022, 2022, : 1135 - 1145
  • [4] Husformer: A Multimodal Transformer for Multimodal Human State Recognition
    Wang, Ruiqi
    Jo, Wonse
    Zhao, Dezhong
    Wang, Weizheng
    Gupte, Arjun
    Yang, Baijian
    Chen, Guohua
    Min, Byung-Cheol
    [J]. IEEE TRANSACTIONS ON COGNITIVE AND DEVELOPMENTAL SYSTEMS, 2024, 16 (04) : 1374 - 1390
  • [5] MULTIMODAL TRANSFORMER FUSION FOR CONTINUOUS EMOTION RECOGNITION
    Huang, Jian
    Tao, Jianhua
    Liu, Bin
    Lian, Zheng
    Niu, Mingyue
    [J]. 2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 3507 - 3511
  • [6] Pedestrian Attribute Recognition Based on Multimodal Transformer
    Liu, Dan
    Song, Wei
    Zhao, Xiaobing
    [J]. PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2023, PT I, 2024, 14425 : 422 - 433
  • [7] MGRFormer: A Multimodal Transformer Approach for Surgical Gesture Recognition
    Feghoul, Kevin
    Maia, Deise Santana
    El Amrani, Mehdi
    Daoudi, Mohamed
    Amad, Ali
    [J]. 2024 IEEE 18TH INTERNATIONAL CONFERENCE ON AUTOMATIC FACE AND GESTURE RECOGNITION, FG 2024, 2024,
  • [8] Multimodal transformer augmented fusion for speech emotion recognition
    Wang, Yuanyuan
    Gu, Yu
    Yin, Yifei
    Han, Yingping
    Zhang, He
    Wang, Shuang
    Li, Chenyu
    Quan, Dou
    [J]. FRONTIERS IN NEUROROBOTICS, 2023, 17
  • [9] TRANSFORMER BASED MULTIMODAL SCENE RECOGNITION IN SOCCER VIDEOS
    Gan, Yaozong
    Togo, Ren
    Ogawa, Takahiro
    Haseyama, Miki
    [J]. 2022 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO WORKSHOPS (IEEE ICMEW 2022), 2022,
  • [10] Building Robust Multimodal Sentiment Recognition via a Simple yet Effective Multimodal Transformer
    Zong, Daoming
    Ding, Chaoyue
    Li, Baoxiang
    Zhou, Dinghao
    Li, Jiakui
    Zheng, Ken
    Zhou, Qunyan
    [J]. PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 9596 - 9600