Learning-Based Multimodal Information Fusion and Behavior Recognition of Vascular Interventionists' Operating Skills

被引:1
|
作者
Wang, Shuang [1 ]
Liu, Zheng [1 ]
Yang, Wentuo [1 ]
Cao, Yongfeng [1 ]
Zhao, Liang [2 ]
Xie, Le [1 ,3 ]
机构
[1] Shanghai Jiao Tong Univ, Sch Mat Sci & Engn, Shanghai 200240, Peoples R China
[2] Shanghai Jiao Tong Univ, Chest Hosp, Shanghai 200030, Peoples R China
[3] Shanghai Jiao Tong Univ, Inst Med Robot, Shanghai 200240, Peoples R China
基金
中国国家自然科学基金;
关键词
Manual percutaneous coronary intervention; multimodal information fusion; behavior recognition; machine learning; deep learning; PERCUTANEOUS CORONARY INTERVENTION; NEURAL-NETWORKS; SAFETY;
D O I
10.1109/JBHI.2023.3289548
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The operating skills of vascular interventionists have an important impact on the effect of surgery. However, current research on behavior recognition and skills learning of interventionists' operating skills is limited. In this study, an innovative deep learning-based multimodal information fusion architecture is proposed for recognizing and analyzing eight common operating behaviors of interventionists. An experimental platform integrating four modal sensors is used to collect multimodal data from interventionists. The ANOVA and Manner-Whitney tests is used for relevance analysis of the data. The analysis results demonstrate that there is almost no significant difference (p <0.001) between the actions related to the unimodal data, which cannot be used for accurate behavior recognition. Therefore, a study of the fusion architecture based on the existing machine learning classifier and the proposed deep learning fusion architecture is carried out. The research findings indicate that the proposed deep learning-based fusion architecture achieves an impressive overall accuracy of 98.5%, surpassing both the machine learning classifier (93.51%) and the unimodal data (90.05%). The deep learning-based multimodal information fusion architecture proves the feasibility of behavior recognition and skills learning of interventionist's operating skills. Furthermore, the application of deep learning-based multimodal fusion technology of surgeon's operating skills will help to improve the autonomy and intelligence of surgical robotic systems.
引用
收藏
页码:4536 / 4547
页数:12
相关论文
共 50 条
  • [1] The Recognition of Teacher Behavior Based on Multimodal Information Fusion
    Wu, Dongli
    Chen, Jia
    Deng, Wei
    Wei, Yantao
    Luo, Heng
    Wei, Yangyu
    [J]. MATHEMATICAL PROBLEMS IN ENGINEERING, 2020, 2020 (2020)
  • [2] Contrastive Learning-Based Multimodal Fusion Model for Automatic Modulation Recognition
    Liu, Fugang
    Pan, Jingyi
    Zhou, Ruolin
    [J]. IEEE COMMUNICATIONS LETTERS, 2024, 28 (01) : 78 - 82
  • [3] A review of deep learning-based information fusion techniques for multimodal medical image classification
    Li, Yihao
    El Habib Daho, Mostafa
    Conze, Pierre-Henri
    Zeghlache, Rachid
    Le Boité, Hugo
    Tadayoni, Ramin
    Cochener, Béatrice
    Lamard, Mathieu
    Quellec, Gwenolé
    [J]. Computers in Biology and Medicine, 2024, 177
  • [4] Deep learning-based late fusion of multimodal information for emotion classification of music video
    Pandeya, Yagya Raj
    Lee, Joonwhoan
    [J]. MULTIMEDIA TOOLS AND APPLICATIONS, 2021, 80 (02) : 2887 - 2905
  • [5] Deep learning-based late fusion of multimodal information for emotion classification of music video
    Yagya Raj Pandeya
    Joonwhoan Lee
    [J]. Multimedia Tools and Applications, 2021, 80 : 2887 - 2905
  • [6] CSE-ARS: Deep Learning-Based Late Fusion of Multimodal Information for Chat-Based Social Engineering Attack Recognition
    Tsinganos, Nikolaos
    Fouliras, Panagiotis
    Mavridis, Ioannis
    Gritzalis, Dimitris
    [J]. IEEE ACCESS, 2024, 12 : 16072 - 16088
  • [7] Multimodal information fusion based human movement recognition
    Shu, Yao
    Zhang, Heng
    [J]. MULTIMEDIA TOOLS AND APPLICATIONS, 2020, 79 (7-8) : 5043 - 5052
  • [8] Multimodal information fusion based human movement recognition
    Yao Shu
    Heng Zhang
    [J]. Multimedia Tools and Applications, 2020, 79 : 5043 - 5052
  • [9] Early, intermediate and late fusion strategies for robust deep learning-based multimodal action recognition
    Said Yacine Boulahia
    Abdenour Amamra
    Mohamed Ridha Madi
    Said Daikh
    [J]. Machine Vision and Applications, 2021, 32
  • [10] Early, intermediate and late fusion strategies for robust deep learning-based multimodal action recognition
    Boulahia, Said Yacine
    Amamra, Abdenour
    Madi, Mohamed Ridha
    Daikh, Said
    [J]. MACHINE VISION AND APPLICATIONS, 2021, 32 (06)