Learning-Based Multimodal Information Fusion and Behavior Recognition of Vascular Interventionists' Operating Skills

被引:1
|
作者
Wang, Shuang [1 ]
Liu, Zheng [1 ]
Yang, Wentuo [1 ]
Cao, Yongfeng [1 ]
Zhao, Liang [2 ]
Xie, Le [1 ,3 ]
机构
[1] Shanghai Jiao Tong Univ, Sch Mat Sci & Engn, Shanghai 200240, Peoples R China
[2] Shanghai Jiao Tong Univ, Chest Hosp, Shanghai 200030, Peoples R China
[3] Shanghai Jiao Tong Univ, Inst Med Robot, Shanghai 200240, Peoples R China
基金
中国国家自然科学基金;
关键词
Manual percutaneous coronary intervention; multimodal information fusion; behavior recognition; machine learning; deep learning; PERCUTANEOUS CORONARY INTERVENTION; NEURAL-NETWORKS; SAFETY;
D O I
10.1109/JBHI.2023.3289548
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The operating skills of vascular interventionists have an important impact on the effect of surgery. However, current research on behavior recognition and skills learning of interventionists' operating skills is limited. In this study, an innovative deep learning-based multimodal information fusion architecture is proposed for recognizing and analyzing eight common operating behaviors of interventionists. An experimental platform integrating four modal sensors is used to collect multimodal data from interventionists. The ANOVA and Manner-Whitney tests is used for relevance analysis of the data. The analysis results demonstrate that there is almost no significant difference (p <0.001) between the actions related to the unimodal data, which cannot be used for accurate behavior recognition. Therefore, a study of the fusion architecture based on the existing machine learning classifier and the proposed deep learning fusion architecture is carried out. The research findings indicate that the proposed deep learning-based fusion architecture achieves an impressive overall accuracy of 98.5%, surpassing both the machine learning classifier (93.51%) and the unimodal data (90.05%). The deep learning-based multimodal information fusion architecture proves the feasibility of behavior recognition and skills learning of interventionist's operating skills. Furthermore, the application of deep learning-based multimodal fusion technology of surgeon's operating skills will help to improve the autonomy and intelligence of surgical robotic systems.
引用
收藏
页码:4536 / 4547
页数:12
相关论文
共 50 条
  • [41] MULTIMODAL INFORMATION FUSION OF AUDIO EMOTION RECOGNITION BASED ON KERNEL ENTROPY COMPONENT ANALYSIS
    Xie, Zhibing
    Guan, Ling
    [J]. INTERNATIONAL JOURNAL OF SEMANTIC COMPUTING, 2013, 7 (01) : 25 - 42
  • [42] Multimodal Information Fusion of Audio Emotion Recognition Based on Kernel Entropy Component Analysis
    Xie, Zhibing
    Guan, Ling
    [J]. 2012 IEEE INTERNATIONAL SYMPOSIUM ON MULTIMEDIA (ISM), 2012, : 1 - 8
  • [43] Fusion of Video and Inertial Sensing for Deep Learning-Based Human Action Recognition
    Wei, Haoran
    Jafari, Roozbeh
    Kehtarnavaz, Nasser
    [J]. SENSORS, 2019, 19 (17)
  • [44] Multimodal Emotion Recognition Based on Feature Fusion
    Xu, Yurui
    Wu, Xiao
    Su, Hang
    Liu, Xiaorui
    [J]. 2022 INTERNATIONAL CONFERENCE ON ADVANCED ROBOTICS AND MECHATRONICS (ICARM 2022), 2022, : 7 - 11
  • [45] Quality Fusion Based Multimodal Eye Recognition
    Zhou, Zhi
    Du, Eliza Yingzi
    Belcher, Craig
    Thomas, N. Luke
    Delp, Edward J.
    [J]. PROCEEDINGS 2012 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2012, : 1297 - 1302
  • [46] A deep learning-based approach for feeding behavior recognition of weanling pigs
    Kim, MinJu
    Choi, YoHan
    Lee, Jeong-nam
    Sa, SooJin
    Cho, Hyun-chong
    [J]. JOURNAL OF ANIMAL SCIENCE AND TECHNOLOGY, 2021, 63 (06) : 1453 - 1463
  • [47] A Deep Learning-Based Posture Estimation Approach for Poultry Behavior Recognition
    Xie, Bo-Xuan
    Chang, Chung-Liang
    [J]. 2022 6TH INTERNATIONAL CONFERENCE ON IMAGING, SIGNAL PROCESSING AND COMMUNICATIONS, ICISPC, 2022, : 32 - 36
  • [48] Few-shot learning-based human behavior recognition model
    Mahalakshmi, V.
    Sandhu, Mukta
    Shabaz, Mohammad
    Keshta, Ismail
    Prasad, K. D. V.
    Kuzieva, Nargiza
    Byeon, Haewon
    Soni, Mukesh
    [J]. COMPUTERS IN HUMAN BEHAVIOR, 2024, 151
  • [49] Multimodal Intelligence: Representation Learning, Information Fusion, and Applications
    Zhang, Chao
    Yang, Zichao
    He, Xiaodong
    Deng, Li
    [J]. IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, 2020, 14 (03) : 478 - 493
  • [50] Research on deep learning-based action recognition and quantitative assessment method for sports skills
    Wang, Tao
    [J]. Applied Mathematics and Nonlinear Sciences, 2024, 9 (01)