Explainable Multi-Modal Deep Learning With Cross-Modal Attention for Diagnosis of Dyssynergic Defecation Using Abdominal X-Ray Images and Symptom Questionnaire

被引:0
|
作者
Sangnark, Sirapob [1 ]
Rattanachaisit, Pakkapon [2 ,3 ]
Patcharatrakul, Tanisa [3 ,4 ]
Vateekul, Peerapon [1 ]
机构
[1] Chulalongkorn Univ, Fac Engn, Dept Comp Engn, Bangkok 10330, Thailand
[2] Chulalongkorn Univ, Fac Med, Dept Physiol, Bangkok 10330, Thailand
[3] Chulalongkorn Univ, Fac Med, Ctr Excellence Neurogastroenterol & Motil, Bangkok 10330, Thailand
[4] King Chulalongkorn Mem Hosp, Dept Med, Div Gastroenterol, Thai Red Cross Soc, Bangkok 10330, Thailand
来源
IEEE ACCESS | 2024年 / 12卷
关键词
Deep learning; X-ray imaging; Data models; Biomedical imaging; Diseases; Medical diagnostic imaging; Task analysis; Dyssynergic defecation; multi-modal; deep learning; attention mechanism; explainable AI; CLASSIFICATION; NETWORK;
D O I
10.1109/ACCESS.2024.3409077
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Dyssynergic defecation (DD) is a type of functional constipation that requires a specialized test for diagnosis. However, these tests are only accessible in tertiary care because they require devices that are not available elsewhere. In this work, we present explainable multi-modal deep learning models that can pre-screen patients with DD, using affordable data accessible in small hospitals i.e. abdominal X-ray images and symptom questionnaires; the output classifies whether DD is present or not. To enhance the model's performance, we apply cross-modal attention to help the model find meaningful interactions between the two modalities. A convolution block attention module (CBAM) is added to obtain more important semantic and spatial features from the images. Masking augmentation is implemented to ignore irrelevant backgrounds in images. Both explainable AI techniques like gradient-weighted class activation mapping (Grad-CAM) and deep shapley additive explanations (DeepSHAP) are also used to explain the important parts of images and the symptom data for each patient. In our experiments, all models are run on 3 patient-based bootstraps. Our model is compared with single-modal models and human experts. Results demonstrate that our multi-modal model outperforms the single-modal model and achieves the highest in terms of sensitivity, specificity, F1, and accuracy (87.37%, 77.01%, 82.17%, and 82.27%), respectively. In addition, our model outperforms human experts, which shows its ability to assist human experts in diagnosing DD. This model is a novel clinical tool that combines symptom and image data for a more accurate diagnosis of DD.
引用
收藏
页码:78132 / 78147
页数:16
相关论文
共 50 条
  • [1] Multi-modal deep learning for predicting dyssynergic defecation (DD) using abdominal X-rays and symptom questionnaires
    Rattanachaisit, Pakkapon
    Sangnark, Sirapob
    Patcharatrakul, Tanisa
    Gonlachanvit, Sutep
    Vateekul, Peerapon
    NEUROGASTROENTEROLOGY AND MOTILITY, 2024, 36
  • [2] MULTI-MODAL DEEP LEARNING MODEL FOR DIAGNOSIS OF DYSSYNERGIC DEFECATION (DD)
    Rattanachaisit, Pakkapon
    Sangnark, Sirapob
    Patcharatrakul, Tanisa
    Gonlachanvit, Sutep
    Vateekul, Peerapon
    GASTROENTEROLOGY, 2024, 166 (05) : S1391 - S1392
  • [3] A cross-modal deep metric learning model for disease diagnosis based on chest x-ray images
    Yufei Jin
    Huijuan Lu
    Zhao Li
    Yanbin Wang
    Multimedia Tools and Applications, 2023, 82 : 33421 - 33442
  • [4] A cross-modal deep metric learning model for disease diagnosis based on chest x-ray images
    Jin, Yufei
    Lu, Huijuan
    Li, Zhao
    Wang, Yanbin
    MULTIMEDIA TOOLS AND APPLICATIONS, 2023, 82 (21) : 33421 - 33442
  • [5] Cross-modal attention network for retinal disease classification based on multi-modal images
    Liu, Zirong
    Hu, Yan
    Qiu, Zhongxi
    Niu, Yanyan
    Zhou, Dan
    Li, Xiaoling
    Shen, Junyong
    Jiang, Hongyang
    Li, Heng
    Liu, Jiang
    BIOMEDICAL OPTICS EXPRESS, 2024, 15 (06): : 3699 - 3714
  • [6] Learning Cross-Modal Deep Representations for Multi-Modal MR Image Segmentation
    Li, Cheng
    Sun, Hui
    Liu, Zaiyi
    Wang, Meiyun
    Zheng, Hairong
    Wang, Shanshan
    MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION - MICCAI 2019, PT II, 2019, 11765 : 57 - 65
  • [7] AI-Assisted Diagnosis of Dyssynergic Defecation Using Deep Learning Approach on Abdominal Radiography and Symptom Questionnaire
    Poovongsaroj, Sornsiri
    Rattanachaisit, Pakkapon
    Patcharatrakul, Tanisa
    Gonlachanvit, Sutep
    Vateekul, Peerapon
    2022 19TH INTERNATIONAL JOINT CONFERENCE ON COMPUTER SCIENCE AND SOFTWARE ENGINEERING (JCSSE 2022), 2022,
  • [8] A Framework for Enabling Unpaired Multi-Modal Learning for Deep Cross-Modal Hashing Retrieval
    Williams-Lekuona, Mikel
    Cosma, Georgina
    Phillips, Iain
    JOURNAL OF IMAGING, 2022, 8 (12)
  • [9] Multi-modal fusion of deep transfer learning based COVID-19 diagnosis and classification using chest x-ray images
    Reddy, A. Siva Krishna
    Rao, K. N. Brahmaji
    Soora, Narasimha Reddy
    Shailaja, Kotte
    Kumar, N. C. Santosh
    Sridharan, Abel
    Uthayakumar, J.
    MULTIMEDIA TOOLS AND APPLICATIONS, 2023, 82 (08) : 12653 - 12677
  • [10] Multi-modal fusion of deep transfer learning based COVID-19 diagnosis and classification using chest x-ray images
    A. Siva Krishna Reddy
    K. N. Brahmaji Rao
    Narasimha Reddy Soora
    Kotte Shailaja
    N. C. Santosh Kumar
    Abel Sridharan
    J. Uthayakumar
    Multimedia Tools and Applications, 2023, 82 : 12653 - 12677