Transformer-Based Multi-Modal Data Fusion Method for COPD Classification and Physiological and Biochemical Indicators Identification

被引:4
|
作者
Xie, Weidong [1 ]
Fang, Yushan [1 ]
Yang, Guicheng [1 ]
Yu, Kun [2 ]
Li, Wei [1 ,3 ]
机构
[1] Northeastern Univ, Sch Comp Sci & Engn, Shenyang 110169, Peoples R China
[2] Northeastern Univ, Coll Med & Bioinformat Engn, Shenyang 110169, Peoples R China
[3] Key Lab Intelligent Comp Med Image MIIC, Shenyang 110169, Peoples R China
关键词
multi-modal fusion; cross-modal transformer; low-rank multi-modal fusion; COPD; PREDICTION; PROGNOSIS;
D O I
10.3390/biom13091391
中图分类号
Q5 [生物化学]; Q7 [分子生物学];
学科分类号
071010 ; 081704 ;
摘要
As the number of modalities in biomedical data continues to increase, the significance of multi-modal data becomes evident in capturing complex relationships between biological processes, thereby complementing disease classification. However, the current multi-modal fusion methods for biomedical data require more effective exploitation of intra- and inter-modal interactions, and the application of powerful fusion methods to biomedical data is relatively rare. In this paper, we propose a novel multi-modal data fusion method that addresses these limitations. Our proposed method utilizes a graph neural network and a 3D convolutional network to identify intra-modal relationships. By doing so, we can extract meaningful features from each modality, preserving crucial information. To fuse information from different modalities, we employ the Low-rank Multi-modal Fusion method, which effectively integrates multiple modalities while reducing noise and redundancy. Additionally, our method incorporates the Cross-modal Transformer to automatically learn relationships between different modalities, facilitating enhanced information exchange and representation. We validate the effectiveness of our proposed method using lung CT imaging data and physiological and biochemical data obtained from patients diagnosed with Chronic Obstructive Pulmonary Disease (COPD). Our method demonstrates superior performance compared to various fusion methods and their variants in terms of disease classification accuracy.
引用
收藏
页数:18
相关论文
共 50 条
  • [41] A multi-modal emotion fusion classification method combined expression and speech based on attention mechanism
    Dong Liu
    Longxi Chen
    Lifeng Wang
    Zhiyong Wang
    Multimedia Tools and Applications, 2022, 81 : 41677 - 41695
  • [42] SERVER: Multi-modal Speech Emotion Recognition using Transformer-based and Vision-based Embeddings
    Nhat Truong Pham
    Duc Ngoc Minh Dang
    Bich Ngoc Hong Pham
    Sy Dzung Nguyen
    PROCEEDINGS OF 2023 8TH INTERNATIONAL CONFERENCE ON INTELLIGENT INFORMATION TECHNOLOGY, ICIIT 2023, 2023, : 234 - 238
  • [43] FNR: a similarity and transformer-based approach to detect multi-modal fake news in social media
    Ghorbanpour, Faeze
    Ramezani, Maryam
    Fazli, Mohammad Amin
    Rabiee, Hamid R.
    SOCIAL NETWORK ANALYSIS AND MINING, 2023, 13 (01)
  • [44] FNR: a similarity and transformer-based approach to detect multi-modal fake news in social media
    Faeze Ghorbanpour
    Maryam Ramezani
    Mohammad Amin Fazli
    Hamid R. Rabiee
    Social Network Analysis and Mining, 13
  • [45] UniTR: A Unified TRansformer-Based Framework for Co-Object and Multi-Modal Saliency Detection
    Guo, Ruohao
    Ying, Xianghua
    Qi, Yanyu
    Qu, Liao
    IEEE TRANSACTIONS ON MULTIMEDIA, 2024, 26 : 7622 - 7635
  • [46] Dual-attention transformer-based hybrid network for multi-modal medical image segmentation
    Zhang, Menghui
    Zhang, Yuchen
    Liu, Shuaibing
    Han, Yahui
    Cao, Honggang
    Qiao, Bingbing
    SCIENTIFIC REPORTS, 2024, 14 (01):
  • [47] Improving motion sickness severity classification through multi-modal data fusion
    Dennison, Mark
    D'Zmura, Mike
    Harrison, Andre
    Lee, Michael
    Raglin, Adrienne
    ARTIFICIAL INTELLIGENCE AND MACHINE LEARNING FOR MULTI-DOMAIN OPERATIONS APPLICATIONS, 2019, 11006
  • [48] Real-time localization and navigation method for autonomous vehicles based on multi-modal data fusion by integrating memory transformer and DDQN
    Zha, Li
    Gong, Chen
    Lv, Kunfeng
    IMAGE AND VISION COMPUTING, 2025, 156
  • [49] Multi-modal person re-identification based on transformer relational regularization
    Zheng, Xiangtian
    Huang, Xiaohua
    Ji, Chen
    Yang, Xiaolin
    Sha, Pengcheng
    Cheng, Liang
    INFORMATION FUSION, 2024, 103
  • [50] A Vision Transformer-Based Framework for Knowledge Transfer From Multi-Modal to Mono-Modal Lymphoma Subtyping Models
    Guetarni, Bilel
    Windal, Feryal
    Benhabiles, Halim
    Petit, Marianne
    Dubois, Romain
    Leteurtre, Emmanuelle
    Collard, Dominique
    IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS, 2024, 28 (09) : 5562 - 5572