Cross-modal alignment and contrastive learning for enhanced cancer survival prediction

被引:0
|
作者
Li, Tengfei [1 ]
Zhou, Xuezhong [1 ]
Xue, Jingyan [1 ]
Zeng, Lili [1 ]
Zhu, Qiang [1 ]
Wang, Ruiping [1 ]
Yu, Haibin [2 ]
Xia, Jianan [1 ]
机构
[1] Beijing Jiaotong Univ, Sch Comp Sci & Technol, Beijing 100044, Peoples R China
[2] Henan Univ Chinese Med, Affiliated Hosp 1, Zhengzhou 450000, Henan, Peoples R China
基金
中国国家自然科学基金;
关键词
Survival prediction; Histopathological image; Multi-omics; Multi-modal fusion; MODEL;
D O I
10.1016/j.cmpb.2025.108633
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Background and Objective: Integrating multimodal data, such as pathology images and genomics, is crucial for understanding cancer heterogeneity, personalized treatment complexity, and enhancing survival prediction. However, most current prognostic methods are limited to a single domain of histopathology or genomics, inevitably reducing their potential for accurate patient outcome prediction. Despite advancements in the concurrent analysis of pathology and genomic data, existing approaches inadequately address the intricate intermodal relationships. Methods: This paper introduces the CPathomic method for multimodal data-based survival prediction. By leveraging whole slide pathology images to guide local pathological features, the method effectively mitigates significant intermodal differences through a cross-modal representational contrastive learning module. Furthermore, it facilitates interactive learning between different modalities through cross-modal and gated attention modules. Results: The extensive experiments on five public TCGA datasets demonstrate that CPathomic framework effectively bridges modality gaps, consistently outperforming alternative multimodal survival prediction methods. Conclusion: The model we propose, CPathomic, unveils the potential of contrastive learning and cross-modal attention in the representation and fusion of multimodal data, enhancing the performance of patient survival prediction.
引用
收藏
页数:10
相关论文
共 50 条
  • [41] Enhanced Multimodal Representation Learning with Cross-modal KD
    Chen, Mengxi
    Xing, Linyu
    Wang, Yu
    Zhang, Ya
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 11766 - 11775
  • [42] Early-Learning regularized Contrastive Learning for Cross-Modal Retrieval with Noisy Labels
    Xu, Tianyuan
    Liu, Xueliang
    Huang, Zhen
    Guo, Dan
    Hong, Richang
    Wang, Meng
    PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2022, 2022,
  • [43] Cross-Modal Joint Prediction and Alignment for Composed Query Image Retrieval
    Yang, Yuchen
    Wang, Min
    Zhou, Wengang
    Li, Houqiang
    PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2021, 2021, : 3303 - 3311
  • [44] Contrastive Cross-Modal Representation Learning Based Active Learning for Visual Question Answer
    Zhang B.-C.
    Li L.
    Zha Z.-J.
    Huang Q.-M.
    Jisuanji Xuebao/Chinese Journal of Computers, 2022, 45 (08): : 1730 - 1745
  • [45] UNIMO: Towards Unified-Modal Understanding and Generation via Cross-Modal Contrastive Learning
    Li, Wei
    Gao, Can
    Niu, Guocheng
    Xiao, Xinyan
    Liu, Hao
    Liu, Jiachen
    Wu, Hua
    Wang, Haifeng
    59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (ACL-IJCNLP 2021), VOL 1, 2021, : 2592 - 2607
  • [46] Gated Multi-modal Fusion with Cross-modal Contrastive Learning for Video Question Answering
    Lyu, Chenyang
    Li, Wenxi
    Ji, Tianbo
    Zhou, Liting
    Gurrin, Cathal
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING, ICANN 2023, PT VII, 2023, 14260 : 427 - 438
  • [47] Improving Speech Translation by Cross-Modal Multi-Grained Contrastive Learning
    Zhang, Hao
    Si, Nianwen
    Chen, Yaqi
    Zhang, Wenlin
    Yang, Xukui
    Qu, Dan
    Zhang, Wei-Qiang
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2023, 31 : 1075 - 1086
  • [48] Contrastive Learning with Cross-Modal Knowledge Mining for Multimodal Human Activity Recognition
    Brinzea, Razvan
    Khaertdinov, Bulat
    Asteriadis, Stylianos
    2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,
  • [49] Graph Information Interaction on Feature and Structure via Cross-modal Contrastive Learning
    Wen, Jinyong
    Wang, Yuhu
    Zhang, Chunxia
    Xiang, Shiming
    Pan, Chunhong
    2023 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO, ICME, 2023, : 1068 - 1073
  • [50] Lip and speech synchronization using supervised contrastive learning and cross-modal attention
    Varshney, Munender
    Mukherji, Mayurakshi
    Senthil, Raja G.
    Ganesh, Ananth
    Banerjee, Kingshuk
    2024 IEEE 18TH INTERNATIONAL CONFERENCE ON AUTOMATIC FACE AND GESTURE RECOGNITION, FG 2024, 2024,