PMFN-SSL: Self-supervised learning-based progressive multimodal fusion network for cancer diagnosis and prognosis

被引:3
|
作者
Li, Le [1 ,2 ]
Pan, Hudan [3 ]
Liang, Yong [1 ,4 ]
Shao, Mingwen [5 ]
Xie, Shengli [6 ]
Lu, Shanghui [2 ]
Liao, Shuilin [2 ]
机构
[1] Peng Cheng Lab, Shenzhen, Peoples R China
[2] Macau Univ Sci & Technol, Sch Fac Innovat Engn, Macau, Peoples R China
[3] Guangzhou Univ Chinese Med, Affiliated Hosp 2, State Key Lab Tradit Chinese Med Syndrome, Guangzhou, Peoples R China
[4] Pazhou Lab Huangpu, Guangzhou, Peoples R China
[5] China Univ Petr, Coll Comp Sci & Technol, Qingdao, Peoples R China
[6] Guangdong Univ Technol, Sch Automat, Guangzhou, Peoples R China
基金
美国国家科学基金会;
关键词
Multimodal learning; Self-supervised learning; Survival analysis; Grade prediction; ARTIFICIAL-INTELLIGENCE; SURVIVAL PREDICTION; CLASSIFICATION; IMAGES;
D O I
10.1016/j.knosys.2024.111502
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The integration of digital pathology images and genetic data is a developing field in cancer research, presenting potential opportunities for predicting survival and classifying grades through multiple source data. However, obtaining comprehensive annotations proves challenging in practical medical settings, and the extraction of features from high -resolution pathology images is hindered by inter-domain disparities. Current data fusion methods ignore the spatio-temporal incongruity among multimodal data. To address the above challenges, we propose a novel self-supervised transformer-based pathology feature extraction strategy, and construct an interpretable Progressive Multimodal Fusion Network (PMFN-SSL) for cancer diagnosis and prognosis. Our contributions are mainly divided into three aspects. Firstly, we propose a joint patch sampling strategy based on the information entropy and HSV components of an image, which reduces the demand for sample annotations and avoid image quality degradation caused by manual contamination. Secondly, a self-supervised transformerbased feature extraction module for pathology images is proposed and innovatively leverages partially weakly supervised labeling to align the extracted features with downstream medical tasks. Further, we improve the existing multimodal feature fusion model with an progressive fusion strategy to reduce the inconsistency between multimodal data due to differences in collection of temporal and spatial. Abundant ablation and comparison experiments demonstrate that the proposed data preprocessing method and multimodal fusion paradigm strengthen the quality of feature extraction and improve the prediction based on real cancer grading and prognosis. Code and trained models are made available at: https://github.com/Mercuriiio/PMFN-SSL.
引用
收藏
页数:12
相关论文
共 50 条
  • [41] Small-sample cucumber disease identification based on multimodal self-supervised learning
    Cao, Yiyi
    Sun, Guangling
    Yuan, Yuan
    Chen, Lei
    CROP PROTECTION, 2025, 188
  • [42] RobustSsF: Robust Missing Modality Brain Tumor Segmentation with Self-supervised Learning-Based Scenario-Specific Fusion
    Lee, Jeongwon
    Kim, Dae-Shik
    MACHINE LEARNING FOR MULTIMODAL HEALTHCARE DATA, ML4MHD 2023, 2024, 14315 : 43 - 53
  • [43] Learning-Based Cancer Treatment Outcome Prognosis Using Multimodal Biomarkers
    Saad, Maliazurina
    He, Shenghua
    Thorstad, Wade
    Gay, Hiram
    Barnett, Daniel
    Zhao, Yujie
    Ruan, Su
    Wang, Xiaowei
    Li, Hua
    IEEE TRANSACTIONS ON RADIATION AND PLASMA MEDICAL SCIENCES, 2022, 6 (02) : 231 - 244
  • [44] OTF: Optimal Transport based Fusion of Supervised and Self-Supervised Learning Models for Automatic Speech Recognition
    Fu, Li
    Li, Siqi
    Li, Qingtao
    Li, Fangzhu
    Deng, Liping
    Fan, Lu
    Chen, Meng
    Wu, Youzheng
    He, Xiaodong
    INTERSPEECH 2023, 2023, : 934 - 938
  • [45] A Self-Supervised Learning-Based 6-DOF Grasp Planning Method for Manipulator
    Peng, Gang
    Ren, Zhenyu
    Wang, Hao
    Li, Xinde
    Khyam, Mohammad Omar
    IEEE TRANSACTIONS ON AUTOMATION SCIENCE AND ENGINEERING, 2022, 19 (04) : 3639 - 3648
  • [46] Self-supervised pretraining enables deep learning-based classification of AMD with fewer annotations
    Holland, Robbie
    Menten, Martin Joseph
    Leingang, Oliver
    Bogunovic, Hrvoje
    Hagag, Ahmed M.
    Kaye, Rebecca
    Riedl, Sophie
    Traber, Ghislaine
    Fritsche, Lars
    Prevost, Toby
    Scholl, Hendrik P.
    Schmidt-Erfurth, Ursula
    Sivaprasad, Sobha
    Rueckert, Daniel
    Lotery, Andrew J.
    INVESTIGATIVE OPHTHALMOLOGY & VISUAL SCIENCE, 2022, 63 (07)
  • [47] A Novel Self-Supervised Learning-Based Method for Dynamic CT Brain Perfusion Imaging
    Liu, Chi-Kuang
    Huang, Hsuan-Ming
    JOURNAL OF IMAGING INFORMATICS IN MEDICINE, 2024,
  • [48] Self-supervised learning-based underwater acoustical signal classification via mask modeling
    Xu, Kele
    Xu, Qisheng
    You, Kang
    Zhu, Boqing
    Feng, Ming
    Feng, Dawei
    Liu, Bo
    JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA, 2023, 154 (01): : 5 - 15
  • [49] SSL-WAEIE:Self-Supervised Learning With Weighted Auto-Encoding and Information Exchange for Infrared and Visible Image Fusion
    Gucheng Zhang
    Rencan Nie
    Jinde Cao
    IEEE/CAA Journal of Automatica Sinica, 2022, 9 (09) : 1694 - 1697
  • [50] SSL-WAEIE: Self-Supervised Learning With Weighted Auto-Encoding and Information Exchange for Infrared and Visible Image Fusion
    Zhang, Gucheng
    Nie, Rencan
    Cao, Jinde
    IEEE-CAA JOURNAL OF AUTOMATICA SINICA, 2022, 9 (09) : 1694 - 1697