Discrepancy and Gradient-Guided Multi-modal Knowledge Distillation for Pathological Glioma Grading

被引:15
|
作者
Xing, Xiaohan [1 ]
Chen, Zhen [1 ]
Zhu, Meilu [1 ]
Hou, Yuenan [2 ]
Gao, Zhifan [3 ]
Yuan, Yixuan [1 ]
机构
[1] City Univ Hong Kong, Dept Elect Engn, Kowloon, Hong Kong, Peoples R China
[2] Shanghai Artificial Intelligence Lab, Shanghai, Peoples R China
[3] Sun Yat Sen Univ, Sch Biomed Engn, Guangzhou, Guangdong, Peoples R China
关键词
Knowledge distillation; Missing modality; Glioma grading;
D O I
10.1007/978-3-031-16443-9_61
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
The fusion of multi-modal data, e.g., pathology slides and genomic profiles, can provide complementary information and benefit glioma grading. However, genomic profiles are difficult to obtain due to the high costs and technical challenges, thus limiting the clinical applications of multi-modal diagnosis. In this work, we address the clinically relevant problem where paired pathology-genomic data are available during training, while only pathology slides are accessible for inference. To improve the performance of pathological grading models, we present a discrepancy and gradient-guided distillation framework to transfer the privileged knowledge from the multi-modal teacher to the pathology student. For the teacher side, to prepare useful knowledge, we propose a Discrepancy-induced Contrastive Distillation (DC-Distill) module that explores reliable contrastive samples with teacher-student discrepancy to regulate the feature distribution of the student. For the student side, as the teacher may include incorrect information, we propose a Gradient-guided Knowledge Refinement (GK-Refine) module that builds a knowledge bank and adaptively absorbs the reliable knowledge according to their agreement in the gradient space. Experiments on the TCGA GBM-LGG dataset show that our proposed distillation framework improves the pathological glioma grading significantly and outperforms other KD methods. Notably, with the sole pathology slides, our method achieves comparable performance with existing multi-modal methods. The code is available at https://github.com/CityU-AIM-Group/MultiModal-learning.
引用
收藏
页码:636 / 646
页数:11
相关论文
共 50 条
  • [41] Gradient structural similarity based gradient filtering for multi-modal image fusion
    Fu, Zhizhong
    Zhao, Yufei
    Xu, Yuwei
    Xu, Lijuan
    Xu, Jin
    INFORMATION FUSION, 2020, 53 (251-268) : 251 - 268
  • [42] MMEA: Entity Alignment for Multi-modal Knowledge Graph
    Chen, Liyi
    Li, Zhi
    Wang, Yijun
    Xu, Tong
    Wang, Zhefeng
    Chen, Enhong
    KNOWLEDGE SCIENCE, ENGINEERING AND MANAGEMENT (KSEM 2020), PT I, 2020, 12274 : 134 - 147
  • [43] A Survey of Multi-modal Knowledge Graphs: Technologies and Trends
    Liang, Wanying
    De Meo, Pasquale
    Tang, Yong
    Zhu, Jia
    ACM COMPUTING SURVEYS, 2024, 56 (11)
  • [44] Multi-Modal Knowledge Hypergraph for Diverse Image Retrieval
    Zeng, Yawen
    Jin, Qin
    Bao, Tengfei
    Li, Wenfeng
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 3, 2023, : 3376 - 3383
  • [45] Contrastive Multi-Modal Knowledge Graph Representation Learning
    Fang, Quan
    Zhang, Xiaowei
    Hu, Jun
    Wu, Xian
    Xu, Changsheng
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2023, 35 (09) : 8983 - 8996
  • [46] Multi-modal news event detection with external knowledge
    Lin, Zehang
    Xie, Jiayuan
    Li, Qing
    INFORMATION PROCESSING & MANAGEMENT, 2024, 61 (03)
  • [47] Combining Knowledge and Multi-modal Fusion for Meme Classification
    Zhong, Qi
    Wang, Qian
    Liu, Ji
    MULTIMEDIA MODELING (MMM 2022), PT I, 2022, 13141 : 599 - 611
  • [48] NativE: Multi-modal Knowledge Graph Completion in the Wild
    Zhang, Yichi
    Chen, Zhuo
    Guo, Lingbing
    Xu, Yajing
    Hu, Binbin
    Liu, Ziqi
    Zhang, Wen
    Chen, Huajun
    PROCEEDINGS OF THE 47TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2024, 2024, : 91 - 101
  • [49] Online Distillation-enhanced Multi-modal Transformer for Sequential Recommendation
    Ji, Wei
    Liu, Xiangyan
    Zhang, An
    Wei, Yinwei
    Ni, Yongxin
    Wang, Xiang
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 955 - 965
  • [50] Generation of Visual Representations for Multi-Modal Mathematical Knowledge
    Wu, Lianlong
    Choi, Seewon
    Raggi, Daniel
    Stockdill, Aaron
    Garcia, Grecia Garcia
    Colarusso, Fiorenzo
    Cheng, Peter C. H.
    Jamnik, Mateja
    THIRTY-EIGTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 21, 2024, : 23850 - 23852