MoMA: Momentum contrastive learning with multi-head attention-based knowledge distillation for histopathology image analysis

被引:0
|
作者
Vuong, Trinh Thi Le [1 ]
Kwak, Jin Tae [1 ]
机构
[1] Korea Univ, Sch Elect Engn, Seoul, South Korea
基金
新加坡国家研究基金会;
关键词
Knowledge distillation; Momentum contrast; Multi-head self-attention; Computational pathology; CANCER;
D O I
10.1016/j.media.2024.103421
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
There is no doubt that advanced artificial intelligence models and high quality data are the keys to success in developing computational pathology tools. Although the overall volume of pathology data keeps increasing, a lack of quality data is a common issue when it comes to a specific task due to several reasons including privacy and ethical issues with patient data. In this work, we propose to exploit knowledge distillation, i.e., utilize the existing model to learn a new, target model, to overcome such issues in computational pathology. Specifically, we employ a student-teacher framework to learn a target model from a pre-trained, teacher model without direct access to source data and distill relevant knowledge via momentum contrastive learning with multi-head attention mechanism, which provides consistent and context-aware feature representations. This enables the target model to assimilate informative representations of the teacher model while seamlessly adapting to the unique nuances of the target data. The proposed method is rigorously evaluated across different scenarios where the teacher model was trained on the same, relevant, and irrelevant classification tasks with the target model. Experimental results demonstrate the accuracy and robustness of our approach in transferring knowledge to different domains and tasks, outperforming other related methods. Moreover, the results provide a guideline on the learning strategy for different types of tasks and scenarios in computational pathology.
引用
收藏
页数:22
相关论文
共 50 条
  • [21] Multi-head attention-based model for reconstructing continuous missing time series data
    Huafeng Wu
    Yuxuan Zhang
    Linian Liang
    Xiaojun Mei
    Dezhi Han
    Bing Han
    Tien-Hsiung Weng
    Kuan-Ching Li
    The Journal of Supercomputing, 2023, 79 : 20684 - 20711
  • [22] Triple-kernel gated attention-based multiple instance learning with contrastive learning for medical image analysis
    Hu, Huafeng
    Ye, Ruijie
    Thiyagalingam, Jeyan
    Coenen, Frans
    Su, Jionglong
    APPLIED INTELLIGENCE, 2023, 53 (17) : 20311 - 20326
  • [23] Triple-kernel gated attention-based multiple instance learning with contrastive learning for medical image analysis
    Huafeng Hu
    Ruijie Ye
    Jeyan Thiyagalingam
    Frans Coenen
    Jionglong Su
    Applied Intelligence, 2023, 53 : 20311 - 20326
  • [24] Multi-head attention-based masked sequence model for mapping functional brain networks
    He, Mengshen
    Hou, Xiangyu
    Ge, Enjie
    Wang, Zhenwei
    Kang, Zili
    Qiang, Ning
    Zhang, Xin
    Ge, Bao
    FRONTIERS IN NEUROSCIENCE, 2023, 17
  • [25] Multimodal sentiment analysis based on multi-head attention mechanism
    Xi, Chen
    Lu, Guanming
    Yan, Jingjie
    ICMLSC 2020: PROCEEDINGS OF THE 4TH INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND SOFT COMPUTING, 2020, : 34 - 39
  • [26] Image-Based Fitness Yoga Pose Recognition: Using Ensemble Learning and Multi-head Attention
    Kou, Yue
    Li, Hai
    INTERNATIONAL JOURNAL OF COMPUTATIONAL INTELLIGENCE SYSTEMS, 2024, 17 (01)
  • [27] A multi-head attention-based transformer model for traffic flow forecasting with a comparative analysis to recurrent neural networks
    Reza, Selim
    Ferreira, Marta Campos
    Machado, J. J. M.
    Tavares, Joao Manuel R. S.
    EXPERT SYSTEMS WITH APPLICATIONS, 2022, 202
  • [28] DeepCKID: A Multi-Head Attention-Based Deep Neural Network Model Leveraging Classwise Knowledge to Handle Imbalanced Textual Data
    Sah, Amit Kumar
    Abulaish, Muhammad
    MACHINE LEARNING WITH APPLICATIONS, 2024, 17
  • [29] Attention-Based Contrastive Learning for Few-Shot Remote Sensing Image Classification
    Xu, Yulong
    Bi, Hanbo
    Yu, Hongfeng
    Lu, Wanxuan
    Li, Peifeng
    Li, Xinming
    Sun, Xian
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2024, 62
  • [30] Estimating ocean thermocline from satellite observations with a multi-head attention-based neural network
    Deng, Fangyu
    Pan, Yanxi
    Wang, Jichao
    OCEAN DYNAMICS, 2025, 75 (02)