Breast cancer diagnosis through knowledge distillation of Swin transformer-based teacher-student models

被引:0
|
作者
Kolla, Bhavannarayanna [1 ]
Venugopal, P. [1 ]
机构
[1] Vellore Inst Technol, Sch Elect Engn, Vellore 632014, Tamil Nadu, India
来源
关键词
teacher model; student model; Swin-transformers; transfer learning; knowledge distillation; breast cancer histopathology;
D O I
10.1088/2632-2153/ad10cc
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Breast cancer is a significant global health concern, emphasizing the crucial need for a timely and accurate diagnosis to enhance survival rates. Traditional diagnostic methods rely on pathologists analyzing whole-slide images (WSIs) to identify and diagnose malignancies. However, this task is complex, demanding specialized expertise and imposing a substantial workload on pathologists. Additionally, existing deep learning models, commonly employed for classifying histopathology images, often need enhancements to ensure their suitability for real-time deployment on WSI, especially when trained for small regions of interest (ROIs). This article introduces two Swin transformer-based architectures: the teacher model, characterized by its moderate size, and the lightweight student model. Both models are trained using a publicly available dataset of breast cancer histopathology images, focusing on ROIs with varying magnification factors. Transfer learning is applied to train the teacher model, and knowledge distillation (KD) transfers its capabilities to the student model. To enhance validation accuracy and minimize the total loss in KD, we employ the state-action-reward-state-action (SARSA) reinforcement learning algorithm. The algorithm dynamically computes temperature and a weighting factor throughout the KD process to achieve high accuracy within a considerably shorter training timeframe. Additionally, the student model is deployed to analyze malignancies in WSI. Despite the student model being only one-third the size and flops of the teacher model, it achieves an impressive accuracy of 98.71%, slightly below the teacher's accuracy of 98.91%. Experimental results demonstrate that the student model can process WSIs at a throughput of 1.67 samples s-1 with an accuracy of 82%. The proposed student model, trained using KD and the SARSA algorithm, exhibits promising breast cancer classification and WSI analysis performance. These findings indicate its potential for assisting pathologists in diagnosing breast cancer accurately and effectively.
引用
收藏
页数:12
相关论文
共 50 条
  • [1] Medical image classification: Knowledge transfer via residual U-Net and vision transformer-based teacher-student model with knowledge distillation
    Song, Yucheng
    Wang, Jincan
    Ge, Yifan
    Li, Lifeng
    Guo, Jia
    Dong, Quanxing
    Liao, Zhifang
    JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2024, 102
  • [2] Transformer fault diagnosis based on relational teacher-student network
    Sihan Y.
    Yalei L.
    Xiaoping L.
    Xu C.
    Huapeng W.
    Journal of China Universities of Posts and Telecommunications, 2023, 30 (03): : 41 - 54
  • [3] Transformer fault diagnosis based on relational teacher-student network
    Yin Sihan
    Li Yalei
    Liu Xiaoping
    Cui Xu
    Wang Huapeng
    The Journal of China Universities of Posts and Telecommunications, 2023, 30 (03) : 41 - 54
  • [4] Teacher-student collaborative knowledge distillation for image classification
    Xu, Chuanyun
    Gao, Wenjian
    Li, Tian
    Bai, Nanlan
    Li, Gang
    Zhang, Yang
    APPLIED INTELLIGENCE, 2023, 53 (02) : 1997 - 2009
  • [5] Exploring the Knowledge Transferred by Response-Based Teacher-Student Distillation
    Song, Liangchen
    Gong, Xuan
    Zhou, Helong
    Chen, Jiajie
    Zhang, Qian
    Doermann, David
    Yuan, Junsong
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 2704 - 2713
  • [6] Teacher-student collaborative knowledge distillation for image classification
    Chuanyun Xu
    Wenjian Gao
    Tian Li
    Nanlan Bai
    Gang Li
    Yang Zhang
    Applied Intelligence, 2023, 53 : 1997 - 2009
  • [7] Hybrid Learning with Teacher-student Knowledge Distillation for Recommenders
    Zhang, Hangbin
    Wong, Raymond K.
    Chu, Victor W.
    20TH IEEE INTERNATIONAL CONFERENCE ON DATA MINING WORKSHOPS (ICDMW 2020), 2020, : 227 - 235
  • [8] Joint learning method with teacher-student knowledge distillation for on-device breast cancer image classification
    Sepahvand, Majid
    Abdali-Mohammadi, Fardin
    COMPUTERS IN BIOLOGY AND MEDICINE, 2023, 155
  • [9] Dual knowledge distillation for visual tracking with teacher-student network
    Wang, Yuanyun
    Sun, Chuanyu
    Wang, Jun
    Chai, Bingfei
    SIGNAL IMAGE AND VIDEO PROCESSING, 2024, 18 (6-7) : 5203 - 5211
  • [10] Teacher-Student Knowledge Distillation for Radar Perception on Embedded Accelerators
    Shaw, Steven
    Tyagi, Kanishka
    Zhang, Shan
    FIFTY-SEVENTH ASILOMAR CONFERENCE ON SIGNALS, SYSTEMS & COMPUTERS, IEEECONF, 2023, : 1035 - 1038