Application of variable temperature gradient TOP-K knowledge distillation with model pruning in lightweight fault diagnosis for bearings

被引:0
|
作者
Cui, Ze [1 ]
Yang, Qishuang [1 ]
Xiong, Zixiang [1 ]
Gu, Rongyang [1 ]
机构
[1] Shanghai Univ, Sch Mechatron Engn & Automat, Shanghai 200444, Peoples R China
关键词
fault diagnosis; knowledge distillation; model pruning;
D O I
10.1088/1361-6501/ada6f3
中图分类号
T [工业技术];
学科分类号
08 ;
摘要
In recent years, deep learning models have been extensively researched and applied in fault diagnosis. However, they often require substantial storage resources, posing challenges for deployment on embedded devices. A prevalent solution to this is leveraging knowledge distillation (KD) between teacher-student models. Through the distillation process, the student model can acquire knowledge from the teacher model without introducing additional parameters, thereby enhancing its performance. Nevertheless, when utilizing a powerful teacher model, the distillation performance is not always optimal. This is attributed to the teacher model's significantly higher complexity compared to the student model, potentially leading to a diminished simulation effect by the student model. To address this issue, the variable-temperature gradient TOP-K KD (VTGTK-KD) method is proposed, which employs multiple pruned, medium-sized teacher models to facilitate a gradual distillation learning process. Furthermore, these models share the same architecture, fostering better knowledge transfer conditions at the logical layer. To further elevate distillation performance, VT distillation is introduced to ensure a balance between distillation speed and accuracy. Additionally, the Gradient TOP-K algorithm is utilized to eliminate erroneous knowledge from the teacher network. Ultimately, classification experiments were conducted on two bearing datasets. The experimental results demonstrate that the proposed VTGTK-KD method enhances distillation performance, surpassing other advanced KD approaches.
引用
收藏
页数:9
相关论文
共 2 条
  • [1] Multiassistant Knowledge Distillation for Lightweight Bearing Fault Diagnosis Based on Decreasing Threshold Channel Pruning
    Zhong, Hongyu
    Yu, Samson
    Trinh, Hieu
    Lv, Yong
    Yuan, Rui
    Wang, Yanan
    IEEE SENSORS JOURNAL, 2024, 24 (01) : 486 - 494
  • [2] A lightweight residual network based on improved knowledge transfer and quantized distillation for cross-domain fault diagnosis of rolling bearings
    Guo, Wei
    Li, Xiang
    Shen, Ziqian
    EXPERT SYSTEMS WITH APPLICATIONS, 2024, 245