共 2 条
Application of variable temperature gradient TOP-K knowledge distillation with model pruning in lightweight fault diagnosis for bearings
被引:0
|作者:
Cui, Ze
[1
]
Yang, Qishuang
[1
]
Xiong, Zixiang
[1
]
Gu, Rongyang
[1
]
机构:
[1] Shanghai Univ, Sch Mechatron Engn & Automat, Shanghai 200444, Peoples R China
关键词:
fault diagnosis;
knowledge distillation;
model pruning;
D O I:
10.1088/1361-6501/ada6f3
中图分类号:
T [工业技术];
学科分类号:
08 ;
摘要:
In recent years, deep learning models have been extensively researched and applied in fault diagnosis. However, they often require substantial storage resources, posing challenges for deployment on embedded devices. A prevalent solution to this is leveraging knowledge distillation (KD) between teacher-student models. Through the distillation process, the student model can acquire knowledge from the teacher model without introducing additional parameters, thereby enhancing its performance. Nevertheless, when utilizing a powerful teacher model, the distillation performance is not always optimal. This is attributed to the teacher model's significantly higher complexity compared to the student model, potentially leading to a diminished simulation effect by the student model. To address this issue, the variable-temperature gradient TOP-K KD (VTGTK-KD) method is proposed, which employs multiple pruned, medium-sized teacher models to facilitate a gradual distillation learning process. Furthermore, these models share the same architecture, fostering better knowledge transfer conditions at the logical layer. To further elevate distillation performance, VT distillation is introduced to ensure a balance between distillation speed and accuracy. Additionally, the Gradient TOP-K algorithm is utilized to eliminate erroneous knowledge from the teacher network. Ultimately, classification experiments were conducted on two bearing datasets. The experimental results demonstrate that the proposed VTGTK-KD method enhances distillation performance, surpassing other advanced KD approaches.
引用
收藏
页数:9
相关论文