共 46 条
Granularity knowledge-sharing supervised contrastive learning framework for long-tailed fault diagnosis of rotating machinery
被引:2
|作者:
Chang, Shuyuan
[1
,2
]
Wang, Liyong
[1
,2
]
Shi, Mingkuan
[3
]
Zhang, Jinle
[4
]
Yang, Li
[1
,2
]
机构:
[1] Beijing Univ Technol, Key Lab Adv Mfg Technol, Beijing 100124, Peoples R China
[2] Beijing Informat Sci & Technol Univ, Key Lab Modern Measurement & Control Technol, Minist Educ, Beijing 100192, Peoples R China
[3] Soochow Univ, Sch Rail Transportat, Suzhou 215131, Peoples R China
[4] China North Vehicle Res Inst, Sci & Technol Vehicle Transmiss Lab, Beijing 100072, Peoples R China
基金:
中国国家自然科学基金;
关键词:
Fault diagnosis;
Fault judgement;
Long-tailed distribution;
Subclass clustering;
Supervised contrast learning;
D O I:
10.1016/j.knosys.2024.112354
中图分类号:
TP18 [人工智能理论];
学科分类号:
081104 ;
0812 ;
0835 ;
1405 ;
摘要:
The long-tailed distribution of monitoring data poses challenges for deep learning-based fault diagnosis (FD). Recent efforts utilizing supervised contrastive learning (SCL) and reweighted loss have made progress, but have overlooked two key issues: 1) prevailing random undersampling introduces sample influence bias and suboptimal model learning; and 2) focusing only on improving the FD average accuracy compromises fundamental fault judgement (FJ), heightening missed-detective and false-alarm risks unsuitable for real-world deployment. To fill these research gaps, this paper proposes a granularity knowledge-sharing SCL (GKSSCL) framework for longtailed FD, encompassing GKS supervised contrasting and GKS classification stages. In the former, normal data are clustered into multiple fine-grained subclasses that are similar in size to the fault categories for balanced contrasting. Moreover, a mixed-granularity contrastive loss facilitates knowledge sharing across granularities. In the latter, FJ and FD tasks were concurrently trained through a knowledge graph-based adaptive sharing strategy. Experiments on two fault datasets showed that the GKSSCL can effectively harness all normal data, eliminate sample influence bias, and enhance FD precision without sacrificing FJ reliability.
引用
收藏
页数:14
相关论文