A multi-modal joint attention network for vibro-acoustic fusion diagnosis of engines

被引:0
|
作者
Zhu, Xiaolong [1 ]
Zhang, Junhong [1 ,2 ]
Wang, Xinwei [1 ,3 ]
Wang, Hui [1 ,3 ]
Lin, Jiewei [1 ]
机构
[1] Tianjin Univ, State Key Lab Engines, Tianjin 300072, Peoples R China
[2] Renai Coll, Tianjin 300072, Peoples R China
[3] Weichai Power Co Ltd, Weifang 261001, Peoples R China
基金
国家重点研发计划;
关键词
multi-modal joint attention network; deep learning; fault diagnosis; internal combustion engine; vibration acoustic fusion diagnosis; COMBUSTION ENGINES; VIBRATION SIGNAL; EXTRACTION;
D O I
10.1088/1361-6501/ad4fb4
中图分类号
T [工业技术];
学科分类号
08 ;
摘要
Deep learning has proven to be effective in diagnosing faults in power machinery and its diagnosis performance relies on a sufficient data set. In practice, a well-labeled data set with sufficient samples is very rare, especially for those machinery running in varying loading cases. The situation is particularly pronounced for multi-cylinder internal combustion engines, where the excitations from cylinders interact with significant background noise, and different data distributions are complicated. To tackle these issues, we propose a novelty multi-modal joint attention network (MJA-Net) for fusing the vibration and acoustic signals for diagnosing multiple faults. In MJA-Net, feature maps from both modalities are input separately into the convolutional module to learn independent features, and joint attention module (JAM) is utilized to enhance the vibro-acoustic information interaction and distribution consistency across modalities. The analysis of multiple loads vibro-acoustic experimental data shows that MJA-Net has a superior classification performance in limited sample tasks, compared to the single-modal methods. Furthermore, MJA-Net outperforms other fusion methods on average accuracy at 97.65%, as well as feature representativeness, and vibro-acoustic feature consistency across loads. JAM has superior diagnosis performance to other alternative modules. The class activation maps (CAM) generated by the Layer CAM highlight the key impact components related to the engine working mechanisms, providing valuable insight into MJA-Net's interpretation for multi-fault recognition.
引用
收藏
页数:17
相关论文
共 50 条
  • [1] Self-supervised multi-modal fusion network for multi-modal thyroid ultrasound image diagnosis
    Xiang, Zhuo
    Zhuo, Qiuluan
    Zhao, Cheng
    Deng, Xiaofei
    Zhu, Ting
    Wang, Tianfu
    Jiang, Wei
    Lei, Baiying
    [J]. COMPUTERS IN BIOLOGY AND MEDICINE, 2022, 150
  • [2] A Tri-Attention fusion guided multi-modal segmentation network
    Zhou, Tongxue
    Ruan, Su
    Vera, Pierre
    Canu, Stephane
    [J]. PATTERN RECOGNITION, 2022, 124
  • [3] Multi-input CNN based vibro-acoustic fusion for accurate fault diagnosis of induction motor
    Choudhary, Anurag
    Mishra, Rismaya Kumar
    Fatima, Shahab
    Panigrahi, B. K.
    [J]. ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2023, 120
  • [4] Joint Attention Estimation during Multi-party Facilitation Using Multi-Modal Fusion
    Chew, Jouh Yeong
    Wang, Xiaohan
    [J]. COMPANION OF THE 2024 ACM/IEEE INTERNATIONAL CONFERENCE ON HUMAN-ROBOT INTERACTION, HRI 2024 COMPANION, 2024, : 322 - 326
  • [5] Bearing fault diagnosis based on vibro-acoustic data fusion and 1D-CNN network
    Wang, Xin
    Mao, Dongxing
    Li, Xiaodong
    [J]. MEASUREMENT, 2021, 173
  • [6] Attention-Based Multi-Modal Fusion Network for Semantic Scene Completion
    Li, Siqi
    Zou, Changqing
    Li, Yipeng
    Zhao, Xibin
    Gao, Yue
    [J]. THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 11402 - 11409
  • [7] MSAFusionNet: Multiple Subspace Attention Based Deep Multi-modal Fusion Network
    Zhang, Sen
    Zhang, Changzheng
    Wang, Lanjun
    Li, Cixing
    Tu, Dandan
    Luo, Rui
    Qi, Guojun
    Luo, Jiebo
    [J]. MACHINE LEARNING IN MEDICAL IMAGING (MLMI 2019), 2019, 11861 : 54 - 62
  • [8] Holographic modal analysis for the investigation of vibro-acoustic systems
    Haberstok, C
    Freymann, R
    Steinbichler, H
    Van der Auweraer, H
    Vanlanduit, S
    [J]. LASER METROLOGY AND INSPECTION, 1999, 3823 : 26 - 37
  • [9] Joint Segmentation and Grasp Pose Detection with Multi-Modal Feature Fusion Network
    Liu, Xiaozheng
    Zhang, Yunzhou
    Cao, He
    Shan, Dexing
    Zhao, Jiaqi
    [J]. 2023 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION, ICRA, 2023, : 1751 - 1756
  • [10] ATTENTION DRIVEN FUSION FOR MULTI-MODAL EMOTION RECOGNITION
    Priyasad, Darshana
    Fernando, Tharindu
    Denman, Simon
    Sridharan, Sridha
    Fookes, Clinton
    [J]. 2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 3227 - 3231