Classifying Excavator Operations with Fusion Network of Multi-modal Deep Learning Models

被引:1
|
作者
Kim, Jin-Young [1 ]
Cho, Sung-Bae [1 ]
机构
[1] Yonsei Univ, Dept Comp Sci, Seoul, South Korea
关键词
Excavator; Classification; Deep learning; Multi-modal data; Autoencoder; Feature extraction; CLASSIFICATION; ALGORITHM;
D O I
10.1007/978-3-030-20055-8_3
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Prognostics and health management (PHM) aims to offer comprehensive solutions for managing equipment health. Classifying the excavator operations plays an important role in measuring the lifetime, which is one of the tasks in PHM because the effect on the lifetime depends on the operations performed by the excavator. Several researchers have struggled with classifying the operations with either sensor or video data, but most of them have difficulties with the use of single modal data only, the surrounding environment, and the exclusive feature extraction for the data in different domains. In this paper, we propose a fusion network that classifies the excavator operations with multi-modal deep learning models. Trained are multiple classifiers with specific type of data, where feature extractors are reused to place at the front of the fusion network. The proposed fusion network combines a video-based model and a sensor-based model based on deep learning. To evaluate the performance of the proposed method, experiments are conducted with the data collected from real construction workplace. The proposed method yields the accuracy of 98.48% which is higher than conventional methods, and the multi-modal deep learning models can complement each other in terms of precision, recall, and F1-score.
引用
收藏
页码:25 / 34
页数:10
相关论文
共 50 条
  • [1] Exploring Fusion Strategies in Deep Learning Models for Multi-Modal Classification
    Zhang, Duoyi
    Nayak, Richi
    Bashar, Md Abul
    [J]. DATA MINING, AUSDM 2021, 2021, 1504 : 102 - 117
  • [2] Robust Deep Multi-modal Learning Based on Gated Information Fusion Network
    Kim, Jaekyum
    Koh, Junho
    Kim, Yecheol
    Choi, Jaehyung
    Hwang, Youngbae
    Choi, Jun Won
    [J]. COMPUTER VISION - ACCV 2018, PT IV, 2019, 11364 : 90 - 106
  • [3] Memory based fusion for multi-modal deep learning
    Priyasad, Darshana
    Fernando, Tharindu
    Denman, Simon
    Sridharan, Sridha
    Fookes, Clinton
    [J]. INFORMATION FUSION, 2021, 67 : 136 - 146
  • [4] A Multi-modal Deep Learning Method for Classifying Chest Radiology Exams
    Nunes, Nelson
    Martins, Bruno
    da Silva, Nuno Andre
    Leite, Francisca
    Silva, Mario J.
    [J]. PROGRESS IN ARTIFICIAL INTELLIGENCE, EPIA 2019, PT I, 2019, 11804 : 323 - 335
  • [5] A deep neural network ensemble of multimodal signals for classifying excavator operations
    Kim, Jin-Young
    Cho, Sung-Bae
    [J]. NEUROCOMPUTING, 2022, 470 : 290 - 299
  • [6] Deep Convolutional Neural Network for Multi-Modal Image Restoration and Fusion
    Deng, Xin
    Dragotti, Pier Luigi
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2021, 43 (10) : 3333 - 3348
  • [7] Deep unsupervised multi-modal fusion network for detecting driver distraction
    Zhang, Yuxin
    Chen, Yiqiang
    Gao, Chenlong
    [J]. NEUROCOMPUTING, 2021, 421 : 26 - 38
  • [8] Deep unsupervised multi-modal fusion network for detecting driver distraction
    Zhang Y.
    Chen Y.
    Gao C.
    [J]. Neurocomputing, 2021, 421 : 26 - 38
  • [9] A comparative review on multi-modal sensors fusion based on deep learning
    Tang, Qin
    Liang, Jing
    Zhu, Fangqi
    [J]. SIGNAL PROCESSING, 2023, 213
  • [10] Multi-Modal Object Tracking and Image Fusion With Unsupervised Deep Learning
    LaHaye, Nicholas
    Ott, Jordan
    Garay, Michael J.
    El-Askary, Hesham Mohamed
    Linstead, Erik
    [J]. IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, 2019, 12 (08) : 3056 - 3066