A comparative review on multi-modal sensors fusion based on deep learning

被引:22
|
作者
Tang, Qin [1 ]
Liang, Jing [1 ]
Zhu, Fangqi [2 ]
机构
[1] Univ Elect Sci & Technol China, Sch Informat & Commun Engn, Xiyuan Ave, Chengdu 611731, Sichuan, Peoples R China
[2] Seagate Technol, Longmont, CO 80501 USA
来源
SIGNAL PROCESSING | 2023年 / 213卷
关键词
Multi-model data fusion; Deep learning; Inference mechanisms; MEDICAL IMAGE FUSION; HUMAN-ROBOT INTERACTION; NEURAL-NETWORK; QUALITY ASSESSMENT; OBJECT DETECTION; TARGET TRACKING; KALMAN FILTER; LAND-USE; CLASSIFICATION; AUTOENCODER;
D O I
10.1016/j.sigpro.2023.109165
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
The wide deployment of multi-modal sensors in various areas generates vast amounts of data with characteristics of high volume, wide variety, and high integrity. However, traditional data fusion methods face immense challenges when dealing with multi-modal data containing abundant intermodality and cross-modality information. Deep learning has the ability to automatically extract and understand the potential association of multi modal information. Despite this, there is a lack of a comprehensive review of the inherent inference mechanisms of deep learning for multi-modal sensor fusion. This work investigates up-to-date developments in multi-modal sensor fusion via deep learning to provide a broad picture of data fusion needs and technologies. It compares the characteristics of multi-modal data for various sensors, summarizes background concepts about data fusion and deep learning, and carefully reviews a large number of investigations in four inference mechanisms: adaptive learning, deep generative, deep discriminative, and algorithms unrolling. The pros and cons of the above methodologies are presented, and several popular application domains are discussed, including medical imaging, autonomous driving, remote sensing, and robotics. A large collection of multi-modal datasets published in recent years is presented, and several tables that quantitatively compare and summarize the performance of fusion algorithms are provided. Finally, by acknowledging the limitations of current research, we establish potential open challenges and future directions as guidance for deep learning-based multi-sensor fusion.
引用
收藏
页数:20
相关论文
共 50 条
  • [1] Memory based fusion for multi-modal deep learning
    Priyasad, Darshana
    Fernando, Tharindu
    Denman, Simon
    Sridharan, Sridha
    Fookes, Clinton
    [J]. INFORMATION FUSION, 2021, 67 : 136 - 146
  • [2] Deep learning based object detection from multi-modal sensors: an overview
    Ye Liu
    Shiyang Meng
    Hongzhang Wang
    Jun Liu
    [J]. Multimedia Tools and Applications, 2024, 83 : 19841 - 19870
  • [3] Deep learning based object detection from multi-modal sensors: an overview
    Liu, Ye
    Meng, Shiyang
    Wang, Hongzhang
    Liu, Jun
    [J]. MULTIMEDIA TOOLS AND APPLICATIONS, 2024, 83 (07) : 19841 - 19870
  • [4] Electromagnetic signal feature fusion and recognition based on multi-modal deep learning
    Hou C.
    Zhang X.
    Chen X.
    [J]. International Journal of Performability Engineering, 2020, 16 (06): : 941 - 949
  • [5] Deep Learning Based Multi-Modal Fusion Architectures for Maritime Vessel Detection
    Farahnakian, Fahimeh
    Heikkonen, Jukka
    [J]. REMOTE SENSING, 2020, 12 (16)
  • [6] Multi-modal Fusion Brain Tumor Detection Method Based on Deep Learning
    Yao Hong-ge
    Shen Xin-xia
    Li Yu
    Yu Jun
    Lei Song-ze
    [J]. ACTA PHOTONICA SINICA, 2019, 48 (07)
  • [7] Cardiovascular disease detection based on deep learning and multi-modal data fusion
    Zhu, Jiayuan
    Liu, Hui
    Liu, Xiaowei
    Chen, Chao
    Shu, Minglei
    [J]. Biomedical Signal Processing and Control, 2025, 99
  • [8] Deep-Learning-Based Multi-Modal Fusion for Fast MR Reconstruction
    Xiang, Lei
    Chen, Yong
    Chang, Weitang
    Zhan, Yiqiang
    Lin, Weili
    Wang, Qian
    Shen, Dinggang
    [J]. IEEE TRANSACTIONS ON BIOMEDICAL ENGINEERING, 2019, 66 (07) : 2105 - 2114
  • [9] Robust Deep Multi-modal Learning Based on Gated Information Fusion Network
    Kim, Jaekyum
    Koh, Junho
    Kim, Yecheol
    Choi, Jaehyung
    Hwang, Youngbae
    Choi, Jun Won
    [J]. COMPUTER VISION - ACCV 2018, PT IV, 2019, 11364 : 90 - 106
  • [10] Multi-Modal Fusion Emotion Recognition Method of Speech Expression Based on Deep Learning
    Liu, Dong
    Wang, Zhiyong
    Wang, Lifeng
    Chen, Longxi
    [J]. FRONTIERS IN NEUROROBOTICS, 2021, 15