Multimodal fusion recognition for digital twin

被引:4
|
作者
Zhou, Tianzhe [1 ]
Zhang, Xuguang [1 ]
Kang, Bing [1 ]
Chen, Mingkai [1 ]
机构
[1] Nanjing Univ Posts & Telecommun, Key Lab Broadband Wireless Commun & Sensor Network, Minist Educ, Nanjing 210003, Peoples R China
关键词
Digital twin; Multimodal fusion; Object recognition; Deep learning; Transfer learning; CLASSIFICATION; NETWORKS; FEATURES;
D O I
10.1016/j.dcan.2022.10.009
中图分类号
TN [电子技术、通信技术];
学科分类号
0809 ;
摘要
The digital twin is the concept of transcending reality, which is the reverse feedback from the real physical space to the virtual digital space. People hold great prospects for this emerging technology. In order to realize the upgrading of the digital twin industrial chain, it is urgent to introduce more modalities, such as vision, haptics, hearing and smell, into the virtual digital space, which assists physical entities and virtual objects in creating a closer connection. Therefore, perceptual understanding and object recognition have become an urgent hot topic in the digital twin. Existing surface material classification schemes often achieve recognition through machine learning or deep learning in a single modality, ignoring the complementarity between multiple modalities. In order to overcome this dilemma, we propose a multimodal fusion network in our article that combines two modalities, visual and haptic, for surface material recognition. On the one hand, the network makes full use of the potential correlations between multiple modalities to deeply mine the modal semantics and complete the data mapping. On the other hand, the network is extensible and can be used as a universal architecture to include more modalities. Experiments show that the constructed multimodal fusion network can achieve 99.42% classification accuracy while reducing complexity.
引用
收藏
页码:337 / 346
页数:10
相关论文
共 50 条
  • [31] The Recognition of Teacher Behavior Based on Multimodal Information Fusion
    Wu, Dongli
    Chen, Jia
    Deng, Wei
    Wei, Yantao
    Luo, Heng
    Wei, Yangyu
    MATHEMATICAL PROBLEMS IN ENGINEERING, 2020, 2020 (2020)
  • [32] Research on a Microexpression Recognition Technology Based on Multimodal Fusion
    Kang, Jie
    Chen, Xiao Ying
    Liu, Qi Yuan
    Jin, Si Han
    Yang, Cheng Han
    Hu, Cong
    COMPLEXITY, 2021, 2021
  • [33] Fusion of Facial Expressions and EEG for Multimodal Emotion Recognition
    Huang, Yongrui
    Yang, Jianhao
    Liao, Pengkai
    Pan, Jiahui
    COMPUTATIONAL INTELLIGENCE AND NEUROSCIENCE, 2017, 2017
  • [34] Multimodal Physiological Signals Fusion for Online Emotion Recognition
    Pan, Tongjie
    Ye, Yalan
    Cai, Hecheng
    Huang, Shudong
    Yang, Yang
    Wang, Guoqing
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 5879 - 5888
  • [35] Review on Multimodal Fusion Techniques for Human Emotion Recognition
    Karani, Ruhina
    Desai, Sharmishta
    INTERNATIONAL JOURNAL OF ADVANCED COMPUTER SCIENCE AND APPLICATIONS, 2022, 13 (10) : 287 - 296
  • [36] Data fusion for driver drowsiness recognition: A multimodal perspective
    Priyanka, S.
    Shanthi, S.
    Kumar, A. Saran
    Praveen, V.
    EGYPTIAN INFORMATICS JOURNAL, 2024, 27
  • [37] Multimodal emotion recognition for the fusion of speech and EEG signals
    Ma J.
    Sun Y.
    Zhang X.
    Xi'an Dianzi Keji Daxue Xuebao/Journal of Xidian University, 2019, 46 (01): : 143 - 150
  • [38] Multimodal information fusion based human movement recognition
    Yao Shu
    Heng Zhang
    Multimedia Tools and Applications, 2020, 79 : 5043 - 5052
  • [39] Context-aware Multimodal Fusion for Emotion Recognition
    Li, Jinchao
    Wang, Shuai
    Chao, Yang
    Liu, Xunying
    Meng, Helen
    INTERSPEECH 2022, 2022, : 2013 - 2017
  • [40] Dynamic Gesture Recognition Based On Multimodal Fusion Model
    Fang, Juan
    Xu, Chao
    Wang, Chao
    Li, Hua
    20TH INT CONF ON UBIQUITOUS COMP AND COMMUNICAT (IUCC) / 20TH INT CONF ON COMP AND INFORMATION TECHNOLOGY (CIT) / 4TH INT CONF ON DATA SCIENCE AND COMPUTATIONAL INTELLIGENCE (DSCI) / 11TH INT CONF ON SMART COMPUTING, NETWORKING, AND SERV (SMARTCNS), 2021, : 172 - 177