Multimodal Fusion Method Based on Self-Attention Mechanism

被引:11
|
作者
Zhu, Hu [1 ]
Wang, Ze [2 ]
Shi, Yu [3 ]
Hua, Yingying [1 ]
Xu, Guoxia [4 ]
Deng, Lizhen [5 ]
机构
[1] Nanjing Univ Posts & Telecommun, Jiangsu Prov Key Lab Image Proc & Image Commun, Nanjing 210003, Peoples R China
[2] China Acad Launch Vehicle Technol, R&D Ctr, Beijing 100176, Peoples R China
[3] Nanjing Univ Posts & Telecommun, Bell Honors Sch, Nanjing 210003, Peoples R China
[4] Norwegian Univ Sci & Technol, Dept Comp Sci, N-2815 Gjovik, Norway
[5] Nanjing Univ Posts & Telecommun, Natl Engn Res Ctr Commun & Network Technol, Nanjing 210003, Peoples R China
来源
WIRELESS COMMUNICATIONS & MOBILE COMPUTING | 2020年 / 2020卷 / 2020期
基金
中国国家自然科学基金;
关键词
D O I
10.1155/2020/8843186
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Multimodal fusion is one of the popular research directions of multimodal research, and it is also an emerging research field of artificial intelligence. Multimodal fusion is aimed at taking advantage of the complementarity of heterogeneous data and providing reliable classification for the model. Multimodal data fusion is to transform data from multiple single-mode representations to a compact multimodal representation. In previous multimodal data fusion studies, most of the research in this field used multimodal representations of tensors. As the input is converted into a tensor, the dimensions and computational complexity increase exponentially. In this paper, we propose a low-rank tensor multimodal fusion method with an attention mechanism, which improves efficiency and reduces computational complexity. We evaluate our model through three multimodal fusion tasks, which are based on a public data set: CMU-MOSI, IEMOCAP, and POM. Our model achieves a good performance while flexibly capturing the global and local connections. Compared with other multimodal fusions represented by tensors, experiments show that our model can achieve better results steadily under a series of attention mechanisms.
引用
收藏
页数:8
相关论文
共 50 条
  • [1] Multi-Sensor Data Fusion Method Based on Self-Attention Mechanism
    Lin, Xuezhu
    Chao, Shihan
    Yan, Dongming
    Guo, Lili
    Liu, Yue
    Li, Lijuan
    [J]. APPLIED SCIENCES-BASEL, 2023, 13 (21):
  • [2] The Multimodal Scene Recognition Method Based on Self-Attention and Distillation
    Sun, Ning
    Xu, Wei
    Liu, Jixin
    Chai, Lei
    Sun, Haian
    [J]. IEEE Multimedia, 2024, 31 (04) : 25 - 36
  • [3] Electrocardiogram signal classification based on fusion method of residual network and self-attention mechanism
    Yuan C.
    Liu Z.
    Wang C.
    Yang F.
    [J]. Shengwu Yixue Gongchengxue Zazhi/Journal of Biomedical Engineering, 2023, 40 (03): : 474 - 481
  • [4] Advancing classroom fatigue recognition: A multimodal fusion approach using self-attention mechanism
    Cao, Lei
    Wang, Wenrong
    Dong, Yilin
    Fan, Chunjiang
    [J]. BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2024, 89
  • [5] SFusion: Self-attention Based N-to-One Multimodal Fusion Block
    Liu, Zecheng
    Wei, Jia
    Li, Rui
    Zhou, Jianlong
    [J]. MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION, MICCAI 2023, PT II, 2023, 14221 : 159 - 169
  • [6] Attributed network embedding based on self-attention mechanism for recommendation method
    Wang, Shuo
    Yang, Jing
    Shang, Fanshu
    [J]. SCIENTIFIC REPORTS, 2023, 13 (01)
  • [7] Crop Diseases Recognition Method via Fusion Color Mask and Self-attention Mechanism
    Yu, Ming
    Li, Ruoxi
    Yan, Gang
    Wang, Yan
    Wang, Jianchun
    Li, Yang
    [J]. Nongye Jixie Xuebao/Transactions of the Chinese Society for Agricultural Machinery, 2022, 53 (08): : 337 - 344
  • [8] Attributed network embedding based on self-attention mechanism for recommendation method
    Shuo Wang
    Jing Yang
    Fanshu Shang
    [J]. Scientific Reports, 13
  • [9] Unsupervised Pansharpening Based on Self-Attention Mechanism
    Qu, Ying
    Baghbaderani, Razieh Kaviani
    Qi, Hairong
    Kwan, Chiman
    [J]. IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2021, 59 (04): : 3192 - 3208
  • [10] Keyphrase Generation Based on Self-Attention Mechanism
    Yang, Kehua
    Wang, Yaodong
    Zhang, Wei
    Yao, Jiqing
    Le, Yuquan
    [J]. CMC-COMPUTERS MATERIALS & CONTINUA, 2019, 61 (02): : 569 - 581