Multimodal Fusion Method Based on Self-Attention Mechanism

被引:10
|
作者
Zhu, Hu [1 ]
Wang, Ze [2 ]
Shi, Yu [3 ]
Hua, Yingying [1 ]
Xu, Guoxia [4 ]
Deng, Lizhen [5 ]
机构
[1] Nanjing Univ Posts & Telecommun, Jiangsu Prov Key Lab Image Proc & Image Commun, Nanjing 210003, Peoples R China
[2] China Acad Launch Vehicle Technol, R&D Ctr, Beijing 100176, Peoples R China
[3] Nanjing Univ Posts & Telecommun, Bell Honors Sch, Nanjing 210003, Peoples R China
[4] Norwegian Univ Sci & Technol, Dept Comp Sci, N-2815 Gjovik, Norway
[5] Nanjing Univ Posts & Telecommun, Natl Engn Res Ctr Commun & Network Technol, Nanjing 210003, Peoples R China
基金
中国国家自然科学基金;
关键词
Computational complexity - Data fusion - Computational efficiency;
D O I
10.1155/2020/8843186
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Multimodal fusion is one of the popular research directions of multimodal research, and it is also an emerging research field of artificial intelligence. Multimodal fusion is aimed at taking advantage of the complementarity of heterogeneous data and providing reliable classification for the model. Multimodal data fusion is to transform data from multiple single-mode representations to a compact multimodal representation. In previous multimodal data fusion studies, most of the research in this field used multimodal representations of tensors. As the input is converted into a tensor, the dimensions and computational complexity increase exponentially. In this paper, we propose a low-rank tensor multimodal fusion method with an attention mechanism, which improves efficiency and reduces computational complexity. We evaluate our model through three multimodal fusion tasks, which are based on a public data set: CMU-MOSI, IEMOCAP, and POM. Our model achieves a good performance while flexibly capturing the global and local connections. Compared with other multimodal fusions represented by tensors, experiments show that our model can achieve better results steadily under a series of attention mechanisms.
引用
下载
收藏
页数:8
相关论文
共 50 条
  • [21] Web service classification based on self-attention mechanism
    Jia, Zhichun
    Zhang, Zhiying
    Dong, Rui
    Yang, Zhongxuan
    Xing, Xing
    2023 35TH CHINESE CONTROL AND DECISION CONFERENCE, CCDC, 2023, : 2164 - 2169
  • [22] Deepfake face discrimination based on self-attention mechanism
    Wang, Shuai
    Zhu, Donghui
    Chen, Jian
    Bi, Jiangbo
    Wang, Wenyi
    PATTERN RECOGNITION LETTERS, 2024, 183 : 92 - 97
  • [23] Progressive Scene Segmentation Based on Self-Attention Mechanism
    Pan, Yunyi
    Gan, Yuan
    Liu, Kun
    Zhang, Yan
    2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, : 3985 - 3992
  • [24] Neural network based on convolution and self-attention fusion mechanism for plant leaves disease recognition
    Zhao, Yun
    Li, Yang
    Wu, Na
    Xu, Xing
    CROP PROTECTION, 2024, 180
  • [25] AttendAffectNet-Emotion Prediction of Movie Viewers Using Multimodal Fusion with Self-Attention
    Thao, Ha Thi Phuong
    Balamurali, B. T.
    Roig, Gemma
    Herremans, Dorien
    SENSORS, 2021, 21 (24)
  • [26] Detection of Rice Pests Based on Self-Attention Mechanism and Multi-Scale Feature Fusion
    Hu, Yuqi
    Deng, Xiaoling
    Lan, Yubin
    Chen, Xin
    Long, Yongbing
    Liu, Cunjia
    INSECTS, 2023, 14 (03)
  • [27] Self-Attention Multilayer Feature Fusion Based on Long Connection
    Yuezhong, Chu
    Jiaqing, Wang
    Heng, Liu
    ADVANCES IN MULTIMEDIA, 2022, 2022
  • [28] Infrared and Visible Image Fusion Method via Interactive Self-attention
    Yang Fan
    Wang Zhishe
    Sun Jing
    Yu Zhaofa
    ACTA PHOTONICA SINICA, 2024, 53 (06)
  • [29] Recognition of piglet postures based on self-attention mechanism and anchor-free method
    Xu C.
    Xue Y.
    Zheng C.
    Hou W.
    Guo J.
    Wang X.
    Nongye Gongcheng Xuebao/Transactions of the Chinese Society of Agricultural Engineering, 2022, 38 (14): : 166 - 173
  • [30] A Coarse-to-Fine Facial Landmark Detection Method Based on Self-attention Mechanism
    Gao, Pengcheng
    Lu, Ke
    Xue, Jian
    Shao, Ling
    Lyu, Jiayi
    IEEE TRANSACTIONS ON MULTIMEDIA, 2021, 23 : 926 - 938