Multimodal Sentiment Analysis Representations Learning via Contrastive Learning with Condense Attention Fusion

被引:18
|
作者
Wang, Huiru [1 ]
Li, Xiuhong [1 ]
Ren, Zenyu [2 ]
Wang, Min [2 ]
Ma, Chunming [1 ]
机构
[1] Xinjiang Univ, Coll Informat Sci & Engn, Xinjiang Key Lab Signal Detect & Proc, Urumqi 830046, Peoples R China
[2] Xinjiang Univ, Coll Informat Sci & Engn, Urumqi 830046, Peoples R China
基金
中国国家自然科学基金;
关键词
multimodal; multimodal sentiment analysis; supervised contrastive learning; MLFC; SCSupCon; CLASSIFICATION;
D O I
10.3390/s23052679
中图分类号
O65 [分析化学];
学科分类号
070302 ; 081704 ;
摘要
Multimodal sentiment analysis has gained popularity as a research field for its ability to predict users' emotional tendencies more comprehensively. The data fusion module is a critical component of multimodal sentiment analysis, as it allows for integrating information from multiple modalities. However, it is challenging to combine modalities and remove redundant information effectively. In our research, we address these challenges by proposing a multimodal sentiment analysis model based on supervised contrastive learning, which leads to more effective data representation and richer multimodal features. Specifically, we introduce the MLFC module, which utilizes a convolutional neural network (CNN) and Transformer to solve the redundancy problem of each modal feature and reduce irrelevant information. Moreover, our model employs supervised contrastive learning to enhance its ability to learn standard sentiment features from data. We evaluate our model on three widely-used datasets, namely MVSA-single, MVSA-multiple, and HFM, demonstrating that our model outperforms the state-of-the-art model. Finally, we conduct ablation experiments to validate the efficacy of our proposed method.
引用
收藏
页数:15
相关论文
共 50 条
  • [1] Multimodal Sentiment Analysis with Preferential Fusion and Distance-aware Contrastive Learning
    Ma, Feipeng
    Zhang, Yueyi
    Sun, Xiaoyan
    2023 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO, ICME, 2023, : 1367 - 1372
  • [2] Multimodal Sentiment Analysis of Government Information Comments Based on Contrastive Learning and Cross-Attention Fusion Networks
    Mu, Guangyu
    Chen, Chuanzhi
    Li, Xiurong
    Li, Jiaxue
    Ju, Xiaoqing
    Dai, Jiaxiu
    IEEE ACCESS, 2024, 12 : 165525 - 165538
  • [3] Multimodal hypergraph network with contrastive learning for sentiment analysis
    Huang, Jian
    Jiang, Kun
    Pu, Yuanyuan
    Zhao, Zhengpeng
    Yang, Qiuxia
    Gu, Jinjing
    Xu, Dan
    NEUROCOMPUTING, 2025, 627
  • [4] Dynamic Weighted Multitask Learning and Contrastive Learning for Multimodal Sentiment Analysis
    Wang, Xingqi
    Zhang, Mengrui
    Chen, Bin
    Wei, Dan
    Shao, Yanli
    ELECTRONICS, 2023, 12 (13)
  • [5] Text-Centric Multimodal Contrastive Learning for Sentiment Analysis
    Peng, Heng
    Gu, Xue
    Li, Jian
    Wang, Zhaodan
    Xu, Hao
    ELECTRONICS, 2024, 13 (06)
  • [6] Tri-CLT: Learning Tri-Modal Representations with Contrastive Learning and Transformer for Multimodal Sentiment Recognition
    Yang, Zhiyong
    Li, Zijian
    Zhu, Dongdong
    Zhou, Yu
    INFORMATION TECHNOLOGY AND CONTROL, 2024, 53 (01): : 206 - 219
  • [7] Transformer-based adaptive contrastive learning for multimodal sentiment analysis
    Hu Y.
    Huang X.
    Wang X.
    Lin H.
    Zhang R.
    Multimedia Tools and Applications, 2025, 84 (3) : 1385 - 1402
  • [8] WConF: Weighted Contrastive Fusion for Multimodal Sentiment Analysis
    Zeng, Biqing
    Li, Ruiyuan
    Lu, Liuxing
    Lu, Liangqi
    Wang, Jiazhen
    Chen, Weihai
    Dong, Huimin
    NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING, PT V, NLPCC 2024, 2025, 15363 : 30 - 42
  • [9] Dynamically Shifting Multimodal Representations via Hybrid-Modal Attention for Multimodal Sentiment Analysis
    Lin, Ronghao
    Hu, Haifeng
    IEEE TRANSACTIONS ON MULTIMEDIA, 2024, 26 : 2740 - 2755
  • [10] Attention fusion network for multimodal sentiment analysis
    Yuanyi Luo
    Rui Wu
    Jiafeng Liu
    Xianglong Tang
    Multimedia Tools and Applications, 2024, 83 : 8207 - 8217