SSLMM: Semi-Supervised Learning with Missing Modalities for Multimodal Sentiment Analysis

被引:0
|
作者
Wang, Yiyu [1 ]
Jian, Haifang [2 ,3 ]
Zhuang, Jian [4 ]
Guo, Huimin [2 ,3 ]
Leng, Yan [1 ]
机构
[1] Shandong Normal Univ, Sch Phys & Elect, Jinan 250358, Peoples R China
[2] Chinese Acad Sci, Lab Solid State Optoelect Informat Technol, Inst Semicond, Beijing 100083, Peoples R China
[3] Chinese Acad Sci, Beijing 100049, Peoples R China
[4] Dalian Univ Technol, Sch Comp Sci & Technol, Dalian 116023, Liaoning, Peoples R China
关键词
Multimodal sentiment analysis; Semi-supervised learning; Missing modalities;
D O I
10.1016/j.inffus.2025.103058
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Multimodal Sentiment Analysis (MSA) integrates information from text, audio, and visuals to understand human emotions, but real-world applications face two challenges: (1) expensive annotation costs reduce the effectiveness of fully supervised methods, and (2) missing modality severely impact model robustness. While there are studies addressing these issues separately, few focus on solving both within a single framework. In real-world scenarios, these challenges often occur together, necessitating an algorithm that can handle both. To address this, we propose a Semi-Supervised Learning with Missing Modalities (SSLMM) framework. SSLMM combines self-supervised learning, alternating interaction information, semi-supervised learning, and modality reconstruction to tackle label scarcity and modality missing simultaneously. Firstly, SSLMM captures latent structural information through self-supervised pre-training. It then fine-tunes the model using semi- supervised learning and modality reconstruction to reduce dependence on labeled data and improve robustness to modality missing. The framework uses a graph-based architecture with an iterative message propagation mechanism to alternately propagate intra-modal and inter-modal messages, capturing emotional associations within and across modalities. Experiments on CMU-MOSI, CMU-MOSEI, and CH-SIMS demonstrate that under the condition where the proportion of labeled samples and the missing modality rate are both 0.5, SSLMM achieves binary classification (negative vs. positive) accuracies of 80.2%, 81.7%, and 77.1%, respectively, surpassing existing methods.
引用
收藏
页数:19
相关论文
共 50 条
  • [1] Multimodal Consistency-Based Teacher for Semi-Supervised Multimodal Sentiment Analysis
    Yuan, Ziqi
    Fang, Jingliang
    Xu, Hua
    Gao, Kai
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2024, 32 : 3669 - 3683
  • [2] Multimodal Prompt Learning with Missing Modalities for Sentiment Analysis and Emotion Recognition
    Guo, Zirun
    Jin, Tao
    Zhao, Zhou
    PROCEEDINGS OF THE 62ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1: LONG PAPERS, 2024, : 1726 - 1736
  • [3] Semi-supervised Learning with Multimodal Perturbation
    Su, Lei
    Liao, Hongzhi
    Yu, Zhengtao
    Tang, Jiahua
    ADVANCES IN NEURAL NETWORKS - ISNN 2009, PT 1, PROCEEDINGS, 2009, 5551 : 651 - +
  • [4] Semi-Supervised Learning for Aspect-Based Sentiment Analysis
    Zheng, Hang
    Zhang, Jianhui
    Suzuki, Yoshimi
    Fukumoto, Fumiyo
    Nishizaki, Hiromitsu
    2021 INTERNATIONAL CONFERENCE ON CYBERWORLDS (CW 2021), 2021, : 209 - 212
  • [5] SemiMemes: A Semi-supervised Learning Approach for Multimodal Memes Analysis
    Pham Thai Hoang Tung
    Nguyen Tan Viet
    Ngo Tien Anh
    Phan Duy Hung
    COMPUTATIONAL COLLECTIVE INTELLIGENCE, ICCCI 2023, 2023, 14162 : 565 - 577
  • [6] Semi-supervised learning with missing values imputation
    Huang, Buliao
    Zhu, Yunhui
    Usman, Muhammad
    Chen, Huanhuan
    KNOWLEDGE-BASED SYSTEMS, 2024, 284
  • [7] Distributed Semi-Supervised Learning With Missing Data
    Xu, Zhen
    Liu, Ying
    Li, Chunguang
    IEEE TRANSACTIONS ON CYBERNETICS, 2021, 51 (12) : 6165 - 6178
  • [8] Multimodal semi-supervised learning for image classification
    Guillaumin, Matthieu
    Verbeek, Jakob
    Schmid, Cordelia
    2010 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2010, : 902 - 909
  • [9] SEMI-SUPERVISED MULTIMODAL IMAGE TRANSLATION FOR MISSING MODALITY IMPUTATION
    Sun, Wangbin
    Ma, Fei
    Li, Yang
    Huang, Shao-Lun
    Ni, Shiguang
    Zhang, Lin
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 4320 - 4324
  • [10] Sentiment analysis using semi-supervised learning with few labeled data
    Pan, Yuhao
    Chen, Zhiqun
    Suzuki, Yoshimi
    Fukumoto, Fumiyo
    Nishizaki, Hiromitsu
    2020 INTERNATIONAL CONFERENCE ON CYBERWORLDS (CW 2020), 2020, : 231 - 234