Combining Knowledge and Multi-modal Fusion for Meme Classification

被引:4
|
作者
Zhong, Qi [1 ]
Wang, Qian [1 ]
Liu, Ji [1 ]
机构
[1] Chongqing Univ, Coll Comp Sci, Chongqing 400044, Peoples R China
来源
关键词
Meme classification; Multi-modal fusion; Self-attention mechanism;
D O I
10.1007/978-3-030-98358-1_47
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Internet memes are widespread on social media platforms such as Twitter and Facebook. Recently, meme classification has been an active research topic, especially meme sentiment classification and meme offensive classification. Internet memes contain multi-modal information, and the meme text is embedded in the meme image. The existing methods classify memes by simply concatenating global visual and textual features to generate a multi-modal representation. However, these approaches ignored the noise introduced by global visual features and the potential common information of meme multi-modal representation. In this paper, we propose a model for meme classification named MeBERT. Our method enhances the semantic representation of the meme by introducing conceptual information through external Knowledge Bases (KBs). Then, to reduce noise, a concept-image attention module is designed to extract concept-sensitive visual representation. In addition, a deep convolution tensor fusion module is built to effectively integrate multi-modal information. To verify the effectiveness of the model in the tasks of meme sentiment classification and meme offensive classification, we designed experiments on the Memotion and MultiOFF datasets. The experimental results show that the MeBERT model achieves better performance than state-of-the-art techniques for meme classification.
引用
收藏
页码:599 / 611
页数:13
相关论文
共 50 条
  • [1] Improved Sentiment Classification by Multi-modal Fusion
    Gan, Lige
    Benlamri, Rachid
    Khoury, Richard
    [J]. 2017 THIRD IEEE INTERNATIONAL CONFERENCE ON BIG DATA COMPUTING SERVICE AND APPLICATIONS (IEEE BIGDATASERVICE 2017), 2017, : 11 - 16
  • [2] Research on multi-modal hateful meme detection
    Li Wanbo
    Liu Suying
    [J]. PROCEEDINGS OF 2021 2ND INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND INFORMATION SYSTEMS (ICAIIS '21), 2021,
  • [3] Research on Emotion Classification Based on Multi-modal Fusion
    Xiang, Zhihua
    Radzi, Nor Haizan Mohamed
    Hashim, Haslina
    [J]. BAGHDAD SCIENCE JOURNAL, 2024, 21 (02) : 548 - 560
  • [4] Image and Encoded Text Fusion for Multi-Modal Classification
    Gallo, I.
    Calefati, A.
    Nawaz, S.
    Janjua, M. K.
    [J]. 2018 INTERNATIONAL CONFERENCE ON DIGITAL IMAGE COMPUTING: TECHNIQUES AND APPLICATIONS (DICTA), 2018, : 203 - 209
  • [5] MULTI-MODAL INFORMATION FUSION FOR CLASSIFICATION OF KIDNEY ABNORMALITIES
    Varsha, S.
    Nasser, Sahar Almahfouz
    Bala, Gouranga
    Kurian, Nikhil Cherian
    Sethi, Amit
    [J]. 2022 IEEE INTERNATIONAL SYMPOSIUM ON BIOMEDICAL IMAGING CHALLENGES (IEEE ISBI 2022), 2022,
  • [6] Multi-modal Fusion
    Liu, Huaping
    Hussain, Amir
    Wang, Shuliang
    [J]. INFORMATION SCIENCES, 2018, 432 : 462 - 462
  • [7] News video classification based on multi-modal information fusion
    Lie, WN
    Su, CK
    [J]. 2005 INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), VOLS 1-5, 2005, : 1021 - 1024
  • [8] Disease Classification Model Based on Multi-Modal Feature Fusion
    Wan, Zhengyu
    Shao, Xinhui
    [J]. IEEE ACCESS, 2023, 11 : 27536 - 27545
  • [9] Multi-Modal Military Event Extraction Based on Knowledge Fusion
    Xiang, Yuyuan
    Jia, Yangli
    Zhang, Xiangliang
    Zhang, Zhenling
    [J]. CMC-COMPUTERS MATERIALS & CONTINUA, 2023, 77 (01): : 97 - 114
  • [10] Incomplete multi-modal brain image fusion for epilepsy classification
    Zhu, Qi
    Li, Huijie
    Ye, Haizhou
    Zhang, Zhiqiang
    Wang, Ran
    Fan, Zizhu
    Zhang, Daoqiang
    [J]. INFORMATION SCIENCES, 2022, 582 : 316 - 333