ConOffense: Multi-modal multitask Contrastive learning for offensive content identification

被引:0
|
作者
Shome, Debaditya [1 ]
Kar, T. [1 ]
机构
[1] KIIT Univ, Sch Elect Engn, Bhubaneswar, Odisha, India
关键词
Multimodal learning; Contrastive learning; Representation learning; Social media; Offensive content identification;
D O I
10.1109/BigData52589.2021.9671427
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Hateful or offensive content has been increasingly common on social media platforms in recent years, and the problem is now widespread. There is a pressing need for effective automatic solutions for detecting such content, especially due to the gigantic size of social media data. Although significant progress has been made in the automated identification of offensive content, most of the focus has been on only using textual information. It can be easily noticed that with the rise in visual information shared on these platforms, it is quite common to have hateful content on images rather than in the associated text. Due to this, present day unimodal text-based methods won't be able to cope up with the multimodal hateful content. In this paper, we propose a novel multimodal neural network powered by contrastive learning for identifying offensive posts on social media utilizing both visual and textual information. We design the text and visual encoders with a lightweight architecture to make the solution efficient for real world use. Evaluation on the MMHS150K dataset shows state-of-the-art performance of 82.6 percent test accuracy, making an improvement of approximately +14.1 percent accuracy over the previous best performing benchmark model on the dataset.
引用
收藏
页码:4524 / 4529
页数:6
相关论文
共 50 条
  • [41] The Performance of a Temporal Multi-Modal Sentiment Analysis Model Based on Multitask Learning in Social Networks
    He, Lin
    Lu, Haili
    INTERNATIONAL JOURNAL OF ADVANCED COMPUTER SCIENCE AND APPLICATIONS, 2024, 15 (05) : 1109 - 1117
  • [42] Unsupervised Multi-modal Learning
    Iqbal, Mohammed Shameer
    ADVANCES IN ARTIFICIAL INTELLIGENCE (AI 2015), 2015, 9091 : 343 - 346
  • [43] Learning Multi-modal Similarity
    McFee, Brian
    Lanckriet, Gert
    JOURNAL OF MACHINE LEARNING RESEARCH, 2011, 12 : 491 - 523
  • [44] Adaptive Remediation with Multi-modal Content
    Tu, Yuwei
    Brinton, Christopher G.
    Lan, Andrew S.
    Chiang, Mung
    ADAPTIVE INSTRUCTIONAL SYSTEMS, AIS 2019, 2019, 11597 : 455 - 468
  • [45] Multitask Collaborative Multi-modal Remote Sensing Target Segmentation Algorithm
    Mao, Xiuhua
    Zhang, Qiang
    Ruan, Hang
    Yang, Yuang
    Dianzi Yu Xinxi Xuebao/Journal of Electronics and Information Technology, 2024, 46 (08): : 3363 - 3371
  • [46] A Multi-Modal Contrastive Diffusion Model for Therapeutic Peptide Generation
    Wang, Yongkang
    Liu, Xuan
    Huang, Feng
    Xiong, Zhankun
    Zhang, Wen
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 1, 2024, : 3 - 11
  • [47] Multi-modal graph contrastive encoding for neural machine translation
    Yin, Yongjing
    Zeng, Jiali
    Su, Jinsong
    Zhou, Chulun
    Meng, Fandong
    Zhou, Jie
    Huang, Degen
    Luo, Jiebo
    ARTIFICIAL INTELLIGENCE, 2023, 323
  • [48] CCGN: consistency contrastive-learning graph network for multi-modal fake news detection
    Cui, Shaodong
    Duan, Kaibo
    Ma, Wen
    Shinnou, Hiroyuki
    MULTIMEDIA SYSTEMS, 2025, 31 (02)
  • [49] Multi-modal human identification system
    Ivanov, Y
    WACV 2005: SEVENTH IEEE WORKSHOP ON APPLICATIONS OF COMPUTER VISION, PROCEEDINGS, 2005, : 164 - 170
  • [50] Towards Accurate and Robust Multi-Modal Medical Image Registration Using Contrastive Metric Learning
    Hu, Jinrong
    Sun, Shanhui
    Yang, Xiaodong
    Zhou, Shuang
    Wang, Xin
    Fu, Ying
    Zhou, Jiliu
    Yin, Youbing
    Cao, Kunlin
    Song, Qi
    Wu, Xi
    IEEE ACCESS, 2019, 7 : 132816 - 132827