Multi-modal sarcasm detection using ensemble net model

被引:0
|
作者
Sukhavasi, Vidyullatha [1 ,2 ]
Dondeti, Venkatesulu [3 ]
机构
[1] Vignans Fdn Sci Technol & Res, Dept CSE, Guntur 522213, Andhra Pradesh, India
[2] BVRIT HYDERABAD Coll Engn Women, Dept CSE, Hyderabad 500090, Telangana, India
[3] Vignans Fdn Sci Technol & Res, Dept Adv CSE, Guntur 522213, Andhra Pradesh, India
关键词
Sarcasm detection; Hybrid EnsembleNet; Weighted fusion modality; Softmax layer; Natural language processing; Deep learning approach;
D O I
10.1007/s10115-024-02227-y
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Generally, sarcasm is expressed via various verbal and non-verbal words. Various existing works on the detection of sarcasm have been performed in either text or video. With the rapid growth of social media and internet technology, people express their emotions and feelings using text. Therefore, a multi-modal sarcasm detection task is crucial to understanding people's real feelings and beliefs. However, it is still a challenge to detect sarcasm from multi-modal features. Therefore, this work presents a new hybrid ensemble deep learning approach for multi-modal sarcasm detection. The major goal of this research is to determine the different classes of sarcasm using a multi-modal dataset. Here, imaging modality-based sarcasm detection is performed using Deep Residual Net, and the visual features are extracted. For the generation of text modality, the text data are pre-processed with punctuation removal, and the textual features are extracted using Term Frequency-Inverse Average Document Frequency. The extracted features are used as input for the bidirectional long short-term memory model. The audio (acoustic) elements are extracted to form acoustic modality, which is subsequently sent to the visual geometry group. Furthermore, the weighted fusion modality process is used to combine all of the collected features. The softmax layer acts as the classification layer for performing multi-modal sarcasm detection. Here, the Tent chaotic snack optimization algorithm is employed to tune the hyperparameter and reduce the complexity of the proposed Hybrid EnsembleNet. PYTHON tool is used to evaluate the performance of the proposed classifier. The proposed hybrid EnsembleNet is trained using two datasets: Memotion 7k and MUStARD.
引用
收藏
页数:23
相关论文
共 50 条
  • [1] Multi-Modal Sarcasm Detection in Twitter with Hierarchical Fusion Model
    Cai, Yitao
    Cai, Huiyu
    Wan, Xiaojun
    [J]. 57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), 2019, : 2506 - 2515
  • [2] Multi-Modal Sarcasm Detection with Sentiment Word Embedding
    Fu, Hao
    Liu, Hao
    Wang, Hongling
    Xu, Linyan
    Lin, Jiali
    Jiang, Dazhi
    [J]. ELECTRONICS, 2024, 13 (05)
  • [3] Multi-modal sarcasm detection based on Multi-Channel Enhanced Fusion model
    Fang, Hong
    Liang, Dahao
    Xiang, Weiyu
    [J]. NEUROCOMPUTING, 2024, 578
  • [4] Multi-Modal Sarcasm Detection with Interactive In-Modal and Cross-Modal Graphs
    Liang, Bin
    Lou, Chenwei
    Li, Xiang
    Gui, Lin
    Yang, Min
    Xu, Ruifeng
    [J]. PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2021, 2021, : 4707 - 4715
  • [5] Attention-based multi-modal fusion sarcasm detection
    Liu, Jing
    Tian, Shengwei
    Yu, Long
    Long, Jun
    Zhou, Tiejun
    Wang, Bo
    [J]. JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2023, 44 (02) : 2097 - 2108
  • [6] Cross-modal incongruity aligning and collaborating for multi-modal sarcasm detection
    Wang, Jie
    Yang, Yan
    Jiang, Yongquan
    Ma, Minbo
    Xie, Zhuyang
    Li, Tianrui
    [J]. INFORMATION FUSION, 2024, 103
  • [7] Multi-Modal Sarcasm Detection via Cross-Modal Graph Convolutional Network
    Liang, Bin
    Lou, Chenwei
    Li, Xiang
    Yang, Min
    Gui, Lin
    He, Yulan
    Pei, Wenjie
    Xu, Ruifeng
    [J]. PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 1767 - 1777
  • [8] Multi-Modal Sarcasm Detection Based on Cross-Modal Composition of Inscribed Entity Relations
    Li, Lingshan
    Jin, Di
    Wang, Xiaobao
    Guo, Fengyu
    Wang, Longbiao
    Dang, Jianwu
    [J]. 2023 IEEE 35TH INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE, ICTAI, 2023, : 918 - 925
  • [9] Single-Stage Extensive Semantic Fusion for multi-modal sarcasm detection
    Fang, Hong
    Liang, Dahao
    Xiang, Weiyu
    [J]. ARRAY, 2024, 22
  • [10] Multi-Modal Sarcasm Detection and Humor Classification in Code-Mixed Conversations
    Bedi, Manjot
    Kumar, Shivani
    Akhtar, Md Shad
    Chakraborty, Tanmoy
    [J]. IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2023, 14 (02) : 1363 - 1375