Multi-modal sarcasm detection using ensemble net model

被引:0
|
作者
Sukhavasi, Vidyullatha [1 ,2 ]
Dondeti, Venkatesulu [3 ]
机构
[1] Vignans Fdn Sci Technol & Res, Dept CSE, Guntur 522213, Andhra Pradesh, India
[2] BVRIT HYDERABAD Coll Engn Women, Dept CSE, Hyderabad 500090, Telangana, India
[3] Vignans Fdn Sci Technol & Res, Dept Adv CSE, Guntur 522213, Andhra Pradesh, India
关键词
Sarcasm detection; Hybrid EnsembleNet; Weighted fusion modality; Softmax layer; Natural language processing; Deep learning approach;
D O I
10.1007/s10115-024-02227-y
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Generally, sarcasm is expressed via various verbal and non-verbal words. Various existing works on the detection of sarcasm have been performed in either text or video. With the rapid growth of social media and internet technology, people express their emotions and feelings using text. Therefore, a multi-modal sarcasm detection task is crucial to understanding people's real feelings and beliefs. However, it is still a challenge to detect sarcasm from multi-modal features. Therefore, this work presents a new hybrid ensemble deep learning approach for multi-modal sarcasm detection. The major goal of this research is to determine the different classes of sarcasm using a multi-modal dataset. Here, imaging modality-based sarcasm detection is performed using Deep Residual Net, and the visual features are extracted. For the generation of text modality, the text data are pre-processed with punctuation removal, and the textual features are extracted using Term Frequency-Inverse Average Document Frequency. The extracted features are used as input for the bidirectional long short-term memory model. The audio (acoustic) elements are extracted to form acoustic modality, which is subsequently sent to the visual geometry group. Furthermore, the weighted fusion modality process is used to combine all of the collected features. The softmax layer acts as the classification layer for performing multi-modal sarcasm detection. Here, the Tent chaotic snack optimization algorithm is employed to tune the hyperparameter and reduce the complexity of the proposed Hybrid EnsembleNet. PYTHON tool is used to evaluate the performance of the proposed classifier. The proposed hybrid EnsembleNet is trained using two datasets: Memotion 7k and MUStARD.
引用
收藏
页数:23
相关论文
共 50 条
  • [41] Multi-modal novelty and familiarity detection
    Christo Panchev
    [J]. BMC Neuroscience, 14 (Suppl 1)
  • [42] Multi-Modal Depression Detection and Estimation
    Yang, Le
    [J]. 2019 8TH INTERNATIONAL CONFERENCE ON AFFECTIVE COMPUTING AND INTELLIGENT INTERACTION WORKSHOPS AND DEMOS (ACIIW), 2019, : 26 - 30
  • [43] Multi-Modal Anomaly Detection by Using Audio and Visual Cues
    Rehman, Ata-Ur
    Ullah, Hafiz Sami
    Farooq, Haroon
    Khan, Muhammad Salman
    Mahmood, Tayyeb
    Khan, Hafiz Owais Ahmed
    [J]. IEEE ACCESS, 2021, 9 : 30587 - 30603
  • [44] Multi-modal Hate Speech Detection using Machine Learning
    Boishakhi, Fariha Tahosin
    Shill, Ponkoj Chandra
    Alam, Md Golam Rabiul
    [J]. 2021 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2021, : 4496 - 4499
  • [45] Sarcasm detection using deep learning and ensemble learning
    Priya Goel
    Rachna Jain
    Anand Nayyar
    Shruti Singhal
    Muskan Srivastava
    [J]. Multimedia Tools and Applications, 2022, 81 : 43229 - 43252
  • [46] Sarcasm detection using deep learning and ensemble learning
    Goel, Priya
    Jain, Rachna
    Nayyar, Anand
    Singhal, Shruti
    Srivastava, Muskan
    [J]. MULTIMEDIA TOOLS AND APPLICATIONS, 2022, 81 (30) : 43229 - 43252
  • [47] What Does Your Smile Mean? Jointly Detecting Multi-Modal Sarcasm and Sentiment Using Quantum Probability
    Liu, Yaochen
    Zhang, Yazhou
    Li, Qiuchi
    Wang, Benyou
    Song, Dawei
    [J]. FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2021, 2021, : 871 - 880
  • [48] Multi-Modal Pedestrian Detection with Large Misalignment Based on Modal-Wise Regression and Multi-Modal IoU
    Wanchaitanawong, Napat
    Tanaka, Masayuki
    Shibata, Takashi
    Okutomi, Masatoshi
    [J]. PROCEEDINGS OF 17TH INTERNATIONAL CONFERENCE ON MACHINE VISION APPLICATIONS (MVA 2021), 2021,
  • [49] DeepMEF: A Deep Model Ensemble Framework for Video Based Multi-modal Person Identification
    Dong, Chuanqi
    Gu, Zheng
    Huang, Zhonghao
    Ji, Wen
    Huo, Jing
    Gao, Yang
    [J]. PROCEEDINGS OF THE 27TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA (MM'19), 2019, : 2531 - 2534
  • [50] Multi-Modal Data Analysis for Alzheimer's Disease Diagnosis: An Ensemble Model Using Imagery and Genetic Features
    Ying, Qi
    Xing, Xin
    Liu, Liangliang
    Lin, Ai-Ling
    Jacobs, Nathan
    Liang, Gongbo
    [J]. 2021 43RD ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE & BIOLOGY SOCIETY (EMBC), 2021, : 3586 - 3591