Multi-modal Sentiment and Emotion Joint Analysis with a Deep Attentive Multi-task Learning Model

被引:4
|
作者
Zhang, Yazhou [1 ]
Rong, Lu [1 ]
Li, Xiang [2 ]
Chen, Rui [1 ]
机构
[1] Zhengzhou Univ Light Ind, Software Engn Coll, Zhengzhou, Peoples R China
[2] Qilu Univ Technol, Shandong Comp Sci Ctr, Shandong Acad Sci, Natl Supercomp Ctr Jinan, Jinan, Peoples R China
来源
基金
美国国家科学基金会;
关键词
Multi-modal sentiment analysis; Emotion recognition; Multi-task learning; Deep learning;
D O I
10.1007/978-3-030-99736-6_35
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Emotion is seen as the external expression of sentiment, while sentiment is the essential nature of emotion. They are tightly entangled with each other in that one helps the understanding of the other, leading to a new research topic, i.e., multi-modal sentiment and emotion joint analysis. There exists two key challenges in this field, i.e., multi-modal fusion and multi-task interaction. Most of the recent approaches treat them as two independent tasks, and fail to model the relationships between them. In this paper, we propose a novel multi-modal multi-task learning model, termed MMT, to generically address such issues. Specially, two attention mechanisms, i.e., cross-modal and cross-task attentions are designed. Cross-modal attention is proposed to model multi-modal feature fusion, while cross-task attention is to capture the interaction between sentiment analysis and emotion recognition. Finally, we empirically show that this method alleviates such problems on two benchmarking datasets, while getting better performance for the main task, i.e., sentiment analysis with the help of the secondary emotion recognition task.
引用
收藏
页码:518 / 532
页数:15
相关论文
共 50 条
  • [41] Multi-modal multi-task learning for joint prediction of multiple regression and classification variables in Alzheimer's disease
    Zhang, Daoqiang
    Shen, Dinggang
    [J]. NEUROIMAGE, 2012, 59 (02) : 895 - 907
  • [42] Attentive Task Interaction Network for Multi-Task Learning
    Sinodinos, Dimitrios
    Armanfard, Narges
    [J]. 2022 26TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2022, : 2885 - 2891
  • [43] YuYin: a multi-task learning model of multi-modal e-commerce background music recommendation
    Ma, Le
    Wu, Xinda
    Tang, Ruiyuan
    Zhong, Chongjun
    Zhang, Kejun
    [J]. EURASIP JOURNAL ON AUDIO SPEECH AND MUSIC PROCESSING, 2023, 2023 (01)
  • [44] YuYin: a multi-task learning model of multi-modal e-commerce background music recommendation
    Le Ma
    Xinda Wu
    Ruiyuan Tang
    Chongjun Zhong
    Kejun Zhang
    [J]. EURASIP Journal on Audio, Speech, and Music Processing, 2023
  • [45] Joint Extraction Model for Entities and Events with Multi-task Deep Learning
    Chuanming, Yu
    Hongjun, Lin
    Zhengang, Zhang
    [J]. Data Analysis and Knowledge Discovery, 2022, 6 (2-3) : 117 - 128
  • [46] Multi-modal fusion and multi-task deep learning for monitoring the growth of film-mulched winter wheat
    Cheng, Zhikai
    Gu, Xiaobo
    Du, Yadan
    Wei, Chunyu
    Xu, Yang
    Zhou, Zhihui
    Li, Wenlong
    Cai, Wenjing
    [J]. PRECISION AGRICULTURE, 2024, 25 (04) : 1933 - 1957
  • [47] Improving sentiment analysis with multi-task learning of negation
    Barnes, Jeremy
    Velldal, Erik
    Ovrelid, Lilja
    [J]. NATURAL LANGUAGE ENGINEERING, 2021, 27 (02) : 249 - 269
  • [48] Multi-Task Model and Feature Joint Learning
    Li, Ya
    Tian, Xinmei
    Liu, Tongliang
    Tao, Dacheng
    [J]. PROCEEDINGS OF THE TWENTY-FOURTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE (IJCAI), 2015, : 3643 - 3649
  • [49] A multi-modal deep learning system for Arabic emotion recognition
    Abu Shaqra F.
    Duwairi R.
    Al-Ayyoub M.
    [J]. International Journal of Speech Technology, 2023, 26 (01) : 123 - 139
  • [50] Adversarial Multi-Task Learning for Mandarin Prosodic Boundary Prediction With Multi-Modal Embeddings
    Yi, Jiangyan
    Tao, Jianhua
    Fu, Ruibo
    Wang, Tao
    Zhang, Chu Yuan
    Wang, Chenglong
    [J]. IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2023, 31 : 2963 - 2973