Cluster-Level Contrastive Learning for Emotion Recognition in Conversations

被引:27
|
作者
Yang, Kailai [1 ,2 ]
Zhang, Tianlin [1 ,2 ]
Alhuzali, Hassan [3 ]
Ananiadou, Sophia [1 ,2 ]
机构
[1] Univ Manchester, NaCTeM, Manchester M13 9PL, England
[2] Univ Manchester, Dept Comp Sci, Manchester M13 9PL, England
[3] Umm Al Qura Univ, Coll Comp & Informat Syst, Mecca 24382, Saudi Arabia
基金
英国生物技术与生命科学研究理事会;
关键词
Emotion recognition; Prototypes; Linguistics; Task analysis; Semantics; Training; Adaptation models; Cluster-level contrastive learning; emotion recognition in conversations; pre-trained knowledge adapters; valence-arousal-dominance; DIALOGUE; FRAMEWORK;
D O I
10.1109/TAFFC.2023.3243463
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
A key challenge for Emotion Recognition in Conversations (ERC) is to distinguish semantically similar emotions. Some works utilise Supervised Contrastive Learning (SCL) which uses categorical emotion labels as supervision signals and contrasts in high-dimensional semantic space. However, categorical labels fail to provide quantitative information between emotions. ERC is also not equally dependent on all embedded features in the semantic space, which makes the high-dimensional SCL inefficient. To address these issues, we propose a novel low-dimensional Supervised Cluster-level Contrastive Learning (SCCL) method, which first reduces the high-dimensional SCL space to a three-dimensional affect representation space Valence-Arousal-Dominance (VAD), then performs cluster-level contrastive learning to incorporate measurable emotion prototypes. To help modelling the dialogue and enriching the context, we leverage the pre-trained knowledge adapters to infuse linguistic and factual knowledge. Experiments show that our method achieves new state-of-the-art results with 69.81% on IEMOCAP, 65.7% on MELD, and 62.51% on DailyDialog datasets. The analysis also proves that the VAD space is not only suitable for ERC but also interpretable, with VAD prototypes enhancing its performance and stabilising the training of SCCL. In addition, the pre-trained knowledge adapters benefit the performance of the utterance encoder and SCCL. Our code is available at: https://github.com/SteveKGYang/SCCLI
引用
收藏
页码:3269 / 3280
页数:12
相关论文
共 50 条
  • [21] Contrastive Learning for Domain Transfer in Cross-Corpus Emotion Recognition
    Yin, Yufeng
    Lu, Liupei
    Xiao, Yao
    Xu, Zhi
    Cai, Kaijie
    Jiang, Haonan
    Gratch, Jonathan
    Soleymani, Mohammad
    2021 9TH INTERNATIONAL CONFERENCE ON AFFECTIVE COMPUTING AND INTELLIGENT INTERACTION (ACII), 2021,
  • [22] Multimodal Prompt Transformer with Hybrid Contrastive Learning for Emotion Recognition in Conversation
    Zou, Shihao
    Huang, Xianying
    Shen, Xudong
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 5994 - 6003
  • [23] Supervised Contrastive Learning with Nearest Neighbor Search for Speech Emotion Recognition
    Wang, Xuechen
    Zhao, Shiwan
    Qin, Yong
    INTERSPEECH 2023, 2023, : 1913 - 1917
  • [24] ScSer: Supervised Contrastive Learning for Speech Emotion Recognition using Transformers
    Alaparthi, Varun Sai
    Pasam, Tejeswara Reddy
    Inagandla, Deepak Abhiram
    Prakash, Jay
    Singh, Pramod Kumar
    2022 15TH INTERNATIONAL CONFERENCE ON HUMAN SYSTEM INTERACTION (HSI), 2022,
  • [25] FCAN : Speech emotion recognition network based on focused contrastive learning
    Kang, Hong
    Xu, Yunfeng
    Jin, Guowei
    Wang, Jialin
    Miao, Borui
    BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2024, 96
  • [26] SigRep: Toward Robust Wearable Emotion Recognition With Contrastive Representation Learning
    Dissanayake, Vipula
    Seneviratne, Sachith
    Rana, Rajib
    Wen, Elliott
    Kaluarachchi, Tharindu
    Nanayakkara, Suranga
    IEEE ACCESS, 2022, 10 : 18105 - 18120
  • [27] Contrastive Learning Reduces Hallucination in Conversations
    Sun, Weiwei
    Shi, Zhengliang
    Gao, Shen
    Ren, Pengjie
    de Rijke, Maarten
    Ren, Zhaochun
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 11, 2023, : 13618 - 13626
  • [28] A Multitask learning model for multimodal sarcasm, sentiment and emotion recognition in conversations
    Zhang, Yazhou
    Wang, Jinglin
    Liu, Yaochen
    Rong, Lu
    Zheng, Qian
    Song, Dawei
    Tiwari, Prayag
    Qin, Jing
    INFORMATION FUSION, 2023, 93 : 282 - 301
  • [29] Improving Contrastive Learning in Emotion Recognition in Conversation via Data Augmentation and Decoupled Neutral Emotion
    Kang, Yujin
    Cho, Yoon-Sik
    PROCEEDINGS OF THE 18TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1: LONG PAPERS, 2024, : 2194 - 2208
  • [30] Cluster failure or power failure? Evaluating sensitivity in cluster-level inference
    Noble, Stephanie
    Scheinost, Dustin
    Constable, R. Todd
    NEUROIMAGE, 2020, 209