Cluster-Level Contrastive Learning for Emotion Recognition in Conversations

被引:27
|
作者
Yang, Kailai [1 ,2 ]
Zhang, Tianlin [1 ,2 ]
Alhuzali, Hassan [3 ]
Ananiadou, Sophia [1 ,2 ]
机构
[1] Univ Manchester, NaCTeM, Manchester M13 9PL, England
[2] Univ Manchester, Dept Comp Sci, Manchester M13 9PL, England
[3] Umm Al Qura Univ, Coll Comp & Informat Syst, Mecca 24382, Saudi Arabia
基金
英国生物技术与生命科学研究理事会;
关键词
Emotion recognition; Prototypes; Linguistics; Task analysis; Semantics; Training; Adaptation models; Cluster-level contrastive learning; emotion recognition in conversations; pre-trained knowledge adapters; valence-arousal-dominance; DIALOGUE; FRAMEWORK;
D O I
10.1109/TAFFC.2023.3243463
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
A key challenge for Emotion Recognition in Conversations (ERC) is to distinguish semantically similar emotions. Some works utilise Supervised Contrastive Learning (SCL) which uses categorical emotion labels as supervision signals and contrasts in high-dimensional semantic space. However, categorical labels fail to provide quantitative information between emotions. ERC is also not equally dependent on all embedded features in the semantic space, which makes the high-dimensional SCL inefficient. To address these issues, we propose a novel low-dimensional Supervised Cluster-level Contrastive Learning (SCCL) method, which first reduces the high-dimensional SCL space to a three-dimensional affect representation space Valence-Arousal-Dominance (VAD), then performs cluster-level contrastive learning to incorporate measurable emotion prototypes. To help modelling the dialogue and enriching the context, we leverage the pre-trained knowledge adapters to infuse linguistic and factual knowledge. Experiments show that our method achieves new state-of-the-art results with 69.81% on IEMOCAP, 65.7% on MELD, and 62.51% on DailyDialog datasets. The analysis also proves that the VAD space is not only suitable for ERC but also interpretable, with VAD prototypes enhancing its performance and stabilising the training of SCCL. In addition, the pre-trained knowledge adapters benefit the performance of the utterance encoder and SCCL. Our code is available at: https://github.com/SteveKGYang/SCCLI
引用
收藏
页码:3269 / 3280
页数:12
相关论文
共 50 条
  • [31] Cluster-level Feedback Power Control for Performance Optimization
    Wang, Xiaorui
    Chen, Ming
    2008 IEEE 14TH INTERNATIONAL SYMPOSIUM ON HIGH PEFORMANCE COMPUTER ARCHITECTURE, 2008, : 93 - 102
  • [32] Integrating distance metric learning and cluster-level constraints in semi-supervised clustering
    Nogueira, Bruno Magalhaes
    Benevides Tomas, Yuri Karan
    Marcacini, Ricardo Marcondes
    2017 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2017, : 4118 - 4125
  • [33] Superpixel Prior Cluster-Level Contrastive Clustering Network for Large-Scale Urban Hyperspectral Images and Vehicle Detection
    Li, Tiancong
    Cai, Yaoming
    Zhang, Yongshan
    Cai, Zhihua
    Jiang, Guozhu
    Liu, Xiaobo
    IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2025, 74 (02) : 2019 - 2031
  • [34] Text-Level Contrastive Learning for Scene Text Recognition
    Zhuang, Junbin
    Ren, Yixuan
    Li, Xia
    Liang, Zhanpeng
    2022 INTERNATIONAL CONFERENCE ON ASIAN LANGUAGE PROCESSING (IALP 2022), 2022, : 231 - 236
  • [35] Does Hard-Negative Contrastive Learning Improve Facial Emotion Recognition?
    Win, Khin Cho
    Akhtar, Zahid
    Mohan, C. Krishna
    PROCEEDINGS OF THE 2024 THE 7TH INTERNATIONAL CONFERENCE ON MACHINE VISION AND APPLICATIONS, ICMVA 2024, 2024, : 162 - 168
  • [36] Emotion recognition of EEG signals based on contrastive learning graph convolutional model
    Zhang, Yiling
    Liao, Yuan
    Chen, Wei
    Zhang, Xiruo
    Huang, Liya
    JOURNAL OF NEURAL ENGINEERING, 2024, 21 (04)
  • [37] Joyful: Joint Modality Fusion and Graph Contrastive Learning for Multimodal Emotion Recognition
    Li, Dongyuan
    Wang, Yusong
    Funakoshi, Kotaro
    Okumura, Manabu
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2023), 2023, : 16051 - 16069
  • [38] A Self-Fusion Network Based on Contrastive Learning for Group Emotion Recognition
    Wang, Xingzhi
    Zhang, Dong
    Tan, Hong-Zhou
    Lee, Dah-Jye
    IEEE TRANSACTIONS ON COMPUTATIONAL SOCIAL SYSTEMS, 2023, 10 (02) : 458 - 469
  • [39] Multimodal Attentive Learning for Real-time Explainable Emotion Recognition in Conversations
    Arumugam, Balaji
    Das Bhattacharjee, Sreyasee
    Yuan, Junsong
    2022 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS 22), 2022, : 1210 - 1214
  • [40] Deep emotion recognition in textual conversations: a survey
    Pereira, Patricia
    Moniz, Helena
    Carvalho, Joao Paulo
    ARTIFICIAL INTELLIGENCE REVIEW, 2024, 58 (01)