User Emotion Recognition Method Based on Facial Expression and Speech Signal Fusion

被引:2
|
作者
Lu, Fei [1 ]
Zhang, Long [1 ]
Tian, Guohui [1 ]
机构
[1] Shandong Univ, Sch Control Sci & Engn, Jinan 250061, Peoples R China
关键词
emotion recognition; gabor transform; transfer learning; multimodal fusion; arousal-valence;
D O I
10.1109/ICIEA51954.2021.9516216
中图分类号
T [工业技术];
学科分类号
08 ;
摘要
In human-computer interaction, it is an urgent problem to use facial expressions and speech information to identify the user's continuous emotions, and the key factors affecting the recognition accuracy are the data deficiencies during the fusion of speech and facial information, and the abnormal frames in the video. In order to solve these problems, a user emotion recognition system based on the fusion of facial expressions and speech multimodality is designed. In the part of facial expressions, Gabor transform continuous emotion recognition method based on data increments is proposed. In the part of speech information, Mel-scale Frequency Cepstral Coefficients (MFCC) is used to extract speech features, and user emotions are recognize through transfer learning. Finally, in the late fusion, multiple linear regression is used for multi-modality to verify the method in this paper. This paper uses the AVEC2013 dataset with Arousal-Valence label to conduct a valid experiment on the proposed method. The experimental results prove that the method improves the accuracy of user emotion recognition.
引用
收藏
页码:1121 / 1126
页数:6
相关论文
共 50 条
  • [1] Feature Fusion Algorithm for Multimodal Emotion Recognition from Speech and Facial Expression Signal
    Han Zhiyan
    Wang Jian
    INTERNATIONAL SEMINAR ON APPLIED PHYSICS, OPTOELECTRONICS AND PHOTONICS (APOP 2016), 2016, 61
  • [2] Multimodal emotion recognition from facial expression and speech based on feature fusion
    Guichen Tang
    Yue Xie
    Ke Li
    Ruiyu Liang
    Li Zhao
    Multimedia Tools and Applications, 2023, 82 : 16359 - 16373
  • [3] Multimodal emotion recognition from facial expression and speech based on feature fusion
    Tang, Guichen
    Xie, Yue
    Li, Ke
    Liang, Ruiyu
    Zhao, Li
    MULTIMEDIA TOOLS AND APPLICATIONS, 2023, 82 (11) : 16359 - 16373
  • [4] Research of emotion recognition based on speech and facial expression
    Wang, Yutai
    Yang, Xinghai
    Zou, Jing
    Telkomnika - Indonesian Journal of Electrical Engineering, 2013, 11 (01): : 83 - 90
  • [5] Bimodal Emotion Recognition Based on Speech Signals and Facial Expression
    Tu, Binbin
    Yu, Fengqin
    FOUNDATIONS OF INTELLIGENT SYSTEMS (ISKE 2011), 2011, 122 : 691 - 696
  • [6] Facial Expression Recognition with Emotion-Based Feature Fusion
    Turan, Cigdem
    Lam, Kin-Man
    He, Xiangjian
    2015 ASIA-PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE (APSIPA), 2015,
  • [7] Multimodal Emotion Recognition Based on Facial Expression and ECG Signal
    NIU Jianwei
    AN Yueqi
    NI Jie
    JIANG Changhua
    包装工程, 2022, 43 (04) : 71 - 79
  • [8] Emotion recognition from the facial image and speech signal
    Go, HJ
    Kwak, KC
    Lee, DJ
    Chun, MG
    SICE 2003 ANNUAL CONFERENCE, VOLS 1-3, 2003, : 2890 - 2895
  • [9] Multi-expression facial animation based on speech emotion recognition
    Research Center far Pervasive Computing, Institute of Computing Technology, Chinese Academy of Sciences, Beijing 100190, China
    Jisuanji Fuzhu Sheji Yu Tuxingxue Xuebao, 2008, 4 (520-525):
  • [10] Multi-Modal Fusion Emotion Recognition Method of Speech Expression Based on Deep Learning
    Liu, Dong
    Wang, Zhiyong
    Wang, Lifeng
    Chen, Longxi
    FRONTIERS IN NEUROROBOTICS, 2021, 15