Multi-Modal Emotion Recognition Fusing Video and Audio

被引:4
|
作者
Xu, Chao [1 ]
Du, Pufeng [2 ]
Feng, Zhiyong [2 ]
Meng, Zhaopeng [1 ]
Cao, Tianyi [2 ]
Dong, Caichao [2 ]
机构
[1] Tianjin Univ, Sch Comp Software, Tianjin 300072, Peoples R China
[2] Tianjin Univ, Sch Comp Sci & Technol, Tianjin 300072, Peoples R China
来源
APPLIED MATHEMATICS & INFORMATION SCIENCES | 2013年 / 7卷 / 02期
基金
美国国家科学基金会;
关键词
Emotion Recognition; Multi-modal Fusion; HMM; Multi-layer Perceptron;
D O I
10.12785/amis/070205
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
Emotion plays an important role in human communications. We construct a framework for multi-modal fusion emotion recognition. Facial expression features and speech features are respectively extracted from image sequences and speech signals. In order to locate and track facial feature points, we construct an Active Appearance Model for facial images with all kinds of expressions. Facial Animation Parameters are calculated from motions of facial feature points as expression features. We extract short-term mean energy, fundamental frequency and formant frequencies from each frame as speech features. An emotion classifier is designed to fuse facial expression and speech based on Hidden Markov Models and Multi-layer Perceptron. Experiments indicate that multi-modal fusion emotion recognition algorithm which is presented in this paper has relatively high recognition accuracy. The proposed approach has better performance and robustness than methods using only video or audio separately.
引用
收藏
页码:455 / 462
页数:8
相关论文
共 50 条
  • [41] A comprehensive video dataset for multi-modal recognition systems
    Handa A.
    Agarwal R.
    Kohli N.
    Data Science Journal, 2019, 18 (01):
  • [42] Contextual and Cross-Modal Interaction for Multi-Modal Speech Emotion Recognition
    Yang, Dingkang
    Huang, Shuai
    Liu, Yang
    Zhang, Lihua
    IEEE SIGNAL PROCESSING LETTERS, 2022, 29 : 2093 - 2097
  • [43] Multi-modal emotion recognition using EEG and speech signals
    Wang, Qian
    Wang, Mou
    Yang, Yan
    Zhang, Xiaolei
    COMPUTERS IN BIOLOGY AND MEDICINE, 2022, 149
  • [44] Dynamic Confidence-Aware Multi-Modal Emotion Recognition
    Zhu, Qi
    Zheng, Chuhang
    Zhang, Zheng
    Shao, Wei
    Zhang, Daoqiang
    IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2024, 15 (03) : 1358 - 1370
  • [45] Multi-modal fusion network with complementarity and importance for emotion recognition
    Liu, Shuai
    Gao, Peng
    Li, Yating
    Fu, Weina
    Ding, Weiping
    INFORMATION SCIENCES, 2023, 619 : 679 - 694
  • [46] A Multi-modal Visual Emotion Recognition Method to Instantiate an Ontology
    Heredia, Juan Pablo A.
    Cardinale, Yudith
    Dongo, Irvin
    Diaz-Amado, Jose
    PROCEEDINGS OF THE 16TH INTERNATIONAL CONFERENCE ON SOFTWARE TECHNOLOGIES (ICSOFT), 2021, : 453 - 464
  • [47] Multi-Modal Fusion Emotion Recognition Based on HMM and ANN
    Xu, Chao
    Cao, Tianyi
    Feng, Zhiyong
    Dong, Caichao
    CONTEMPORARY RESEARCH ON E-BUSINESS TECHNOLOGY AND STRATEGY, 2012, 332 : 541 - 550
  • [48] A multi-modal deep learning system for Arabic emotion recognition
    Abu Shaqra F.
    Duwairi R.
    Al-Ayyoub M.
    International Journal of Speech Technology, 2023, 26 (01) : 123 - 139
  • [49] Dense Attention Memory Network for Multi-modal emotion recognition
    Ma, Gailing
    Guo, Xiao
    2022 5TH INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND NATURAL LANGUAGE PROCESSING, MLNLP 2022, 2022, : 48 - 53
  • [50] Calibration of audio-video sensors for multi-modal event indexing
    Kuehnapfel, Thorsten
    Tan, Tele
    Venkatesh, Svetha
    Lehmann, Eric
    2007 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, VOL II, PTS 1-3, 2007, : 741 - +