Multimodal emotion recognition based on feature selection and extreme learning machine in video clips

被引:0
|
作者
Bei Pan
Kaoru Hirota
Zhiyang Jia
Linhui Zhao
Xiaoming Jin
Yaping Dai
机构
[1] Beijing Institute of Technology,School of Automation
[2] Beijing Union University,College of Robotics
[3] Beijing Engineering Research Center of Smart Mechanical Innovation Design Service,undefined
来源
Journal of Ambient Intelligence and Humanized Computing | 2023年 / 14卷
关键词
Emotion recognition; Multimodal fusion; Evolutionary optimization; Feature selection; Extreme learning machine;
D O I
暂无
中图分类号
学科分类号
摘要
Multimodal fusion-based emotion recognition has attracted increasing attention in affective computing because different modalities can achieve information complementation. One of the main challenges for reliable and effective model design is to define and extract appropriate emotional features from different modalities. In this paper, we present a novel multimodal emotion recognition framework to estimate categorical emotions, where visual and audio signals are utilized as multimodal input. The model learns neural appearance and key emotion frame using a statistical geometric method, which acts as a pre-processer for saving computation power. Discriminative emotion features expressed from visual and audio modalities are extracted through evolutionary optimization, and then fed to the optimized extreme learning machine (ELM) classifiers for unimodal emotion recognition. Finally, a decision-level fusion strategy is applied to integrate the results of predicted emotions by the different classifiers to enhance the overall performance. The effectiveness of the proposed method is demonstrated through three public datasets, i.e., the acted CK+ dataset, the acted Enterface05 dataset, and the spontaneous BAUM-1s dataset. An average recognition rate of 93.53%\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\%$$\end{document} on CK+, 91.62%\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\%$$\end{document} on Enterface05, and 60.77%\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\%$$\end{document} on BAUM-1s are obtained. The emotion recognition results acquired by fusing visual and audio predicted emotions are superior to both recognition of unimodality and concatenation of individual features.
引用
收藏
页码:1903 / 1917
页数:14
相关论文
共 50 条
  • [21] FEATURE SELECTION AND RECOGNITION OF ELECTROENCEPHALOGRAM SIGNALS: AN EXTREME LEARNING MACHINE AND GENETIC ALGORITHM-BASED APPROACH
    Lin, Qin
    Huang, Jia-Bo
    Zhong, Jian
    Lin, Si-Da
    Xue, Yun
    PROCEEDINGS OF 2015 INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND CYBERNETICS, VOL. 2, 2015, : 499 - 504
  • [22] Emotion recognition from multimodal physiological measurements based on an interpretable feature selection method
    Polo, Edoardo Maria
    Mollura, Maximiliano
    Lenatti, Marta
    Zanet, Marco
    Paglialonga, Alessia
    Barbieri, Riccardo
    2021 43RD ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE & BIOLOGY SOCIETY (EMBC), 2021, : 989 - 992
  • [23] SVM-based feature selection methods for emotion recognition from multimodal data
    Cristian Torres-Valencia
    Mauricio Álvarez-López
    Álvaro Orozco-Gutiérrez
    Journal on Multimodal User Interfaces, 2017, 11 : 9 - 23
  • [24] SVM-based feature selection methods for emotion recognition from multimodal data
    Torres-Valencia, Cristian
    Alvarez-Lopez, Mauricio
    Orozco-Gutierrez, Alvaro
    JOURNAL ON MULTIMODAL USER INTERFACES, 2017, 11 (01) : 9 - 23
  • [25] Emotion Recognition Related to Stock Trading Using Machine Learning Algorithms With Feature Selection
    Torres, Edgar P.
    Torres, Edgar Alejandro
    Hernandez-Alvarez, Myriam
    Yoo, Sang Guun
    IEEE ACCESS, 2020, 8 (08): : 199719 - 199732
  • [26] Multimodal interaction enhanced representation learning for video emotion recognition
    Xia, Xiaohan
    Zhao, Yong
    Jiang, Dongmei
    FRONTIERS IN NEUROSCIENCE, 2022, 16
  • [27] EmoNets: Multimodal deep learning approaches for emotion recognition in video
    Kahou, Samira Ebrahimi
    Bouthillier, Xavier
    Lamblin, Pascal
    Gulcehre, Caglar
    Michalski, Vincent
    Konda, Kishore
    Jean, Sebastien
    Froumenty, Pierre
    Dauphin, Yann
    Boulanger-Lewandowski, Nicolas
    Ferrari, Raul Chandias
    Mirza, Mehdi
    Warde-Farley, David
    Courville, Aaron
    Vincent, Pascal
    Memisevic, Roland
    Pal, Christopher
    Bengio, Yoshua
    JOURNAL ON MULTIMODAL USER INTERFACES, 2016, 10 (02) : 99 - 111
  • [28] EmoNets: Multimodal deep learning approaches for emotion recognition in video
    Samira Ebrahimi Kahou
    Xavier Bouthillier
    Pascal Lamblin
    Caglar Gulcehre
    Vincent Michalski
    Kishore Konda
    Sébastien Jean
    Pierre Froumenty
    Yann Dauphin
    Nicolas Boulanger-Lewandowski
    Raul Chandias Ferrari
    Mehdi Mirza
    David Warde-Farley
    Aaron Courville
    Pascal Vincent
    Roland Memisevic
    Christopher Pal
    Yoshua Bengio
    Journal on Multimodal User Interfaces, 2016, 10 : 99 - 111
  • [29] Feature Selection for Multimodal Emotion Recognition in the Arousal-Valence Space
    Torres, Cristian A.
    Orozco, Alvaro A.
    Alvarez, Mauricio A.
    2013 35TH ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY (EMBC), 2013, : 4330 - 4333
  • [30] EEG-Based Emotion Recognition with Manifold Regularized Extreme Learning Machine
    Peng, Yong
    Zhu, Jia-Yi
    Zheng, Wei-Long
    Lu, Bao-Liang
    2014 36TH ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY (EMBC), 2014, : 974 - 977