Multimodal emotion recognition with evolutionary computation for human-robot interaction

被引:52
|
作者
Perez-Gaspar, Luis-Alberto [1 ]
Caballero-Morales, Santiago-Omar [1 ]
Trujillo-Romero, Felipe [1 ]
机构
[1] Technol Univ Mixteca, Rd Acatlima Km 2-5, Mexico City 69000, DF, Mexico
关键词
Emotion recognition; Principal Component Analysis; Hidden Markov Models; Genetic Algorithms; Artificial Neural Networks; Finite state machines; SPEECH; CLASSIFIERS; FEATURES; FUSION;
D O I
10.1016/j.eswa.2016.08.047
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Service robotics is an important field of research for the development of assistive technologies. Particularly, humanoid robots will play an increasing and important role in our society. More natural assistive interaction with humanoid robots can be achieved if the emotional aspect is considered. However emotion recognition is one of the most challenging topics in pattern recognition and improved intelligent techniques have to be developed to accomplish this goal. Recent research has addressed the emotion recognition problem with techniques such as Artificial Neural Networks (ANNs)/Hidden Markov Models (HMMs) and reliability of proposed approaches has been assessed (in most cases) with standard databases. In this work we (1) explored on the implications of using standard databases for assessment of emotion recognition techniques, (2) extended on the evolutionary optimization of ANNs and HMMs for the development of a multimodal emotion recognition system, (3) set the guidelines for the development of emotional databases of speech and facial expressions, (4) rules were set for phonetic transcription of Mexican speech, and (5) evaluated the suitability of the multimodal system within the context of spoken dialogue between a humanoid robot and human users. The development of intelligent systems for emotion recognition can be improved by the findings of the present work: (a) emotion recognition depends on the structure of the database sub-sets used for training and testing, and it also depends on the type of technique used for recognition where a specific emotion can be highly recognized by a specific technique, (b) optimization of HMMs led to a Bakis structure which is more suitable for acoustic modeling of emotion-specific vowels while optimization of ANNs led to a more suitable ANN structure for recognition of facial expressions, (c) some emotions can be better recognized based on speech patterns instead of visual patterns, and (d) the weighted integration of the multimodal emotion recognition system optimized with these observations can achieve a recognition rate up to 97.00 % in live dialogue tests with a humanoid robot. (C) 2016 Elsevier Ltd. All rights reserved.
引用
收藏
页码:42 / 61
页数:20
相关论文
共 50 条
  • [41] Effects of Emotion Grouping for Recognition in Human-Robot Interactions
    Tozadore, Daniel C.
    Ranieri, Caetano M.
    Nardari, Guilherme V.
    Romero, Roseli A. F.
    Guizilini, Vitor C.
    [J]. 2018 7TH BRAZILIAN CONFERENCE ON INTELLIGENT SYSTEMS (BRACIS), 2018, : 438 - 443
  • [42] Investigating Effects of Multimodal Topic-continuance Recognition on Human-Robot Interviewing Interaction
    Nagasawa, Fuminori
    Okada, Shogo
    [J]. ACM/IEEE International Conference on Human-Robot Interaction, : 779 - 783
  • [43] Investigating Effects of Multimodal Topic-continuance Recognition on Human-Robot Interviewing Interaction
    Nagasawa, Fuminori
    Okada, Shogo
    [J]. COMPANION OF THE 2024 ACM/IEEE INTERNATIONAL CONFERENCE ON HUMAN-ROBOT INTERACTION, HRI 2024 COMPANION, 2024, : 779 - 783
  • [44] A unified multimodal control framework for human-robot interaction
    Cherubini, Andrea
    Passama, Robin
    Fraisse, Philippe
    Crosnier, Andre
    [J]. ROBOTICS AND AUTONOMOUS SYSTEMS, 2015, 70 : 106 - 115
  • [45] Knowledge acquisition through human-robot multimodal interaction
    Randelli, Gabriele
    Bonanni, Taigo Maria
    Iocchi, Luca
    Nardi, Daniele
    [J]. INTELLIGENT SERVICE ROBOTICS, 2013, 6 (01) : 19 - 31
  • [46] Multimodal Engagement Prediction in Multiperson Human-Robot Interaction
    Abdelrahman, Ahmed A.
    Strazdas, Dominykas
    Khalifa, Aly
    Hintz, Jan
    Hempel, Thorsten
    Al-Hamadi, Ayoub
    [J]. IEEE ACCESS, 2022, 10 : 61980 - 61991
  • [47] Challenges of Multimodal Interaction in the Era of Human-Robot Coexistence
    Zhang, Zhengyou
    [J]. ICMI'19: PROCEEDINGS OF THE 2019 INTERNATIONAL CONFERENCE ON MULTIMODAL INTERACTION, 2019, : 2 - 2
  • [48] DiGeTac Unit for Multimodal Communication in Human-Robot Interaction
    Al, Gorkem Anil
    Martinez-Hernandez, Uriel
    [J]. IEEE SENSORS LETTERS, 2024, 8 (05) : 1 - 4
  • [49] A Multimodal Human-Robot Interaction Manager for Assistive Robots
    Abbasi, Bahareh
    Monaikul, Natawut
    Rysbek, Zhanibek
    Di Eugenio, Barbara
    Zefran, Milos
    [J]. 2019 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2019, : 6756 - 6762
  • [50] Multimodal QOL Estimation During Human-Robot Interaction
    Nakagawa, Satoshi
    Kuniyoshi, Yasuo
    [J]. 2024 IEEE INTERNATIONAL CONFERENCE ON DIGITAL HEALTH, ICDH 2024, 2024, : 23 - 32