Minimal representation of speech signals for generation of emotion speech and human-robot interaction

被引:0
|
作者
Lee, Heyoung [1 ]
Bien, Z. Zenn [2 ]
机构
[1] Seoul Natl Univ Technol, Dept Control & Instrumentat Engn, Seoul, South Korea
[2] Korea Adv Inst Sci & Technol, Dept Elect Engn & Comp Sci, Daejon, South Korea
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper minimal representation of voiced speech based on decomposition into ANI-FM components is proposed for generation of emotion speech. For the decomposition, firstly time-frequency boundaries of AM-FM components are estimated and secondary each AM-FM component is extracted by using the variable bandwidth filter [171 adaptive to the estimated time-frequency boundaries. Finally, two parameters, that is, instantaneous frequency and instantaneous amplitude of each AM-FM component are estimated. The set composed of instantaneous amplitudes and instantaneous frequencies is the minimal representation of voiced speech signals. The minimal representation is optimal feature set since the set describes effectively the biomechanical characteristics of the vocal codes and the vocal track. Raw speech signals are modified by changing the parameters for generation of emotion speech.
引用
收藏
页码:137 / +
页数:2
相关论文
共 50 条
  • [1] Emotion Sensitive Speech Control for Human-Robot Interaction in Minimal Invasive Surgery
    Schuller, Bjoern
    Rigoll, Gerhard
    Can, Salman
    Feussner, Hubertus
    [J]. 2008 17TH IEEE INTERNATIONAL SYMPOSIUM ON ROBOT AND HUMAN INTERACTIVE COMMUNICATION, VOLS 1 AND 2, 2008, : 453 - +
  • [2] Emotion Recognition From Speech to Improve Human-robot Interaction
    Zhu, Changrui
    Ahamd, Wasim
    [J]. IEEE 17TH INT CONF ON DEPENDABLE, AUTONOM AND SECURE COMP / IEEE 17TH INT CONF ON PERVAS INTELLIGENCE AND COMP / IEEE 5TH INT CONF ON CLOUD AND BIG DATA COMP / IEEE 4TH CYBER SCIENCE AND TECHNOLOGY CONGRESS (DASC/PICOM/CBDCOM/CYBERSCITECH), 2019, : 370 - 375
  • [3] The Necessity of Emotion Recognition from Speech Signals for Natural and Effective Human-Robot Interaction in Society 5.0
    Ulgen Sonmez, Yesim
    Varol, Asaf
    [J]. 2022 10TH INTERNATIONAL SYMPOSIUM ON DIGITAL FORENSICS AND SECURITY (ISDFS), 2022,
  • [4] A Selection Method of Speech Vocabulary for Human-Robot Speech Interaction
    Liu, Hong
    Li, Xiaofei
    [J]. IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN AND CYBERNETICS (SMC 2010), 2010,
  • [5] Speech emotion recognition in real static and dynamic human-robot interaction scenarios
    Grageda, Nicolas
    Busso, Carlos
    Alvarado, Eduardo
    Garcia, Ricardo
    Mahu, Rodrigo
    Huenupan, Fernando
    Yoma, Nestor Becerra
    [J]. COMPUTER SPEECH AND LANGUAGE, 2025, 89
  • [6] Speech Emotion Recognition Using an Enhanced Kernel Isomap for Human-Robot Interaction
    Zhang, Shiqing
    Zhao, Xiaoming
    Lei, Bicheng
    [J]. INTERNATIONAL JOURNAL OF ADVANCED ROBOTIC SYSTEMS, 2013, 10
  • [7] On the Robustness of Speech Emotion Recognition for Human-Robot Interaction with Deep Neural Networks
    Lakomkin, Egor
    Zamani, Mohammad Ali
    Weber, Cornelius
    Magg, Sven
    Wermter, Stefan
    [J]. 2018 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2018, : 854 - 860
  • [8] GENERATION OF NODDING, HEAD TILTING AND GAZING FOR HUMAN-ROBOT SPEECH INTERACTION
    Liu, Chaoran
    Ishi, Carlos T.
    Ishiguro, Hiroshi
    Hagita, Norihiro
    [J]. INTERNATIONAL JOURNAL OF HUMANOID ROBOTICS, 2013, 10 (01)
  • [9] Space, Speech, and Gesture in Human-Robot Interaction
    Mead, Ross
    [J]. ICMI '12: PROCEEDINGS OF THE ACM INTERNATIONAL CONFERENCE ON MULTIMODAL INTERACTION, 2012, : 333 - 336
  • [10] Integration of Gestures and Speech in Human-Robot Interaction
    Meena, Raveesh
    Jokinen, Kristiina
    Wilcock, Graham
    [J]. 3RD IEEE INTERNATIONAL CONFERENCE ON COGNITIVE INFOCOMMUNICATIONS (COGINFOCOM 2012), 2012, : 673 - 678