USING EMOTIONAL NOISE TO UNCLOUD AUDIO-VISUAL EMOTION PERCEPTUAL EVALUATION

被引:0
|
作者
Provost, Emily Mower [1 ]
Zhu, Irene [1 ]
Narayanan, Shrikanth [2 ]
机构
[1] Univ Michigan, Ann Arbor, MI 48109 USA
[2] Univ Southernn California, Elect Engn, Los Angeles, CA USA
基金
美国国家科学基金会;
关键词
Emotion perception; McGurk effect; EAR;
D O I
暂无
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Emotion perception underlies communication and social interaction, shaping how we interpret our world. However, there are many aspects of this process that we still do not fully understand. Notably, we have not yet identified how audio and video information are integrated during the perception of emotion. In this work we present an approach to enhance our understanding of this process using the McGurk effect paradigm, a framework in which stimuli composed of mismatched audio and video cues are presented to human evaluators. Our stimuli set contain sentence-level emotional stimuli with either the same emotion on each channel ("matched") or different emotions on each channel ("mismatched", for example, an angry face with a happy voice). We obtain dimensional evaluations (valence and activation) of these emotionally consistent and noisy stimuli using crowd sourcing via Amazon Mechanical Turk. We use these data to investigate the audio-visual feature bias that underlies the evaluation process. We demonstrate that both audio and video information individually contribute to the perception of these dimensional properties. We further demonstrate that the change in perception from the emotionally matched to emotionally mismatched stimuli can be modeled using only unimodal feature variation. These results provide insight into the nature of audio-visual feature integration in emotion perception.
引用
收藏
页数:6
相关论文
共 50 条
  • [41] Audio-Visual Emotion Recognition using Gaussian Mixture Models for Face and Voice
    Metallinou, Angeliki
    Lee, Sungbok
    Narayanan, Shrikanth
    [J]. ISM: 2008 IEEE INTERNATIONAL SYMPOSIUM ON MULTIMEDIA, 2008, : 250 - 257
  • [42] Joint modelling of audio-visual cues using attention mechanisms for emotion recognition
    Esam Ghaleb
    Jan Niehues
    Stylianos Asteriadis
    [J]. Multimedia Tools and Applications, 2023, 82 : 11239 - 11264
  • [43] Joint modelling of audio-visual cues using attention mechanisms for emotion recognition
    Ghaleb, Esam
    Niehues, Jan
    Asteriadis, Stylianos
    [J]. MULTIMEDIA TOOLS AND APPLICATIONS, 2023, 82 (08) : 11239 - 11264
  • [44] Audio-Visual Emotion Recognition System Using Multi-Modal Features
    Handa, Anand
    Agarwal, Rashi
    Kohli, Narendra
    [J]. INTERNATIONAL JOURNAL OF COGNITIVE INFORMATICS AND NATURAL INTELLIGENCE, 2021, 15 (04)
  • [45] Audio-visual based emotion recognition using tripled hidden Markov model
    Song, ML
    Chen, C
    You, MY
    [J]. 2004 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, VOL V, PROCEEDINGS: DESIGN AND IMPLEMENTATION OF SIGNAL PROCESSING SYSTEMS INDUSTRY TECHNOLOGY TRACKS MACHINE LEARNING FOR SIGNAL PROCESSING MULTIMEDIA SIGNAL PROCESSING SIGNAL PROCESSING FOR EDUCATION, 2004, : 877 - 880
  • [46] METHODS AND CHALLENGES FOR CREATING AN EMOTIONAL AUDIO-VISUAL DATABASE
    Pandharipande, Meghna A.
    Chakraborty, Rupayan
    Kopparapu, Sunil Kumar
    [J]. 2017 20TH CONFERENCE OF THE ORIENTAL CHAPTER OF THE INTERNATIONAL COORDINATING COMMITTEE ON SPEECH DATABASES AND SPEECH I/O SYSTEMS AND ASSESSMENT (O-COCOSDA), 2017, : 183 - 188
  • [47] Integrative interaction of emotional speech in audio-visual modality
    Dong, Haibin
    Li, Na
    Fan, Lingzhong
    Wei, Jianguo
    Xu, Junhai
    [J]. FRONTIERS IN NEUROSCIENCE, 2022, 16
  • [48] A Robust Audio-visual Speech Recognition Using Audio-visual Voice Activity Detection
    Tamura, Satoshi
    Ishikawa, Masato
    Hashiba, Takashi
    Takeuchi, Shin'ichi
    Hayamizu, Satoru
    [J]. 11TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION 2010 (INTERSPEECH 2010), VOLS 3 AND 4, 2010, : 2702 - +
  • [49] CHEAVD: a Chinese natural emotional audio-visual database
    Li, Ya
    Tao, Jianhua
    Chao, Linlin
    Bao, Wei
    Liu, Yazhu
    [J]. JOURNAL OF AMBIENT INTELLIGENCE AND HUMANIZED COMPUTING, 2017, 8 (06) : 913 - 924
  • [50] EMID: An Emotional Aligned Dataset in Audio-Visual Modality
    Zou, Jialing
    Mei, Jiahao
    Ye, Guangze
    Huai, Tianyu
    Shen, Qiwei
    Dong, Daoguo
    [J]. PROCEEDINGS OF THE 1ST INTERNATIONAL WORKSHOP ON MULTIMEDIA CONTENT GENERATION AND EVALUATION, MCGE 2023: New Methods and Practice, 2023, : 41 - 48