AFFECT BURST RECOGNITION USING MULTI-MODAL CUES

被引:0
|
作者
Turker, Bekir Berker [1 ]
Marzban, Shabbir [1 ]
Erzin, Engin [1 ]
Yemez, Yucel [1 ]
Sezgin, Tevfik Metin [1 ]
机构
[1] Koc Univ, Muhendisl Fak, Istanbul, Turkey
来源
2014 22ND SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS CONFERENCE (SIU) | 2014年
关键词
D O I
暂无
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Affect bursts, which are nonverbal expressions of emotions in conversations, play a critical role in analyzing affective states. Although there exist a number of methods on affect burst detection and recognition using only audio information, little effort has been spent for combining cues in a multimodal setup. We suggest that facial gestures constitute a key component to characterize affect bursts, and hence have potential for more robust affect burst detection and recognition. We take a data-driven approach to characterize affect bursts using Hidden Markov Models (HMI, and employ a multimodal decision fusion scheme that combines cues from audio and facial gestures for classification of affect bursts. We demonstrate the contribution of facial gestures to affect burst recognition by conducting experiments on an audiovisual database which comprise speech and facial motion data belonging to various dyadic conversations. Keywords: affect burst, multimodal recognition
引用
收藏
页码:1608 / 1611
页数:4
相关论文
共 50 条
  • [41] Multi-modal Emotion Recognition using Speech Features and Text Embedding
    Kim J.-H.
    Lee S.-P.
    Transactions of the Korean Institute of Electrical Engineers, 2021, 70 (01): : 108 - 113
  • [42] A Multi-Modal Approach to Emotion Recognition using Undirected Topic Models
    Shah, Mohit
    Chakrabarti, Chaitali
    Spanias, Andreas
    2014 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS), 2014, : 754 - 757
  • [43] MULTI-MODAL EAR AND FACE MODELING AND RECOGNITION
    Mahoor, Mohammad H.
    Cadavid, Steven
    Abdel-Mottaleb, Mohamed
    2009 16TH IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, VOLS 1-6, 2009, : 4137 - +
  • [44] Dig into Multi-modal Cues for Video Retrieval with Hierarchical Alignment
    Wang, Wenzhe
    Zhang, Mengdan
    Chen, Runnan
    Cai, Guanyu
    Zhou, Penghao
    Peng, Pai
    Guo, Xiaowei
    Wu, Jian
    Sun, Xing
    PROCEEDINGS OF THE THIRTIETH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2021, 2021, : 1113 - 1121
  • [45] Learning to Generate Object Segment Proposals with Multi-modal Cues
    Zhang, Haoyang
    He, Xuming
    Porikli, Fatih
    COMPUTER VISION - ACCV 2016, PT I, 2017, 10111 : 121 - 136
  • [46] Multi-Modal Fused-Attention Network for Depression Level Recognition Based on Enhanced Audiovisual Cues
    Zhou, Yihan
    Yu, Xiaokang
    Huang, Zixi
    Palati, Feierdun
    Zhao, Zeyu
    He, Zihan
    Feng, Yuan
    Luo, Yuxi
    IEEE ACCESS, 2025, 13 : 37913 - 37923
  • [47] Automatic temporal ranking of children's engagement levels using multi-modal cues
    Kim, Jaebok
    Truong, Khiet P.
    Evers, Vanessa
    COMPUTER SPEECH AND LANGUAGE, 2018, 50 : 16 - 39
  • [48] MUSE: MULTI-MODAL TARGET SPEAKER EXTRACTION WITH VISUAL CUES
    Pan, Zexu
    Tao, Ruijie
    Xu, Chenglin
    Li, Haizhou
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 6678 - 6682
  • [49] MMSNet: Multi-modal scene recognition using multi-scale encoded features
    Caglayan, Ali
    Imamoglu, Nevrez
    Nakamura, Ryosuke
    IMAGE AND VISION COMPUTING, 2022, 122
  • [50] Multi-Modal Physiological Data Fusion for Affect Estimation Using Deep Learning
    Hssayeni, Murtadha D.
    Ghoraani, Behnaz
    IEEE ACCESS, 2021, 9 : 21642 - 21652