AFFECT BURST RECOGNITION USING MULTI-MODAL CUES

被引:0
|
作者
Turker, Bekir Berker [1 ]
Marzban, Shabbir [1 ]
Erzin, Engin [1 ]
Yemez, Yucel [1 ]
Sezgin, Tevfik Metin [1 ]
机构
[1] Koc Univ, Muhendisl Fak, Istanbul, Turkey
关键词
D O I
暂无
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Affect bursts, which are nonverbal expressions of emotions in conversations, play a critical role in analyzing affective states. Although there exist a number of methods on affect burst detection and recognition using only audio information, little effort has been spent for combining cues in a multimodal setup. We suggest that facial gestures constitute a key component to characterize affect bursts, and hence have potential for more robust affect burst detection and recognition. We take a data-driven approach to characterize affect bursts using Hidden Markov Models (HMI, and employ a multimodal decision fusion scheme that combines cues from audio and facial gestures for classification of affect bursts. We demonstrate the contribution of facial gestures to affect burst recognition by conducting experiments on an audiovisual database which comprise speech and facial motion data belonging to various dyadic conversations. Keywords: affect burst, multimodal recognition
引用
下载
收藏
页码:1608 / 1611
页数:4
相关论文
共 50 条
  • [41] A multi-modal Eliza using natural language processing and emotion recognition
    Fitrianie, S
    Wiggers, P
    Rothkrantz, LJM
    TEXT, SPEECH AND DIALOGUE, PROCEEDINGS, 2003, 2807 : 394 - 399
  • [42] Multi-modal Emotion Recognition using Speech Features and Text Embedding
    Kim J.-H.
    Lee S.-P.
    Transactions of the Korean Institute of Electrical Engineers, 2021, 70 (01): : 108 - 113
  • [43] Learning to Generate Object Segment Proposals with Multi-modal Cues
    Zhang, Haoyang
    He, Xuming
    Porikli, Fatih
    COMPUTER VISION - ACCV 2016, PT I, 2017, 10111 : 121 - 136
  • [44] Dig into Multi-modal Cues for Video Retrieval with Hierarchical Alignment
    Wang, Wenzhe
    Zhang, Mengdan
    Chen, Runnan
    Cai, Guanyu
    Zhou, Penghao
    Peng, Pai
    Guo, Xiaowei
    Wu, Jian
    Sun, Xing
    PROCEEDINGS OF THE THIRTIETH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2021, 2021, : 1113 - 1121
  • [45] Automatic temporal ranking of children's engagement levels using multi-modal cues
    Kim, Jaebok
    Truong, Khiet P.
    Evers, Vanessa
    COMPUTER SPEECH AND LANGUAGE, 2018, 50 : 16 - 39
  • [46] MUSE: MULTI-MODAL TARGET SPEAKER EXTRACTION WITH VISUAL CUES
    Pan, Zexu
    Tao, Ruijie
    Xu, Chenglin
    Li, Haizhou
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 6678 - 6682
  • [47] MMSNet: Multi-modal scene recognition using multi-scale encoded features
    Caglayan, Ali
    Imamoglu, Nevrez
    Nakamura, Ryosuke
    IMAGE AND VISION COMPUTING, 2022, 122
  • [48] Multi-Modal Physiological Data Fusion for Affect Estimation Using Deep Learning
    Hssayeni, Murtadha D.
    Ghoraani, Behnaz
    IEEE ACCESS, 2021, 9 : 21642 - 21652
  • [49] A new multi-modal dataset for human affect analysis
    1600, Springer Verlag (8749):
  • [50] A New Multi-modal Dataset for Human Affect Analysis
    Wei, Haolin
    Monaghan, David S.
    O'Connor, Noel E.
    Scanlon, Patricia
    HUMAN BEHAVIOR UNDERSTANDING (HBU 2014), 2014, 8749 : 42 - 51