The uulmMAC Database-A Multimodal Affective Corpus for Affective Computing in Human-Computer Interaction

被引:20
|
作者
Hazer-Rau, Dilana [1 ]
Meudt, Sascha [2 ]
Daucher, Andreas [1 ]
Spohrs, Jennifer [1 ]
Hoffmann, Holger [1 ]
Schwenker, Friedhelm [2 ]
Traue, Harald C. [1 ]
机构
[1] Univ Ulm, Sect Med Psychol, Frauensteige 6, D-89075 Ulm, Germany
[2] Univ Ulm, Inst Neural Informat Proc, D-89081 Ulm, Germany
关键词
affective corpus; multimodal sensors; overload; underload; interest; frustration; cognitive load; emotion recognition; stress research; affective computing; machine learning; human-computer interaction; COGNITIVE LOAD; MENTAL WORKLOAD; EMOTION; QUESTIONNAIRE; TECHNOLOGIES;
D O I
10.3390/s20082308
中图分类号
O65 [分析化学];
学科分类号
070302 ; 081704 ;
摘要
In this paper, we present a multimodal dataset for affective computing research acquired in a human-computer interaction (HCI) setting. An experimental mobile and interactive scenario was designed and implemented based on a gamified generic paradigm for the induction of dialog-based HCI relevant emotional and cognitive load states. It consists of six experimental sequences, inducing Interest, Overload, Normal, Easy, Underload, and Frustration. Each sequence is followed by subjective feedbacks to validate the induction, a respiration baseline to level off the physiological reactions, and a summary of results. Further, prior to the experiment, three questionnaires related to emotion regulation (ERQ), emotional control (TEIQue-SF), and personality traits (TIPI) were collected from each subject to evaluate the stability of the induction paradigm. Based on this HCI scenario, the University of Ulm Multimodal Affective Corpus (uulmMAC), consisting of two homogenous samples of 60 participants and 100 recording sessions was generated. We recorded 16 sensor modalities including 4 x video, 3 x audio, and 7 x biophysiological, depth, and pose streams. Further, additional labels and annotations were also collected. After recording, all data were post-processed and checked for technical and signal quality, resulting in the final uulmMAC dataset of 57 subjects and 95 recording sessions. The evaluation of the reported subjective feedbacks shows significant differences between the sequences, well consistent with the induced states, and the analysis of the questionnaires shows stable results. In summary, our uulmMAC database is a valuable contribution for the field of affective computing and multimodal data analysis: Acquired in a mobile interactive scenario close to real HCI, it consists of a large number of subjects and allows transtemporal investigations. Validated via subjective feedbacks and checked for quality issues, it can be used for affective computing and machine learning applications.
引用
收藏
页数:33
相关论文
共 50 条
  • [41] Multimodal sentence similarity in human-computer interaction systems
    Ferri, Fernando
    Grifoni, Patrizia
    Paolozzi, Stefano
    [J]. KNOWLEDGE-BASED INTELLIGENT INFORMATION AND ENGINEERING SYSTEMS: KES 2007 - WIRN 2007, PT II, PROCEEDINGS, 2007, 4693 : 403 - 410
  • [42] A Software Framework for Multimodal Human-Computer Interaction Systems
    Shen, Jie
    Pantic, Maja
    [J]. 2009 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN AND CYBERNETICS (SMC 2009), VOLS 1-9, 2009, : 2038 - 2045
  • [43] Multimodal human-computer interaction system for speech rehabilitation
    Zhao, Jian
    Wang, Liu
    Shi, Li-Juan
    Kuang, Zhe-Jun
    Wang, Si-Bo
    Sun, Yuan-Yuan
    [J]. Jilin Daxue Xuebao (Gongxueban)/Journal of Jilin University (Engineering and Technology Edition), 2020, 50 (04): : 1478 - 1486
  • [44] Physicians in virtual environments -: multimodal human-computer interaction
    Krapichler, C
    Haubner, M
    Lösch, A
    Schuhmann, D
    Seemann, M
    Englmeier, KH
    [J]. INTERACTING WITH COMPUTERS, 1999, 11 (04) : 427 - 452
  • [45] Error resolution during multimodal human-computer interaction
    Oviatt, S
    VanGent, R
    [J]. ICSLP 96 - FOURTH INTERNATIONAL CONFERENCE ON SPOKEN LANGUAGE PROCESSING, PROCEEDINGS, VOLS 1-4, 1996, : 204 - 207
  • [46] The JESTKOD database: an affective multimodal database of dyadic interactions
    Bozkurt, Elif
    Khaki, Hossein
    Kececi, Sinan
    Turker, B. Berker
    Yemez, Yucel
    Erzin, Engin
    [J]. LANGUAGE RESOURCES AND EVALUATION, 2017, 51 (03) : 857 - 872
  • [47] The JESTKOD database: an affective multimodal database of dyadic interactions
    Elif Bozkurt
    Hossein Khaki
    Sinan Keçeci
    B. Berker Türker
    Yücel Yemez
    Engin Erzin
    [J]. Language Resources and Evaluation, 2017, 51 : 857 - 872
  • [48] Speech and language processing for multimodal human-computer interaction
    Deng, L
    Wang, Y
    Wang, K
    Acero, A
    Hon, H
    Droppo, J
    Boulis, C
    Mahajan, M
    Huang, XD
    [J]. JOURNAL OF VLSI SIGNAL PROCESSING SYSTEMS FOR SIGNAL IMAGE AND VIDEO TECHNOLOGY, 2004, 36 (2-3): : 161 - 187
  • [49] A Multimodal Human-Computer Interaction for Smart Learning System
    Alzubi, Tareq Mahmod
    Alzubi, Jafar A.
    Singh, Ashish
    Alzubi, Omar A.
    Subramanian, Murali
    [J]. INTERNATIONAL JOURNAL OF HUMAN-COMPUTER INTERACTION, 2023,
  • [50] Speech and Language Processing for Multimodal Human-Computer Interaction
    L. Deng
    Y. Wang
    K. Wang
    A. Acero
    H. Hon
    J. Droppo
    C. Boulis
    M. Mahajan
    X.D. Huang
    [J]. Journal of VLSI signal processing systems for signal, image and video technology, 2004, 36 : 161 - 187