Multimodal assessment of apparent personality using feature attention and error consistency constraint

被引:19
|
作者
Aslan, Suleyman [1 ]
Gudukbay, Ugur [1 ]
Dibeklioglu, Hamdi [1 ]
机构
[1] Bilkent Univ, Dept Comp Engn, Ankara, Turkey
关键词
Deep learning; Apparent personality; Multimodal modeling; Information fusion; Feature attention; Error consistency; 5-FACTOR MODEL; TRAITS; RECOGNITION; PERCEPTION; AUDIO; MUSIC; LIFE;
D O I
10.1016/j.imavis.2021.104163
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Personality computing and affective computing, where the recognition of personality traits is essential, have gained increasing interest and attention in many research areas recently. We propose a novel approach to recognize the Big Five personality traits of people from videos. To this end, we use four different modalities, namely, ambient appearance (scene), facial appearance, voice, and transcribed speech. Through a specialized subnetwork for each of these modalities, our model learns reliable modality-specific representations and fuse them using an attention mechanism that re-weights each dimension of these representations to obtain an optimal combination of multimodal information. A novel loss function is employed to enforce the proposed model to give an equivalent importance for each of the personality traits to be estimated through a consistency constraint that keeps the trait-specific errors as close as possible. To further enhance the reliability of our model, we employ (pre-trained) state-of-the-art architectures (i.e., ResNet, VGGish, ELMo) as the backbones of the modality-specific subnetworks, which are complemented by multilayered Long Short-Term Memory networks to capture temporal dynamics. To minimize the computational complexity of multimodal optimization, we use two-stage modeling, where the modality-specific subnetworks are first trained individually, and the whole network is then finetuned to jointly model multimodal data. On the large scale ChaLearn First Impressions V2 challenge dataset, we evaluate the reliability of our model as well as investigating the informativeness of the considered modalities. Experimental results show the effectiveness of the proposed attention mechanism and the error consistency constraint. While the best performance is obtained using facial information among individual modalities, with the use of all four modalities, our model achieves a mean accuracy of 91.8%, improving the state of the art in automatic personality analysis. (c) 2021 Elsevier B.V. All rights reserved.
引用
收藏
页数:9
相关论文
共 25 条
  • [1] Personality Assessment Based on Multimodal Attention Network Learning With Category-Based Mean Square Error
    Sun, Xiao
    Huang, Jie
    Zheng, Shixin
    Rao, Xuanheng
    Wang, Meng
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2022, 31 : 2162 - 2174
  • [2] Multimodal assessment of visual attention using the Bethesda Eye & Attention Measure (BEAM)
    Ettenhofer, Mark L.
    Hershaw, Jamie N.
    Barry, David M.
    JOURNAL OF CLINICAL AND EXPERIMENTAL NEUROPSYCHOLOGY, 2016, 38 (01) : 96 - 110
  • [3] Unsupervised Building Extraction from Multimodal Aerial Data Based on Accurate Vegetation Removal and Image Feature Consistency Constraint
    Meng, Yan
    Chen, Shanxiong
    Liu, Yuxuan
    Li, Li
    Zhang, Zemin
    Ke, Tao
    Hu, Xiangyun
    REMOTE SENSING, 2022, 14 (08)
  • [4] Multimodal Personality Recognition using Cross-attention Transformer and Behaviour Encoding
    Agrawal, Tanay
    Agarwal, Dhruv
    Balazia, Michal
    Sinha, Neelabh
    Bremond, Francois
    PROCEEDINGS OF THE 17TH INTERNATIONAL JOINT CONFERENCE ON COMPUTER VISION, IMAGING AND COMPUTER GRAPHICS THEORY AND APPLICATIONS (VISAPP), VOL 5, 2022, : 501 - 508
  • [5] A blind quality assessment method for images using shape consistency feature
    Chu, Ying
    Mou, Xuanqin
    Hong, Wei
    Mou, Xuanqin, 1600, Xi'an Jiaotong University (48): : 12 - 17
  • [6] Feature Consistency-Based Style Transfer for Landscape Images Using Dual-Channel Attention
    Zhang, Qiang
    Wang, Shuai
    Cui, Dong
    IEEE ACCESS, 2024, 12 : 164018 - 164027
  • [7] Determination of Patient Setup Error in Radiotherapy using Feature-Based Multimodal Rigid Image Registration
    Hakim, Aayesha
    Talele, K. T. V.
    Harsh, Rajesh
    Verma, Dharmesh
    2015 IEEE BOMBAY SECTION SYMPOSIUM (IBSS), 2015,
  • [8] A Comparison of Multimodal Biomarkers for Chronic Hepatitis B Assessment Using Recursive Feature Elimination
    Bai, Yanru
    Chen, Xin
    Dong, Changfeng
    Liu, Yingxia
    Zhang, Zhiguo
    2016 38TH ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY (EMBC), 2016, : 2448 - 2451
  • [9] Detecting Attention Deficit Hyperactivity Disorder and its feigning using the Personality Assessment Inventory
    Maffly-Kipp, Joseph
    Morey, Leslie C.
    APPLIED NEUROPSYCHOLOGY-ADULT, 2023,
  • [10] A Residual Multihead Self-Attention Network Using Multimodal Shallow Feature Fusion for Motor Fault Diagnosis
    Feng, Juan
    Su, Jinya
    Feng, Xiaoliang
    IEEE SENSORS JOURNAL, 2023, 23 (23) : 29131 - 29142