DriveSense: A Multi-modal Emotion Recognition and Regulation System for a Car Driver

被引:0
|
作者
Zhu, Lei [1 ]
Zhong, Zhinan [1 ]
Dai, Wan [1 ]
Chen, Yunfei [1 ]
Zhang, Yan [1 ]
Chen, Mo [2 ]
机构
[1] Southeast Univ, Sch Mech Engn, Nanjing 211189, Peoples R China
[2] Nanjing Tech Univ, Coll Art & Design, Nanjing 210096, Peoples R China
关键词
Driver emotion; Emotion recognition; Emotion regulation; Human-machine interaction; Deep learning; FACE;
D O I
10.1007/978-3-031-60477-5_7
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Negative emotions significantly impact cognitive behavior and are a critical factor in driver safety and road traffic security. As intelligent driving systems evolve, the recognition and management of driver emotions has emerged as a crucial focus in automotive Human-Machine Interaction (HMI). We introduce DriveSense, an innovative emotion recognition and regulation system. DriveSense utilizes multi-modal data from onboard sensors, processes this data via deep learning in the cloud, and communicates with the HMI to implement regulation strategies. Our multi-modal emotion recognition model combines facial expression analysis and speech processing through Mel-frequency cepstral coefficients (MFCC), achieving a 60.37% accuracy on the RAVDESS dataset. We further validate DriveSense's utility through an experiment with 40 participants using a simulated driving scenario to test an adaptive music-based emotion regulation strategy. The results indicate that adaptive music can mitigate negative emotions effectively, underscoring DriveSense's potential to improve driver safety and secure driving practices.
引用
收藏
页码:82 / 97
页数:16
相关论文
共 50 条
  • [1] A multi-modal deep learning system for Arabic emotion recognition
    Abu Shaqra F.
    Duwairi R.
    Al-Ayyoub M.
    [J]. International Journal of Speech Technology, 2023, 26 (01) : 123 - 139
  • [2] Multi-modal Attention for Speech Emotion Recognition
    Pan, Zexu
    Luo, Zhaojie
    Yang, Jichen
    Li, Haizhou
    [J]. INTERSPEECH 2020, 2020, : 364 - 368
  • [3] Towards Efficient Multi-Modal Emotion Recognition
    Dobrisek, Simon
    Gajsek, Rok
    Mihelic, France
    Pavesic, Nikola
    Struc, Vitomir
    [J]. INTERNATIONAL JOURNAL OF ADVANCED ROBOTIC SYSTEMS, 2013, 10
  • [4] Emotion Recognition from Multi-Modal Information
    Wu, Chung-Hsien
    Lin, Jen-Chun
    Wei, Wen-Li
    Cheng, Kuan-Chun
    [J]. 2013 ASIA-PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE (APSIPA), 2013,
  • [5] Evaluation and Discussion of Multi-modal Emotion Recognition
    Rabie, Ahmad
    Wrede, Britta
    Vogt, Thurid
    Hanheide, Marc
    [J]. SECOND INTERNATIONAL CONFERENCE ON COMPUTER AND ELECTRICAL ENGINEERING, VOL 1, PROCEEDINGS, 2009, : 598 - +
  • [6] Multi-modal Emotion Recognition Based on Hypergraph
    Zong L.-L.
    Zhou J.-H.
    Xie Q.-J.
    Zhang X.-C.
    Xu B.
    [J]. Jisuanji Xuebao/Chinese Journal of Computers, 2023, 46 (12): : 2520 - 2534
  • [7] Multi-modal Emotion Recognition Based on Speech and Image
    Li, Yongqiang
    He, Qi
    Zhao, Yongping
    Yao, Hongxun
    [J]. ADVANCES IN MULTIMEDIA INFORMATION PROCESSING - PCM 2017, PT I, 2018, 10735 : 844 - 853
  • [8] Multi-Modal Emotion Recognition Fusing Video and Audio
    Xu, Chao
    Du, Pufeng
    Feng, Zhiyong
    Meng, Zhaopeng
    Cao, Tianyi
    Dong, Caichao
    [J]. APPLIED MATHEMATICS & INFORMATION SCIENCES, 2013, 7 (02): : 455 - 462
  • [9] Multi-modal Correlated Network for emotion recognition in speech
    Ren, Minjie
    Nie, Weizhi
    Liu, Anan
    Su, Yuting
    [J]. VISUAL INFORMATICS, 2019, 3 (03) : 150 - 155
  • [10] ATTENTION DRIVEN FUSION FOR MULTI-MODAL EMOTION RECOGNITION
    Priyasad, Darshana
    Fernando, Tharindu
    Denman, Simon
    Sridharan, Sridha
    Fookes, Clinton
    [J]. 2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 3227 - 3231