A Deep Learning-Based Multimodal Architecture to predict Signs of Dementia

被引:2
|
作者
Ortiz-Perez, David [1 ]
Ruiz-Ponce, Pablo [1 ]
Tomas, David [2 ]
Garcia-Rodriguez, Jose [1 ]
Vizcaya-Moreno, M. Flores [3 ]
Leo, Marco [4 ]
机构
[1] Univ Alicante, Dept Comp Sci & Technol, Carretera San Vicente Raspeig, Alicante 03690, Spain
[2] Univ Alicante, Dept Software & Comp Syst, Carretera San Vicente Raspeig, Alicante 03690, Spain
[3] Univ Alicante, Fac Hlth Sci, Unit Clin Nursing Res, Carretera San Vicente Raspeig, Alicante 03690, Spain
[4] Natl Res Council Italy, Inst Appl Sci & Intelligent Syst, I-73100 Lecce, Italy
关键词
Multimodal; Deep learning; Transformers; Dementia prediction;
D O I
10.1016/j.neucom.2023.126413
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper proposes a multimodal deep learning architecture combining text and audio information to predict dementia, a disease which affects around 55 million people all over the world and makes them in some cases dependent people. The system was evaluated on the DementiaBank Pitt Corpus dataset, which includes audio recordings as well as their transcriptions for healthy people and people with dementia. Different models have been used and tested, including Convolutional Neural Networks (CNN) for audio classification, Transformers for text classification, and a combination of both in a multimodal ensemble. These models have been evaluated on a test set, obtaining the best results by using the text modality, achieving 90.36% accuracy on the task of detecting dementia. Additionally, an analysis of the corpus has been conducted for the sake of explainability, aiming to obtain more information about how the models generate their predictions and identify patterns in the data. & COPY; 2023 The Author(s). Published by Elsevier B.V. This is an open access article under the CC BY-NC-ND license (http://creativecommons.org/licenses/by-nc-nd/4.0/).
引用
收藏
页数:10
相关论文
共 50 条
  • [1] A new model driven architecture for deep learning-based multimodal lifelog retrieval
    Ben Abdallah, Fatma
    Feki, Ghada
    Ben Ammar, Anis
    Ben Amar, Chokri
    [J]. 26. INTERNATIONAL CONFERENCE IN CENTRAL EUROPE ON COMPUTER GRAPHICS, VISUALIZATION AND COMPUTER VISION (WSCG 2018), 2018, 2803 : 8 - 17
  • [2] Multimodal deep network learning-based image annotation
    Zhu, Songhao
    Li, Xiangxiang
    Shen, Shuhan
    [J]. ELECTRONICS LETTERS, 2015, 51 (12) : 905 - 906
  • [3] Multimodal deep learning-based diagnostic model for BPPV
    Hang Lu
    Yuxing Mao
    Jinsen Li
    Lin Zhu
    [J]. BMC Medical Informatics and Decision Making, 24
  • [4] Multimodal deep learning-based diagnostic model for BPPV
    Lu, Hang
    Mao, Yuxing
    Li, Jinsen
    Zhu, Lin
    [J]. BMC MEDICAL INFORMATICS AND DECISION MAKING, 2024, 24 (01)
  • [5] Deep Learning-Based Image Segmentation on Multimodal Medical Imaging
    Guo, Zhe
    Li, Xiang
    Huang, Heng
    Guo, Ning
    Li, Quanzheng
    [J]. IEEE TRANSACTIONS ON RADIATION AND PLASMA MEDICAL SCIENCES, 2019, 3 (02) : 162 - 169
  • [6] Deep Learning-Based Multimodal Trajectory Prediction with Traffic Light
    Lee, Seoyoung
    Park, Hyogyeong
    You, Yeonhwi
    Yong, Sungjung
    Moon, Il-Young
    [J]. APPLIED SCIENCES-BASEL, 2023, 13 (22):
  • [7] Deep learning-based sign language recognition system for static signs
    Ankita Wadhawan
    Parteek Kumar
    [J]. Neural Computing and Applications, 2020, 32 : 7957 - 7968
  • [8] Deep learning-based sign language recognition system for static signs
    Wadhawan, Ankita
    Kumar, Parteek
    [J]. NEURAL COMPUTING & APPLICATIONS, 2020, 32 (12): : 7957 - 7968
  • [9] A comprehensive survey on deep learning-based approaches for multimodal sentiment analysis
    Alireza Ghorbanali
    Mohammad Karim Sohrabi
    [J]. Artificial Intelligence Review, 2023, 56 : 1479 - 1512
  • [10] A Multimodal Deep Learning-Based Distributed Network Latency Measurement System
    Mohammed, Shady A.
    Shirmohammadi, Shervin
    Altamimi, Sa'di
    [J]. IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2020, 69 (05) : 2487 - 2494