Optimizing and Evaluating Pre-Trained Large Language Models for Alzheimer's Disease Detection

被引:0
|
作者
Casu, Filippo [1 ]
Grosso, Enrico [1 ]
Lagorio, Andrea [1 ]
Trunfio, Giuseppe A. [1 ]
机构
[1] Univ Sassari, Dept Biomed Sci, Sassari, Italy
关键词
Natural Language Processing; Large Language Model; Alzheimer's Disease; Model Fine-Tuning; Machine Learning;
D O I
10.1109/PDP62718.2024.00046
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
This research focuses on developing improved diagnostic tools for Alzheimer's Disease (AD), a condition impacting approximately 50 million individuals globally. In the paper, we achieve automatic AD detection by leveraging pre-trained Large Language Models (LLMs) for linguistic analysis applied to the ADReSS/ADReSSo (Alzheimer's Dementia Recognition through Spontaneous Speech/only) Challenges datasets, following speech-to-text conversion. While the recent advancements in LLMs offer a robust foundation for their application in healthcare, fine-tuning these models for specific tasks, like AD detection, requires optimization to balance performance and computational efficiency. Also in response to data privacy concerns in healthcare, we implement our methodology on consumer-level GPU cards, which offer a practical solution for local data processing. Our approach uses fine-tuning techniques such as Low Ranking Adaptation and Parameter-Efficient Fine-Tuning to enhance the capabilities of Large Language Models within the limits of consumer-grade hardware. Additionally, we incorporate quantization to reduce computational demands while preserving model accuracy. Conducted on setups with RTX 4090 and dual RTX 3090 GPUs, our experiments demonstrate promising results that align with or surpass existing benchmarks in dementia recognition.
引用
收藏
页码:277 / 284
页数:8
相关论文
共 50 条
  • [1] Disfluencies and Fine-Tuning Pre-trained Language Models for Detection of Alzheimer's Disease
    Yuan, Jiahong
    Bian, Yuchen
    Cai, Xingyu
    Huang, Jiaji
    Ye, Zheng
    Church, Kenneth
    [J]. INTERSPEECH 2020, 2020, : 2162 - 2166
  • [2] Evaluating Commonsense in Pre-Trained Language Models
    Zhou, Xuhui
    Zhang, Yue
    Cui, Leyang
    Huang, Dandan
    [J]. THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 9733 - 9740
  • [3] Evaluating the Summarization Comprehension of Pre-Trained Language Models
    Chernyshev, D. I.
    Dobrov, B. V.
    [J]. LOBACHEVSKII JOURNAL OF MATHEMATICS, 2023, 44 (08) : 3028 - 3039
  • [4] Evaluating and Inducing Personality in Pre-trained Language Models
    Jiang, Guangyuan
    Xu, Manjie
    Zhu, Song-Chun
    Han, Wenjuan
    Zhang, Chi
    Zhu, Yixin
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [5] Evaluating the Summarization Comprehension of Pre-Trained Language Models
    D. I. Chernyshev
    B. V. Dobrov
    [J]. Lobachevskii Journal of Mathematics, 2023, 44 : 3028 - 3039
  • [6] Pre-trained language models evaluating themselves - A comparative study
    Koch, Philipp
    Assenmacher, Matthias
    Heumann, Christian
    [J]. PROCEEDINGS OF THE THIRD WORKSHOP ON INSIGHTS FROM NEGATIVE RESULTS IN NLP (INSIGHTS 2022), 2022, : 180 - 187
  • [7] Probing Pre-Trained Language Models for Disease Knowledge
    Alghanmi, Israa
    Espinosa-Anke, Luis
    Schockaert, Steven
    [J]. FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL-IJCNLP 2021, 2021, : 3023 - 3033
  • [8] Probing Toxic Content in Large Pre-Trained Language Models
    Ousidhoum, Nedjma
    Zhao, Xinran
    Fang, Tianqing
    Song, Yangqiu
    Yeung, Dit-Yan
    [J]. 59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (ACL-IJCNLP 2021), VOL 1, 2021, : 4262 - 4274
  • [9] Pre-Trained Language Models and Their Applications
    Wang, Haifeng
    Li, Jiwei
    Wu, Hua
    Hovy, Eduard
    Sun, Yu
    [J]. ENGINEERING, 2023, 25 : 51 - 65
  • [10] Adapting Pre-trained Language Models to Rumor Detection on Twitter
    Slimi, Hamda
    Bounhas, Ibrahim
    Slimani, Yahya
    [J]. JOURNAL OF UNIVERSAL COMPUTER SCIENCE, 2021, 27 (10) : 1128 - 1148