Leveraging large language models for medical text classification: a hospital readmission prediction case

被引:0
|
作者
Nazyrova, Nodira [1 ]
Chahed, Salma [1 ]
Chausalet, Thierry [1 ]
Dwek, Miriam [2 ]
机构
[1] Univ Westminster, Sch Comp Sci & Engn, London, England
[2] Univ Westminster, Sch Life Sci, London, England
关键词
hospital readmission prediction; domain-specific transformer models; BERT; ClinicalBERT; SciBERT; BioBERT; large language models;
D O I
10.1109/ICPRS62101.2024.10677826
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In recent years, the intersection of natural language processing (NLP) and healthcare informatics has witnessed a revolutionary transformation. One of the most groundbreaking developments in this realm is the advent of large language models (LLM), which have demonstrated remarkable capabilities in analysing clinical data. This paper aims to explore the potential of large language models in medical text classification, shedding light on their ability to discern subtle patterns, grasp domain-specific terminology, and adapt to the dynamic nature of medical information. This research focuses on the application of transformer-based models, such as Bidirectional Encoder Representations from Transformers (BERT), on hospital discharge summaries to predict 30-day readmissions among older adults. In particular, we explore the role of transfer learning in medical text classification and compare domain-specific transformer models, such as SciBERT, BioBERT and ClinicalBERT. We also analyse how data preprocessing techniques affect the performance of language models. Our comparative analysis shows that removing parts of text with a large proportion of out-of-vocabulary words improves the classification results. We also investigate how the input sequence length affects the model performance, varying sequence length from 128 to 512 for BERT-based models and 4096 sequence length for the Longformers. The results of the investigation showed that among compared models SciBERT yields the best performance when applied in the medical domain, improving current hospital readmission predictions using clinical notes on MIMIC data from 0.714 to 0.735 AUROC. Our next step is pretraining a model with a large corpus of clinical notes to potentially improve the adaptability of a language model in the medical domain and achieve better results in downstream tasks.
引用
收藏
页数:7
相关论文
共 50 条
  • [21] Synthetic Data Generation with Large Language Models for Text Classification: Potential and Limitations
    Li, Zhuoyan
    Zhu, Hangxiao
    Lu, Zhuoran
    Yin, Ming
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2023), 2023, : 10443 - 10461
  • [22] Leveraging large language models for predictive chemistry
    Jablonka, Kevin Maik
    Schwaller, Philippe
    Ortega-Guerrero, Andres
    Smit, Berend
    NATURE MACHINE INTELLIGENCE, 2024, 6 (02) : 122 - 123
  • [23] SPIN: Sparsifying and Integrating Internal Neurons in Large Language Models for Text Classification
    Jiao, Difan
    Liu, Yilun
    Tang, Zhenwei
    Matter, Daniel
    Pfeffer, Jurgen
    Anderson, Ashton
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 4666 - 4682
  • [24] LEVERAGING LARGE LANGUAGE MODELS FOR CARDIOVASCULAR MORTALITY PREDICTION FROM CT CHEST REPORTS
    James, Jose
    Brooks, Hunter
    Kullo, Iftikhar J.
    JOURNAL OF THE AMERICAN COLLEGE OF CARDIOLOGY, 2024, 83 (13) : 2441 - 2441
  • [25] Beyond Detection: Leveraging Large Language Models for Cyber Attack Prediction in IoT Networks
    Diaf, Alaeddine
    Korba, Abdelaziz Amara
    Karabadji, Nour Elislem
    Ghamri-Doudane, Yacine
    2024 20TH INTERNATIONAL CONFERENCE ON DISTRIBUTED COMPUTING IN SMART SYSTEMS AND THE INTERNET OF THINGS, DCOSS-IOT 2024, 2024, : 117 - 123
  • [26] Large Language Models for Clinical Text Cleansing Enhance Medical Concept Normalization
    Abdulnazar, Akhila
    Roller, Roland
    Schulz, Stefan
    Kreuzthaler, Markus
    IEEE ACCESS, 2024, 12 : 147981 - 147990
  • [27] Causality Extraction from Medical Text Using Large Language Models (LLMs)
    Gopalakrishnan, Seethalakshmi
    Garbayo, Luciana
    Zadrozny, Wlodek
    INFORMATION, 2025, 16 (01)
  • [28] Leveraging Text-to-Text Pretrained Language Models for Question Answering in Chemistry
    Tran, Dan
    Pascazio, Laura
    Akroyd, Jethro
    Mosbach, Sebastian
    Kraft, Markus
    ACS OMEGA, 2024, 9 (12): : 13883 - 13896
  • [29] Exploring the Potential of Large Language Models for Text-Based Personality Prediction
    Molchanova, Maria
    NATURAL LANGUAGE PROCESSING AND INFORMATION SYSTEMS, PT II, NLDB 2024, 2024, 14763 : 291 - 301
  • [30] Transforming legal text interactions: leveraging natural language processing and large language models for legal support in Palestinian cooperatives
    Maree M.
    Al-Qasem R.
    Tantour B.
    International Journal of Information Technology, 2024, 16 (1) : 551 - 558