Learning to match patients to clinical trials using large language models

被引:2
|
作者
Rybinski, Maciej [1 ]
Kusa, Wojciech [2 ]
Karimi, Sarvnaz [1 ]
Hanbury, Allan [2 ]
机构
[1] CSIRO Data61, 26 Pembroke Rd, Marsfield, NSW 2122, Australia
[2] TU Wien, Favoritenstr 9-11, A-1040 Vienna, Austria
基金
欧盟地平线“2020”;
关键词
Clinical trials; Patient to trials matching; TCRR; TREC CT; Large language models; Information retrieval; Learning-to-rank;
D O I
10.1016/j.jbi.2024.104734
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Objective: This study investigates the use of Large Language Models (LLMs) for matching patients to clinical trials (CTs) within an information retrieval pipeline. Our objective is to enhance the process of patient-trial matching by leveraging the semantic processing capabilities of LLMs, thereby improving the effectiveness of patient recruitment for clinical trials. Methods: We employed a multi-stage retrieval pipeline integrating various methodologies, including BM25 and Transformer-based rankers, along with LLM-based methods. Our primary datasets were the TREC Clinical Trials 2021-23 track collections. We compared LLM-based approaches, focusing on methods that leverage LLMs in query formulation, filtering, relevance ranking, and re-ranking of CTs. Results: Our results indicate that LLM-based systems, particularly those involving re-ranking with a fine-tuned LLM, outperform traditional methods in terms of nDCG and Precision measures. The study demonstrates that fine-tuning LLMs enhances their ability to find eligible trials. Moreover, our LLM-based approach is competitive with state-of-the-art systems in the TREC challenges. The study shows the effectiveness of LLMs in CT matching, highlighting their potential in handling complex semantic analysis and improving patient-trial matching. However, the use of LLMs increases the computational cost and reduces efficiency. We provide a detailed analysis of effectiveness-efficiency trade-offs. Conclusion: This research demonstrates the promising role of LLMs in enhancing the patient-to-clinical trial matching process, offering a significant advancement in the automation of patient recruitment. Future work should explore optimising the balance between computational cost and retrieval effectiveness in practical applications.
引用
收藏
页数:12
相关论文
共 50 条
  • [21] USING LARGE LANGUAGE MODELS TO ANNOTATE SUBSTANCE USE BEHAVIOR FROM ICU PATIENTS' CLINICAL NOTES
    Mathur, Piyush
    Maslinksi, Julia
    Dirosa, Izabella
    Cohen, Anabelle
    Arshad, Hajra
    Mahapatra, Dwarikanath
    Mishra, Shreya
    Awasthi, Raghav
    CRITICAL CARE MEDICINE, 2025, 53 (01)
  • [22] Leveraging machine learning technology to efficiently identify and match patients for precision oncology clinical trials.
    Sachse, Laura
    Dasari, Smriti
    Ackermann, Marc
    Patnaude, Emily
    OLeary, Stephanie
    Al-Olimat, Hussein
    Grigorenko, Alexander
    Stewart, Andrew
    Ward, AnnaJane
    Darmofal, Annie
    Ballakur, Sowmya
    Bennett, William
    Franzen, Amy
    Blau, Sibel
    Chandra, Abhinav Binod
    Nikolinakos, Petros
    Orsini, James Michael
    Peguero, Julio Antonio
    Blackwell, Kimberly L.
    Cooney, Matthew M.
    JOURNAL OF CLINICAL ONCOLOGY, 2021, 39 (15)
  • [23] Federated and edge learning for large language models
    Piccialli, Francesco
    Chiaro, Diletta
    Qi, Pian
    Bellandi, Valerio
    Damiani, Ernesto
    INFORMATION FUSION, 2025, 117
  • [24] Tool learning with large language models: a survey
    Qu, Changle
    Dai, Sunhao
    Wei, Xiaochi
    Cai, Hengyi
    Wang, Shuaiqiang
    Yin, Dawei
    Xu, Jun
    Wen, Ji-rong
    FRONTIERS OF COMPUTER SCIENCE, 2025, 19 (08)
  • [25] An Investigation of Applying Large Language Models to Spoken Language Learning
    Gao, Yingming
    Nuchged, Baorian
    Li, Ya
    Peng, Linkai
    APPLIED SCIENCES-BASEL, 2024, 14 (01):
  • [26] Large Language Models Demonstrate the Potential of Statistical Learning in Language
    Contreras Kallens, Pablo
    Kristensen-McLachlan, Ross Deans
    Christiansen, Morten H.
    COGNITIVE SCIENCE, 2023, 47 (03) : e13256
  • [27] Shortcut Learning of Large Language Models in Natural Language Understanding
    Du, Mengnan
    He, Fengxiang
    Zou, Na
    Tao, Dacheng
    Hu, Xia
    COMMUNICATIONS OF THE ACM, 2024, 67 (01) : 110 - 120
  • [28] Distillation Matters: Empowering Sequential Recommenders to Match the Performance of Large Language Models
    Cui, Yu
    Liu, Feng
    Wang, Pengbo
    Wang, Bohao
    Tang, Heng
    Wan, Yi
    Wang, Jun
    Chen, Jiawei
    PROCEEDINGS OF THE EIGHTEENTH ACM CONFERENCE ON RECOMMENDER SYSTEMS, RECSYS 2024, 2024, : 507 - 517
  • [29] Demo: Accelerating Patient Screening for Clinical Trials using Large Language Model Prompting
    Gopeekrishnan, Anand
    Arif, Shibbir Ahmed
    Liu, Hao
    2024 IEEE/ACM CONFERENCE ON CONNECTED HEALTH: APPLICATIONS, SYSTEMS AND ENGINEERING TECHNOLOGIES, CHASE 2024, 2024, : 214 - 215
  • [30] Using large language models in psychology
    Demszky, Dorottya
    Yang, Diyi
    Yeager, David
    Bryan, Christopher
    Clapper, Margarett
    Chandhok, Susannah
    Eichstaedt, Johannes
    Hecht, Cameron
    Jamieson, Jeremy
    Johnson, Meghann
    Jones, Michaela
    Krettek-Cobb, Danielle
    Lai, Leslie
    Jonesmitchell, Nirel
    Ong, Desmond
    Dweck, Carol
    Gross, James
    Pennebaker, James
    NATURE REVIEWS PSYCHOLOGY, 2023, 2 (11): : 688 - 701