Large language models leverage external knowledge to extend clinical insight beyond language boundaries

被引:1
|
作者
Wu, Jiageng [1 ]
Wu, Xian [2 ]
Qiu, Zhaopeng [2 ]
Li, Minghui
Lin, Shixu [1 ]
Zhang, Yingying [2 ]
Zheng, Yefeng [2 ]
Yuan, Changzheng [1 ,3 ,5 ]
Yang, Jie [1 ,4 ,6 ,7 ]
机构
[1] Zhejiang Univ, Sch Med, Sch Publ Hlth, Hangzhou 310058, Peoples R China
[2] Tencent YouTu Lab, Jarvis Res Ctr, 1 Tianchen East Rd, Beijing 100101, Peoples R China
[3] Harvard TH Chan Sch Publ Hlth, Dept Nutr, Boston, MA 02115 USA
[4] Harvard Med Sch, Brigham & Womens Hosp, Dept Med, Div Pharmacoepidemiol & Pharmacoecon, Boston, MA 02115 USA
[5] Zhejiang Univ, Sch Publ Hlth, 866 Yuhangtang Rd, Hangzhou, Zhejiang, Peoples R China
[6] Brigham & Womens Hosp, Dept Med, 75 Francis St, Boston, MA 02115 USA
[7] Harvard Med Sch, 75 Francis St, Boston, MA 02115 USA
关键词
large language models; clinical knowledge; natural language processing; medical examination;
D O I
10.1093/jamia/ocae079
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Objectives Large Language Models (LLMs) such as ChatGPT and Med-PaLM have excelled in various medical question-answering tasks. However, these English-centric models encounter challenges in non-English clinical settings, primarily due to limited clinical knowledge in respective languages, a consequence of imbalanced training corpora. We systematically evaluate LLMs in the Chinese medical context and develop a novel in-context learning framework to enhance their performance.Materials and Methods The latest China National Medical Licensing Examination (CNMLE-2022) served as the benchmark. We collected 53 medical books and 381 149 medical questions to construct the medical knowledge base and question bank. The proposed Knowledge and Few-shot Enhancement In-context Learning (KFE) framework leverages the in-context learning ability of LLMs to integrate diverse external clinical knowledge sources. We evaluated KFE with ChatGPT (GPT-3.5), GPT-4, Baichuan2-7B, Baichuan2-13B, and QWEN-72B in CNMLE-2022 and further investigated the effectiveness of different pathways for incorporating LLMs with medical knowledge from 7 distinct perspectives.Results Directly applying ChatGPT failed to qualify for the CNMLE-2022 at a score of 51. Cooperated with the KFE framework, the LLMs with varying sizes yielded consistent and significant improvements. The ChatGPT's performance surged to 70.04 and GPT-4 achieved the highest score of 82.59. This surpasses the qualification threshold (60) and exceeds the average human score of 68.70, affirming the effectiveness and robustness of the framework. It also enabled a smaller Baichuan2-13B to pass the examination, showcasing the great potential in low-resource settings.Discussion and Conclusion This study shed light on the optimal practices to enhance the capabilities of LLMs in non-English medical scenarios. By synergizing medical knowledge through in-context learning, LLMs can extend clinical insight beyond language barriers in healthcare, significantly reducing language-related disparities of LLM applications and ensuring global benefit in this field.
引用
收藏
页码:2054 / 2064
页数:11
相关论文
共 50 条
  • [41] From text to insight: large language models for chemical data extraction
    Schilling-Wilhelmi, Mara
    Rios-Garcia, Martino
    Shabih, Sherjeel
    Gil, Maria Victoria
    Miret, Santiago
    Koch, Christoph T.
    Marquez, Jose A.
    Jablonka, Kevin Maik
    CHEMICAL SOCIETY REVIEWS, 2025, 54 (03) : 1125 - 1150
  • [42] Large Language Models
    Vargas, Diego Collarana
    Katsamanis, Nassos
    ERCIM NEWS, 2024, (136): : 12 - 13
  • [43] Large Language Models
    Cerf, Vinton G.
    COMMUNICATIONS OF THE ACM, 2023, 66 (08) : 7 - 7
  • [44] How Proficient Are Large Language Models in Formal Languages? An In-Depth Insight for Knowledge Base Question Answering
    Liu, Jinxin
    Cao, Shulin
    Shi, Jiaxin
    Zhang, Tingjian
    Nie, Lunyiu
    Liu, Linmei
    Hou, Lei
    Li, Juanzi
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 792 - 815
  • [45] Swahili Beyond the Boundaries: Literature, Language, and Identity
    Ntarangwi, Mwenda
    AFRICAN STUDIES REVIEW, 2009, 52 (01) : 220 - 222
  • [46] Swahili Beyond the Boundaries: literature, language and identity
    Becker, Felicitas
    AFRICA, 2008, 78 (04): : 621 - 623
  • [47] The Life of "Alan Ford" beyond Language Boundaries
    Rocco, Goranka
    Scukanec, Aleksandra
    SLAVICA TERGESTINA, 2021, 27 : 406 - 427
  • [48] Swahili Beyond the Boundaries: Literature, Language, Identity
    Rettova, Alena
    BULLETIN OF THE SCHOOL OF ORIENTAL AND AFRICAN STUDIES-UNIVERSITY OF LONDON, 2008, 71 : 608 - 609
  • [49] Swahili Beyond the Boundaries: Language, Literature and Identity
    Waliaula, Kennedy
    CANADIAN JOURNAL OF AFRICAN STUDIES, 2008, 42 (01) : 186 - 188
  • [50] Workshop on Enterprise Knowledge Graphs using Large Language Models
    Gupta, Rajeev
    Srinivasa, Srinath
    PROCEEDINGS OF THE 32ND ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2023, 2023, : 5271 - 5272