PMC-LLaMA: toward building open-source language models for medicine

被引:13
|
作者
Wu, Chaoyi [1 ,2 ]
Lin, Weixiong [1 ,2 ]
Zhang, Xiaoman [1 ,2 ]
Zhang, Ya [1 ,2 ]
Xie, Weidi [1 ,2 ]
Wang, Yanfeng [1 ,2 ]
机构
[1] Shanghai Jiao Tong Univ, Cooperat Medianet Innovat Ctr CM, Shanghai 200240, Peoples R China
[2] Shanghai AI Lab, Shanghai 200232, Peoples R China
基金
国家重点研发计划;
关键词
large language models; biomedical NLP; generative language models; ChatGPT;
D O I
10.1093/jamia/ocae045
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Objective Recently, large language models (LLMs) have showcased remarkable capabilities in natural language understanding. While demonstrating proficiency in everyday conversations and question-answering (QA) situations, these models frequently struggle in domains that require precision, such as medical applications, due to their lack of domain-specific knowledge. In this article, we describe the procedure for building a powerful, open-source language model specifically designed for medicine applications, termed as PMC-LLaMA.Materials and methods We adapt a general-purpose LLM toward the medical domain, involving data-centric knowledge injection through the integration of 4.8M biomedical academic papers and 30K medical textbooks, as well as comprehensive domain-specific instruction fine-tuning, encompassing medical QA, rationale for reasoning, and conversational dialogues with 202M tokens.Results While evaluating various public medical QA benchmarks and manual rating, our lightweight PMC-LLaMA, which consists of only 13B parameters, exhibits superior performance, even surpassing ChatGPT. All models, codes, and datasets for instruction tuning will be released to the research community.Discussion Our contributions are 3-fold: (1) we build up an open-source LLM toward the medical domain. We believe the proposed PMC-LLaMA model can promote further development of foundation models in medicine, serving as a medical trainable basic generative language backbone; (2) we conduct thorough ablation studies to demonstrate the effectiveness of each proposed component, demonstrating how different training data and model scales affect medical LLMs; (3) we contribute a large-scale, comprehensive dataset for instruction tuning.Conclusion In this article, we systematically investigate the process of building up an open-source medical-specific LLM, PMC-LLaMA.
引用
收藏
页码:1833 / 1843
页数:11
相关论文
共 50 条
  • [41] Evaluation of Language Runtimes in Open-source Serverless Platforms
    Djemame, Karim
    Datsev, Daniel
    Kelefouras, Vasilios
    PROCEEDINGS OF THE 12TH INTERNATIONAL CONFERENCE ON CLOUD COMPUTING AND SERVICES SCIENCE (CLOSER), 2022, : 123 - 132
  • [42] FreeSML: Delivering on the open-source simulation language promise
    DiLeo, JJ
    Proceedings of the 2005 Winter Simulation Conference, Vols 1-4, 2005, : 2513 - 2523
  • [43] The Language of Nuclear Security: Language Diversity in Open-Source Internet Searches
    Homan, Zenobia
    Shaban, Yara
    Rane, Shraddha
    INTERNATIONAL JOURNAL OF INTELLIGENCE AND COUNTERINTELLIGENCE, 2023, 36 (03) : 870 - 891
  • [44] ManPy: an open-source software tool for building discrete event simulation models of manufacturing systems
    Dagkakis, Georgios
    Papagiannopoulos, Ioannis
    Heavey, Cathal
    SOFTWARE-PRACTICE & EXPERIENCE, 2016, 46 (07): : 955 - 981
  • [45] Toward development of open-source models of decision support systems for water demand management in Egypt
    Mohamed Ghareeb
    Ahmed Khaled Seif
    Innovative Infrastructure Solutions, 2020, 5
  • [46] Toward development of open-source models of decision support systems for water demand management in Egypt
    Ghareeb, Mohamed
    Seif, Ahmed Khaled
    INNOVATIVE INFRASTRUCTURE SOLUTIONS, 2020, 5 (02)
  • [47] ArcheType: A Novel Framework for Open-Source Column Type Annotation using Large Language Models
    Feuer, Benjamin
    Liu, Yurong
    Hegde, Chinmay
    Freire, Juliana
    PROCEEDINGS OF THE VLDB ENDOWMENT, 2024, 17 (09): : 2279 - 2292
  • [48] Need of Fine-Tuned Radiology Aware Open-Source Large Language Models for Neuroradiology
    Ray, Partha Pratim
    CLINICAL NEURORADIOLOGY, 2024,
  • [49] Toponym resolution leveraging lightweight and open-source large language models and geo-knowledge
    Hu, Xuke
    Kersten, Jens
    Klan, Friederike
    Farzana, Sheikh Mastura
    INTERNATIONAL JOURNAL OF GEOGRAPHICAL INFORMATION SCIENCE, 2024,
  • [50] Open-source modelling infrastructure: Building decarbonization capacity in Canada
    McPherson, Madeleine
    Monroe, Jacob
    Jurasz, Jakub
    Rowe, Andrew
    Hendriks, Richard
    Stanislaw, Lauren
    Awais, Muhammad
    Seatle, Madeleine
    Xu, Robert
    Crownshaw, Timothy
    Miri, Mohammad
    Aldana, Dustin
    Esfahlani, Moe
    Arjmand, Reza
    Saffari, Mohammadali
    Cusi, Tristan
    Toor, Kanwarpreet Singh
    Grieco, Joel
    ENERGY STRATEGY REVIEWS, 2022, 44