Exploring the effectiveness of instruction tuning in biomedical language processing

被引:1
|
作者
Rohanian, Omid [1 ,2 ]
Nouriborji, Mohammadmahdi [2 ,3 ]
Kouchaki, Samaneh [4 ]
Nooralahzadeh, Farhad [5 ,6 ]
Clifton, Lei [7 ]
Clifton, David A. [1 ,8 ]
机构
[1] Univ Oxford, Dept Engn Sci, Oxford, England
[2] NLPie Res, Oxford, England
[3] Sharif Univ Technol, Tehran, Iran
[4] Univ Surrey, Dept Elect & Elect Engn, Guildford, England
[5] Univ Zurich, Zurich, Switzerland
[6] Univ Hosp Zurich, Zurich, Switzerland
[7] Univ Oxford, Nuffield Dept Populat Hlth, Oxford, England
[8] Oxford Suzhou Ctr Adv Res, Suzhou, Peoples R China
关键词
Instruction tuning; Biomedical NLP; Named entity recognition; Relation extraction; Medical NLI; Llama2-MedTuned;
D O I
10.1016/j.artmed.2024.103007
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Large Language Models (LLMs), particularly those similar to ChatGPT, have significantly influenced the field of Natural Language Processing (NLP). While these models excel in general language tasks, their performance in domain-specific downstream tasks such as biomedical and clinical Named Entity Recognition (NER), Relation Extraction (RE), and Medical Natural Language Inference (NLI) is still evolving. In this context, our study investigates the potential of instruction tuning for biomedical language processing, applying this technique to two general LLMs of substantial scale. We present a comprehensive, instruction-based model trained on a dataset that consists of approximately 200,000 instruction-focused samples. This dataset represents a carefully curated compilation of existing data, meticulously adapted and reformatted to align with the specific requirements of our instruction-based tasks. This initiative represents an important step in utilising such models to achieve results on par with specialised encoder-only models like BioBERT and BioClinicalBERT for various classical biomedical NLP tasks. Our work includes an analysis of the dataset's composition and its impact on model performance, providing insights into the intricacies of instruction tuning. By sharing our codes, models, and the distinctively assembled instruction-based dataset, we seek to encourage ongoing research and development in this area.2
引用
收藏
页数:9
相关论文
共 50 条
  • [1] BioInstruct: instruction tuning of large language models for biomedical natural language processing
    Tran, Hieu
    Yang, Zhichao
    Yao, Zonghai
    Yu, Hong
    JOURNAL OF THE AMERICAN MEDICAL INFORMATICS ASSOCIATION, 2024, 31 (09) : 1821 - 1832
  • [2] Exploring the Benefits of Training Expert Language Models over Instruction Tuning
    Jang, Joel
    Kim, Seungone
    Ye, Seonghyeon
    Kim, Doyoung
    Logeswaran, Lajanugen
    Lee, Moontae
    Lee, Kyungjae
    Seo, Minjoon
    Proceedings of Machine Learning Research, 2023, 202 : 14702 - 14729
  • [3] Exploring the Benefits of Training Expert Language Models over Instruction Tuning
    Jang, Joel
    Kim, Seungone
    Ye, Seonghyeon
    Kim, Doyoung
    Logeswaran, Lajanugen
    Lee, Moontae
    Lee, Kyungjae
    Seo, Minjoon
    arXiv, 2023,
  • [4] Fine-tuning large neural language models for biomedical natural language processing
    Tinn, Robert
    Cheng, Hao
    Gu, Yu
    Usuyama, Naoto
    Liu, Xiaodong
    Naumann, Tristan
    Gao, Jianfeng
    Poon, Hoifung
    PATTERNS, 2023, 4 (04):
  • [5] THE ROLE OF EXPLICIT GRAMMAR IN SECOND LANGUAGE TEACHING: INSTRUCTIONAL EFFECTIVENESS OF PROCESSING INSTRUCTION
    Nunez Delgado, Maria Pilar
    Alonso Aparicio, Irene
    RIVISTA DI PSICOLINGUISTICA APPLICATA-JOURNAL OF APPLIED PSYCHOLINGUISTICS, 2009, 9 (1-2): : 93 - 108
  • [6] Biomedical Natural Language Processing
    Kim, Jin-Dong
    COMPUTATIONAL LINGUISTICS, 2017, 43 (01) : 265 - 267
  • [7] Biomedical Natural Language Processing
    Hamon, Thierry
    TRAITEMENT AUTOMATIQUE DES LANGUES, 2013, 54 (03): : 77 - 79
  • [8] Exploring Effectiveness and Moderators of Language Learning Strategy Instruction on Second Language and Self-Regulated Learning Outcomes
    Ardasheva, Yuliya
    Wang, Zhe
    Adesope, Olusola O.
    Valentine, Jeffrey C.
    REVIEW OF EDUCATIONAL RESEARCH, 2017, 87 (03) : 544 - 582
  • [9] Exploring subdomain variation in biomedical language
    Thomas Lippincott
    Diarmuid Ó Séaghdha
    Anna Korhonen
    BMC Bioinformatics, 12
  • [10] Exploring subdomain variation in biomedical language
    Lippincott, Thomas
    Seaghdha, Diarmuid O.
    Korhonen, Anna
    BMC BIOINFORMATICS, 2011, 12