Exploring the effectiveness of instruction tuning in biomedical language processing

被引:1
|
作者
Rohanian, Omid [1 ,2 ]
Nouriborji, Mohammadmahdi [2 ,3 ]
Kouchaki, Samaneh [4 ]
Nooralahzadeh, Farhad [5 ,6 ]
Clifton, Lei [7 ]
Clifton, David A. [1 ,8 ]
机构
[1] Univ Oxford, Dept Engn Sci, Oxford, England
[2] NLPie Res, Oxford, England
[3] Sharif Univ Technol, Tehran, Iran
[4] Univ Surrey, Dept Elect & Elect Engn, Guildford, England
[5] Univ Zurich, Zurich, Switzerland
[6] Univ Hosp Zurich, Zurich, Switzerland
[7] Univ Oxford, Nuffield Dept Populat Hlth, Oxford, England
[8] Oxford Suzhou Ctr Adv Res, Suzhou, Peoples R China
关键词
Instruction tuning; Biomedical NLP; Named entity recognition; Relation extraction; Medical NLI; Llama2-MedTuned;
D O I
10.1016/j.artmed.2024.103007
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Large Language Models (LLMs), particularly those similar to ChatGPT, have significantly influenced the field of Natural Language Processing (NLP). While these models excel in general language tasks, their performance in domain-specific downstream tasks such as biomedical and clinical Named Entity Recognition (NER), Relation Extraction (RE), and Medical Natural Language Inference (NLI) is still evolving. In this context, our study investigates the potential of instruction tuning for biomedical language processing, applying this technique to two general LLMs of substantial scale. We present a comprehensive, instruction-based model trained on a dataset that consists of approximately 200,000 instruction-focused samples. This dataset represents a carefully curated compilation of existing data, meticulously adapted and reformatted to align with the specific requirements of our instruction-based tasks. This initiative represents an important step in utilising such models to achieve results on par with specialised encoder-only models like BioBERT and BioClinicalBERT for various classical biomedical NLP tasks. Our work includes an analysis of the dataset's composition and its impact on model performance, providing insights into the intricacies of instruction tuning. By sharing our codes, models, and the distinctively assembled instruction-based dataset, we seek to encourage ongoing research and development in this area.2
引用
收藏
页数:9
相关论文
共 50 条
  • [21] Cheap and Quick: Efficient Vision-Language Instruction Tuning for Large Language Models
    Luo, Gen
    Zhou, Yiyi
    Ren, Tianhe
    Chen, Shengxin
    Sun, Xiaoshuai
    Ji, Rongrong
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [22] On Transferability of Prompt Tuning for Natural Language Processing
    Su, Yusheng
    Wang, Xiaozhi
    Qin, Yujia
    Chan, Chi-Min
    Lin, Yankai
    Wang, Huadong
    Wen, Kaiyue
    Liu, Zhiyuan
    Li, Peng
    Li, Juanzi
    Hou, Lei
    Sun, Maosong
    Zhou, Jie
    NAACL 2022: THE 2022 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES, 2022, : 3949 - 3969
  • [23] Database Tuning using Natural Language Processing
    Trummer, Immanuel
    SIGMOD RECORD, 2021, 50 (03) : 27 - 28
  • [24] SECOND LANGUAGE PROCESSING AND DELIVERING PROCESSING INSTRUCTION IN CLASSROOMS AND VIRTUAL CONTEXTS
    Dyson, Bronwen
    AUSTRALIAN REVIEW OF APPLIED LINGUISTICS, 2010, 33 (01)
  • [25] Exploring the amount and type of writing instruction during language arts instruction in kindergarten classrooms
    Cynthia S. Puranik
    Stephanie Al Otaiba
    Jessica Folsom Sidler
    Luana Greulich
    Reading and Writing, 2014, 27 : 213 - 236
  • [26] Exploring the amount and type of writing instruction during language arts instruction in kindergarten classrooms
    Puranik, Cynthia S.
    Al Otaiba, Stephanie
    Sidler, Jessica Folsom
    Greulich, Luana
    READING AND WRITING, 2014, 27 (02) : 213 - 236
  • [27] Input processing and grammar instruction in second language acquisition
    Jelinski, JB
    HISPANIA-A JOURNAL DEVOTED TO THE TEACHING OF SPANISH AND PORTUGUESE, 1997, 80 (04): : 811 - 812
  • [28] Processing instruction and a role for output in second language acquisition
    Toth, PD
    LANGUAGE LEARNING, 2006, 56 (02) : 319 - 385
  • [29] Cross-Language Similarity Modulates Effectiveness of Second Language Grammar Instruction
    Tolentino, Leida C.
    Tokowicz, Natasha
    LANGUAGE LEARNING, 2014, 64 (02) : 279 - 309
  • [30] Demystifying Instruction Mixing for Fine-tuning Large Language Models
    Wang, Renxi
    Li, Haonan
    Wu, Minghao
    Wang, Yuxia
    Han, Xudong
    Zhang, Chiyu
    Baldwin, Timothy
    PROCEEDINGS OF THE 62ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 4: STUDENT RESEARCH WORKSHOP, 2024, : 86 - 93