Balancing Speciality and Versatility: a Coarse to Fine Framework for Supervised Fine-tuning Large Language Model

被引:0
|
作者
Zhang, Hengyuan [1 ]
Wu, Yanru [1 ]
Li, Dawei [3 ]
Yang, Sak
Zhao, Rui [2 ]
Jiang, Yong [1 ]
Tan, Fei [2 ]
机构
[1] Tsinghua Univ, Beijing, Peoples R China
[2] SenseTime Res, Beijing, Peoples R China
[3] Univ Calif San Diego, San Diego, CA USA
关键词
D O I
暂无
中图分类号
学科分类号
摘要
Aligned Large Language Models (LLMs) showcase remarkable versatility, capable of handling diverse real-world tasks. Meanwhile, aligned LLMs are also expected to exhibit speciality, excelling in specific applications. However, fine-tuning with extra data, a common practice to gain speciality, often leads to catastrophic forgetting (CF) of previously acquired versatility, hindering the model's performance across diverse tasks. In response to this challenge, we propose CoFiTune, a coarse to fine framework in an attempt to strike the balance between speciality and versatility. At the coarse-grained level, an empirical tree-search algorithm is utilized to pinpoint and update specific modules that are crucial for speciality, while keeping other parameters frozen; at the fine-grained level, a soft-masking mechanism regulates the update to the LLMs, mitigating the CF issue without compromising speciality. In an overall evaluation of both speciality and versatility, CoFiTune consistently outperforms baseline methods across diverse tasks and model scales. When compared to the full-parameter SFT, CoFiTune offers an average versatility improvement of 14%, while only incurring a marginal loss in speciality. Lastly, based on further analysis, we provide a speculative insight into the information forwarding process in LLMs, which helps explain the effectiveness of the proposed method. The code is available at https: //github.com/rattlesnakey/CoFiTune.
引用
收藏
页码:7467 / 7509
页数:43
相关论文
共 50 条
  • [31] CLEFT: Language-Image Contrastive Learning with Efficient Large Language Model and Prompt Fine-Tuning
    Dui, Yuexi
    Chang, Brian
    Dvornek, Nicha C.
    MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION - MICCAI 2024, PT XII, 2024, 15012 : 465 - 475
  • [32] Applying SoftTriple Loss for Supervised Language Model Fine Tuning
    Sosnowski, Witold
    Wroblewska, Anna
    Gawrysiak, Piotr
    PROCEEDINGS OF THE 2022 17TH CONFERENCE ON COMPUTER SCIENCE AND INTELLIGENCE SYSTEMS (FEDCSIS), 2022, : 141 - 147
  • [33] Fine-Tuning Large Language Models for Private Document Retrieval: A Tutorial
    Sommers, Frank
    Kongthon, Alisa
    Kongyoung, Sarawoot
    PROCEEDINGS OF THE 4TH ANNUAL ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA RETRIEVAL, ICMR 2024, 2024, : 1319 - 1320
  • [34] Large language models in Radiology: The importance of fine-tuning and the fable of the luthier
    Martin-Noguerol, Teodoro
    Lopez-Ubeda, Pilar
    Luna, Antonio
    EUROPEAN JOURNAL OF RADIOLOGY, 2024, 178
  • [35] Distributed Inference and Fine-tuning of Large Language Models Over The Internet
    Borzunov, Alexander
    Ryabinin, Max
    Chumachenko, Artem
    Baranchuk, Dmitry
    Dettmers, Tim
    Belkada, Younes
    Samygin, Pavel
    Raffel, Colin
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [36] Fine-tuning large language models for rare disease concept normalization
    Wang, Andy
    Liu, Cong
    Yang, Jingye
    Weng, Chunhua
    JOURNAL OF THE AMERICAN MEDICAL INFORMATICS ASSOCIATION, 2024, 31 (09) : 2076 - 2083
  • [37] Repeatability of Fine-Tuning Large Language Models Illustrated Using QLoRA
    Alahmari, Saeed S.
    Hall, Lawrence O.
    Mouton, Peter R.
    Goldgof, Dmitry B.
    IEEE ACCESS, 2024, 12 : 153221 - 153231
  • [38] Fine-Tuning Large Enterprise Language Models via Ontological Reasoning
    Baldazzi, Teodoro
    Bellomarini, Luigi
    Ceri, Stefano
    Colombo, Andrea
    Gentili, Andrea
    Sallinger, Emanuel
    RULES AND REASONING, RULEML+RR 2023, 2023, 14244 : 86 - 94
  • [39] BloomLLM: Large Language Models Based Question Generation Combining Supervised Fine-Tuning and Bloom's Taxonomy
    Nghia Duong-Trung
    Wang, Xia
    Kravcik, Milos
    TECHNOLOGY ENHANCED LEARNING FOR INCLUSIVE AND EQUITABLE QUALITY EDUCATION, PT II, EC-TEL 2024, 2024, 15160 : 93 - 98
  • [40] How fine can fine-tuning be? Learning efficient language models
    Radiya-Dixit, Evani
    Wang, Xin
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 108, 2020, 108 : 2435 - 2442