Balancing Speciality and Versatility: a Coarse to Fine Framework for Supervised Fine-tuning Large Language Model

被引:0
|
作者
Zhang, Hengyuan [1 ]
Wu, Yanru [1 ]
Li, Dawei [3 ]
Yang, Sak
Zhao, Rui [2 ]
Jiang, Yong [1 ]
Tan, Fei [2 ]
机构
[1] Tsinghua Univ, Beijing, Peoples R China
[2] SenseTime Res, Beijing, Peoples R China
[3] Univ Calif San Diego, San Diego, CA USA
关键词
D O I
暂无
中图分类号
学科分类号
摘要
Aligned Large Language Models (LLMs) showcase remarkable versatility, capable of handling diverse real-world tasks. Meanwhile, aligned LLMs are also expected to exhibit speciality, excelling in specific applications. However, fine-tuning with extra data, a common practice to gain speciality, often leads to catastrophic forgetting (CF) of previously acquired versatility, hindering the model's performance across diverse tasks. In response to this challenge, we propose CoFiTune, a coarse to fine framework in an attempt to strike the balance between speciality and versatility. At the coarse-grained level, an empirical tree-search algorithm is utilized to pinpoint and update specific modules that are crucial for speciality, while keeping other parameters frozen; at the fine-grained level, a soft-masking mechanism regulates the update to the LLMs, mitigating the CF issue without compromising speciality. In an overall evaluation of both speciality and versatility, CoFiTune consistently outperforms baseline methods across diverse tasks and model scales. When compared to the full-parameter SFT, CoFiTune offers an average versatility improvement of 14%, while only incurring a marginal loss in speciality. Lastly, based on further analysis, we provide a speculative insight into the information forwarding process in LLMs, which helps explain the effectiveness of the proposed method. The code is available at https: //github.com/rattlesnakey/CoFiTune.
引用
收藏
页码:7467 / 7509
页数:43
相关论文
共 50 条
  • [21] Fine-Tuning Large Language Model Based Explainable Recommendation with Explainable Quality Reward
    Yang, Mengyuan
    Zhu, Mengying
    Wang, Yan
    Chen, Linxun
    Zhao, Yilei
    Wang, Xiuyuan
    Han, Bing
    Zheng, Xiaolin
    Yin, Jianwei
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 8, 2024, : 9250 - 9259
  • [22] Multi-phase Fine-Tuning: A New Fine-Tuning Approach for Sign Language Recognition
    Sarhan, Noha
    Lauri, Mikko
    Frintrop, Simone
    KUNSTLICHE INTELLIGENZ, 2022, 36 (01): : 91 - 98
  • [23] Drive as Veteran: Fine-tuning of an Onboard Large Language Model for Highway Autonomous Driving
    Wang, Yujin
    Huang, Zhaoyan
    Liu, Quanfeng
    Zheng, Yutong
    Hong, Jinlong
    Chen, Junyi
    Xiong, Lu
    Gao, Bingzhao
    Chen, Hong
    2024 35TH IEEE INTELLIGENT VEHICLES SYMPOSIUM, IEEE IV 2024, 2024, : 502 - 508
  • [24] Evaluating the Effectiveness of Fine-Tuning Large Language Model for Domain-Specific Task
    Dabhi, Saumya
    Martinez, Joseph
    Poursardar, Faryaneh
    2024 IEEE INTERNATIONAL CONFERENCE ON INFORMATION REUSE AND INTEGRATION FOR DATA SCIENCE, IRI 2024, 2024, : 176 - 177
  • [25] Multi-phase Fine-Tuning: A New Fine-Tuning Approach for Sign Language Recognition
    Noha Sarhan
    Mikko Lauri
    Simone Frintrop
    KI - Künstliche Intelligenz, 2022, 36 : 91 - 98
  • [26] Fine-tuning large neural language models for biomedical natural language processing
    Tinn, Robert
    Cheng, Hao
    Gu, Yu
    Usuyama, Naoto
    Liu, Xiaodong
    Naumann, Tristan
    Gao, Jianfeng
    Poon, Hoifung
    PATTERNS, 2023, 4 (04):
  • [27] Selecting Informative Contexts Improves Language Model Fine-tuning
    Antonello, Richard
    Beckage, Nicole M.
    Turek, Javier S.
    Huth, Alexander G.
    59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING, VOL 1 (ACL-IJCNLP 2021), 2021, : 1072 - 1085
  • [28] On Transferability of Bias Mitigation Effects in Language Model Fine-Tuning
    Jin, Xisen
    Barbieri, Francesco
    Kennedy, Brendan
    Davani, Aida Mostafazadeh
    Neves, Leonardo
    Ren, Xiang
    2021 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL-HLT 2021), 2021, : 3770 - 3783
  • [29] SFMD: A Semi-supervised Framework for Pre-trained Language Models Fine-Tuning with Noisy Samples
    Yang, Yiwen
    Duan, Pengfei
    Li, Yongbing
    Zhang, Yifang
    Xiong, Shengwu
    ADVANCED INTELLIGENT COMPUTING TECHNOLOGY AND APPLICATIONS, PT III, ICIC 2024, 2024, 14877 : 316 - 328
  • [30] Balancing Differential Privacy and Utility: A Relevance-Based Adaptive Private Fine-Tuning Framework for Language Models
    Wang, Naiyu
    Wang, Shen
    Li, Meng
    Wu, Longfei
    Zhang, Zijian
    Guan, Zhitao
    Zhu, Liehuang
    IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2025, 20 : 207 - 220