Balancing Speciality and Versatility: a Coarse to Fine Framework for Supervised Fine-tuning Large Language Model

被引:0
|
作者
Zhang, Hengyuan [1 ]
Wu, Yanru [1 ]
Li, Dawei [3 ]
Yang, Sak
Zhao, Rui [2 ]
Jiang, Yong [1 ]
Tan, Fei [2 ]
机构
[1] Tsinghua Univ, Beijing, Peoples R China
[2] SenseTime Res, Beijing, Peoples R China
[3] Univ Calif San Diego, San Diego, CA USA
关键词
D O I
暂无
中图分类号
学科分类号
摘要
Aligned Large Language Models (LLMs) showcase remarkable versatility, capable of handling diverse real-world tasks. Meanwhile, aligned LLMs are also expected to exhibit speciality, excelling in specific applications. However, fine-tuning with extra data, a common practice to gain speciality, often leads to catastrophic forgetting (CF) of previously acquired versatility, hindering the model's performance across diverse tasks. In response to this challenge, we propose CoFiTune, a coarse to fine framework in an attempt to strike the balance between speciality and versatility. At the coarse-grained level, an empirical tree-search algorithm is utilized to pinpoint and update specific modules that are crucial for speciality, while keeping other parameters frozen; at the fine-grained level, a soft-masking mechanism regulates the update to the LLMs, mitigating the CF issue without compromising speciality. In an overall evaluation of both speciality and versatility, CoFiTune consistently outperforms baseline methods across diverse tasks and model scales. When compared to the full-parameter SFT, CoFiTune offers an average versatility improvement of 14%, while only incurring a marginal loss in speciality. Lastly, based on further analysis, we provide a speculative insight into the information forwarding process in LLMs, which helps explain the effectiveness of the proposed method. The code is available at https: //github.com/rattlesnakey/CoFiTune.
引用
收藏
页码:7467 / 7509
页数:43
相关论文
共 50 条
  • [41] Leveraging Large Language Models Knowledge Enhancement Dual-Stage Fine-Tuning Framework for Recommendation
    Zeng, Biqing
    Shi, Hao
    Li, Yangyu
    Li, Ruizhe
    Deng, Huimin
    NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING, PT II, NLPCC 2024, 2025, 15360 : 333 - 345
  • [42] CRaSh: Clustering, Removing, and Sharing Enhance Fine-tuning without Full Large Language Model
    Zhang, Kaiyan
    Ding, Ning
    Qi, Biqing
    Zhu, Xuekai
    Long, Xinwei
    Zhou, Bowen
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2023), 2023, : 9612 - 9637
  • [43] Framework welcome, but could do with fine-tuning
    Frédéric Sgard
    Nature, 2001, 412 : 375 - 375
  • [44] Guided Recommendation for Model Fine-Tuning
    Li, Hao
    Fowlkes, Charless
    Yang, Hao
    Dabeer, Onkar
    Tu, Zhuowen
    Soatto, Stefano
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR, 2023, : 3633 - 3642
  • [45] Model Editing by Standard Fine-Tuning
    Gangadhar, Govind
    Stratos, Karl
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 5907 - 5913
  • [46] Enhancing Code Language Models for Program Repair by Curricular Fine-tuning Framework
    Hao, Sichong
    Shi, Xianjun
    Liu, Hongwei
    Shu, Yanjun
    2023 IEEE INTERNATIONAL CONFERENCE ON SOFTWARE MAINTENANCE AND EVOLUTION, ICSME, 2023, : 136 - 146
  • [47] CSS-LM: A Contrastive Framework for Semi-Supervised Fine-Tuning of Pre-Trained Language Models
    Su, Yusheng
    Han, Xu
    Lin, Yankai
    Zhang, Zhengyan
    Liu, Zhiyuan
    Li, Peng
    Zhou, Jie
    Sun, Maosong
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2021, 29 : 2930 - 2941
  • [48] MediBioDeBERTa: Biomedical Language Model With Continuous Learning and Intermediate Fine-Tuning
    Kim, Eunhui
    Jeong, Yuna
    Choi, Myung-Seok
    IEEE ACCESS, 2023, 11 : 141036 - 141044
  • [49] Improving Universal Language Model Fine-Tuning using Attention Mechanism
    Santos, Flavio A. O.
    Ponce-Guevara, K. L.
    Macedo, David
    Zanchettin, Cleber
    2019 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2019,
  • [50] MultiFiT: Efficient Multi-lingual Language Model Fine-tuning
    Eisenschlos, Julian
    Ruder, Sebastian
    Czapla, Piotr
    Kardas, Marcin
    Gugger, Sylvain
    Howard, Jeremy
    2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019): PROCEEDINGS OF THE CONFERENCE, 2019, : 5702 - 5707