Balancing Speciality and Versatility: a Coarse to Fine Framework for Supervised Fine-tuning Large Language Model

被引:0
|
作者
Zhang, Hengyuan [1 ]
Wu, Yanru [1 ]
Li, Dawei [3 ]
Yang, Sak
Zhao, Rui [2 ]
Jiang, Yong [1 ]
Tan, Fei [2 ]
机构
[1] Tsinghua Univ, Beijing, Peoples R China
[2] SenseTime Res, Beijing, Peoples R China
[3] Univ Calif San Diego, San Diego, CA USA
关键词
D O I
暂无
中图分类号
学科分类号
摘要
Aligned Large Language Models (LLMs) showcase remarkable versatility, capable of handling diverse real-world tasks. Meanwhile, aligned LLMs are also expected to exhibit speciality, excelling in specific applications. However, fine-tuning with extra data, a common practice to gain speciality, often leads to catastrophic forgetting (CF) of previously acquired versatility, hindering the model's performance across diverse tasks. In response to this challenge, we propose CoFiTune, a coarse to fine framework in an attempt to strike the balance between speciality and versatility. At the coarse-grained level, an empirical tree-search algorithm is utilized to pinpoint and update specific modules that are crucial for speciality, while keeping other parameters frozen; at the fine-grained level, a soft-masking mechanism regulates the update to the LLMs, mitigating the CF issue without compromising speciality. In an overall evaluation of both speciality and versatility, CoFiTune consistently outperforms baseline methods across diverse tasks and model scales. When compared to the full-parameter SFT, CoFiTune offers an average versatility improvement of 14%, while only incurring a marginal loss in speciality. Lastly, based on further analysis, we provide a speculative insight into the information forwarding process in LLMs, which helps explain the effectiveness of the proposed method. The code is available at https: //github.com/rattlesnakey/CoFiTune.
引用
收藏
页码:7467 / 7509
页数:43
相关论文
共 50 条
  • [1] Comprehensive Review of Large Language Model Fine-Tuning
    Zhang, Qintong
    Wang, Yuchao
    Wang, Hexi
    Wang, Junxin
    Chen, Hai
    Computer Engineering and Applications, 2024, 60 (17) : 17 - 33
  • [2] Enhancing AAV Viability Prediction: A Generalizable Fine-Tuning Framework with Large Language Model xTrimoPGLM
    Yang, Qirong
    Zou, Diming
    Guo, Yucheng
    Xu, Chenrui
    Marsic, Damien
    Cai, Zhongshan
    Liu, Yawen
    Xu, Ziyao
    Qu, Vicky
    Garces, Fernando
    Greisen, Per
    Ji, Qingzhou
    Song, Le
    MOLECULAR THERAPY, 2024, 32 (04) : 693 - 694
  • [3] Selective privacy-preserving framework for large language models fine-tuning
    Wang, Teng
    Zhai, Lindong
    Yang, Tengfei
    Luo, Zhucheng
    Liu, Shuanggen
    INFORMATION SCIENCES, 2024, 678
  • [4] Phased Instruction Fine-Tuning for Large Language Models
    Pang, Wei
    Zhou, Chuan
    Zhou, Xiao-Hua
    Wang, Xiaojie
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 5735 - 5748
  • [5] HackMentor: Fine-Tuning Large Language Models for Cybersecurity
    Zhang, Jie
    Wen, Hui
    Deng, Liting
    Xin, Mingfeng
    Li, Zhi
    Li, Lun
    Zhu, Hongsong
    Sun, Limin
    2023 IEEE 22ND INTERNATIONAL CONFERENCE ON TRUST, SECURITY AND PRIVACY IN COMPUTING AND COMMUNICATIONS, TRUSTCOM, BIGDATASE, CSE, EUC, ISCI 2023, 2024, : 452 - 461
  • [6] WalkLM: A Uniform Language Model Fine-tuning Framework for Attributed Graph Embedding
    Tan, Yanchao
    Zhou, Zihao
    Lv, Hang
    Liu, Weiming
    Yang, Carl
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [7] Fine-Tuning Language Models For Semi-Supervised Text Mining
    Chen, Xinyu
    Beaver, Ian
    Freeman, Cynthia
    2020 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2020, : 3608 - 3617
  • [8] Efficient fine-tuning of short text classification based on large language model
    Wang, Likun
    PROCEEDINGS OF INTERNATIONAL CONFERENCE ON MODELING, NATURAL LANGUAGE PROCESSING AND MACHINE LEARNING, CMNM 2024, 2024, : 33 - 38
  • [9] Fine-Tuning a Large Language Model with Reinforcement Learning for Educational Question Generation
    Lamsiyah, Salima
    El Mahdaouy, Abdelkader
    Nourbakhsh, Aria
    Schommer, Christoph
    ARTIFICIAL INTELLIGENCE IN EDUCATION, PT I, AIED 2024, 2024, 14829 : 424 - 438
  • [10] Raise a Child in Large Language Model: Towards Effective and Generalizable Fine-tuning
    Xu, Runxin
    Luo, Fuli
    Zhang, Zhiyuan
    Tan, Chuanqi
    Chang, Baobao
    Huang, Songfang
    Huang, Fei
    2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 9514 - 9528