Enhancing Task Performance in Continual Instruction Fine-tuning Through Format Uniformity

被引:0
|
作者
Tan, Xiaoyu [1 ]
Cheng, Leijun [2 ]
Qiu, Xihe [2 ]
Shi, Shaojie [2 ]
Cheng, Yuan [3 ]
Chu, Wei [1 ]
Xu, Yinghui [3 ]
Qi, Yuan [3 ]
机构
[1] INF Technol Shanghai Co Ltd, Shanghai, Peoples R China
[2] Shanghai Univ Engn Sci, Sch Elect & Elect Engn, Shanghai, Peoples R China
[3] Fudan Univ, AI3 Inst, Shanghai, Peoples R China
基金
中国国家自然科学基金;
关键词
Large Language Models; Continual Instruction Fine-tuning; Format Uniformity; Catastrophic Forgetting;
D O I
10.1145/3626772.3657920
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In recent advancements, large language models (LLMs) have demonstrated remarkable capabilities in diverse tasks, primarily through interactive question-answering with humans. This development marks significant progress towards artificial general intelligence (AGI). Despite their superior performance, LLMs often exhibit limitations when adapted to domain-specific tasks through instruction fine-tuning (IF). The primary challenge lies in the discrepancy between the data distribution in general and domain-specific contexts, leading to suboptimal accuracy in specialized tasks. To address this, continual instruction fine-tuning (CIF), particularly supervised finetuning (SFT), on targeted domain-specific instruction datasets is necessary. Our ablation study reveals that the structure of these instruction datasets critically influences CIF performance, with substantial data distributional shifts resulting in notable performance degradation. In this paper, we introduce a novel framework that enhances CIF by promoting format uniformity. We assess our approach using the Llama2 chat model across various domain-specific instruction datasets. The results demonstrate not only an improvement in task-specific performance under CIF but also a reduction in catastrophic forgetting (CF). This study contributes to the optimization of LLMs for domain-specific applications, highlighting the significance of data structure and distribution in CIF.
引用
收藏
页码:2384 / 2389
页数:6
相关论文
共 50 条
  • [1] Alleviating Representational Shift for Continual Fine-tuning
    Jie, Shibo
    Deng, Zhi-Hong
    Li, Ziheng
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS, CVPRW 2022, 2022, : 3809 - 3818
  • [2] Chain-of-LoRA: Enhancing the Instruction Fine-Tuning Performance of Low-Rank Adaptation on Diverse Instruction Set
    Qiu, Xihe
    Hao, Teqi
    Shi, Shaojie
    Tan, Xiaoyu
    Xiong, Yu-Jie
    IEEE SIGNAL PROCESSING LETTERS, 2024, 31 : 875 - 879
  • [3] MotIF: Motion Instruction Fine-Tuning
    Hwang, Minyoung
    Hejna, Joey
    Sadigh, Dorsa
    Bisk, Yonatan
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2025, 10 (03): : 2287 - 2294
  • [4] Tangent Model Composition for Ensembling and Continual Fine-tuning
    Liu, Tian Yu
    Soatto, Stefano
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 18630 - 18640
  • [5] Kaizen: Practical self-supervised continual learning with continual fine-tuning
    Tang, Chi Ian
    Qendrol, Lorena
    Spathis, Dimitris
    Kawsar, Fahim
    Mascolo, Cecilia
    Mathur, Akhil
    2024 IEEE/CVF WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION, WACV 2024, 2024, : 2829 - 2838
  • [6] Learning Task-Specific Initialization for Effective Federated Continual Fine-Tuning of Foundation Model Adapters
    Peng, Danni
    Wang, Yuan
    Fu, Huazhu
    Wee, Qingsong
    Liu, Yong
    Goh, Rick Siow Mong
    2024 IEEE CONFERENCE ON ARTIFICIAL INTELLIGENCE, CAI 2024, 2024, : 811 - 816
  • [7] Phased Instruction Fine-Tuning for Large Language Models
    Pang, Wei
    Zhou, Chuan
    Zhou, Xiao-Hua
    Wang, Xiaojie
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 5735 - 5748
  • [8] DLCFT: Deep Linear Continual Fine-Tuning for General Incremental Learning
    Shon, Hyounguk
    Lee, Janghyeon
    Kim, Seung Hwan
    Kim, Junmo
    COMPUTER VISION - ECCV 2022, PT XXXIII, 2022, 13693 : 513 - 529
  • [9] Improving CLIP Fine-tuning Performance
    Wei, Yixuan
    Hu, Han
    Xie, Zhenda
    Liu, Ze
    Zhang, Zheng
    Cao, Yue
    Bao, Jianmin
    Chen, Dong
    Guo, Baining
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION, ICCV, 2023, : 5416 - 5426
  • [10] Fine-tuning CARs for best performance
    Hotblack, Alastair
    Straathof, Karin
    CANCER CELL, 2022, 40 (01) : 11 - 13