Enhancing Task Performance in Continual Instruction Fine-tuning Through Format Uniformity

被引:0
|
作者
Tan, Xiaoyu [1 ]
Cheng, Leijun [2 ]
Qiu, Xihe [2 ]
Shi, Shaojie [2 ]
Cheng, Yuan [3 ]
Chu, Wei [1 ]
Xu, Yinghui [3 ]
Qi, Yuan [3 ]
机构
[1] INF Technol Shanghai Co Ltd, Shanghai, Peoples R China
[2] Shanghai Univ Engn Sci, Sch Elect & Elect Engn, Shanghai, Peoples R China
[3] Fudan Univ, AI3 Inst, Shanghai, Peoples R China
基金
中国国家自然科学基金;
关键词
Large Language Models; Continual Instruction Fine-tuning; Format Uniformity; Catastrophic Forgetting;
D O I
10.1145/3626772.3657920
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In recent advancements, large language models (LLMs) have demonstrated remarkable capabilities in diverse tasks, primarily through interactive question-answering with humans. This development marks significant progress towards artificial general intelligence (AGI). Despite their superior performance, LLMs often exhibit limitations when adapted to domain-specific tasks through instruction fine-tuning (IF). The primary challenge lies in the discrepancy between the data distribution in general and domain-specific contexts, leading to suboptimal accuracy in specialized tasks. To address this, continual instruction fine-tuning (CIF), particularly supervised finetuning (SFT), on targeted domain-specific instruction datasets is necessary. Our ablation study reveals that the structure of these instruction datasets critically influences CIF performance, with substantial data distributional shifts resulting in notable performance degradation. In this paper, we introduce a novel framework that enhances CIF by promoting format uniformity. We assess our approach using the Llama2 chat model across various domain-specific instruction datasets. The results demonstrate not only an improvement in task-specific performance under CIF but also a reduction in catastrophic forgetting (CF). This study contributes to the optimization of LLMs for domain-specific applications, highlighting the significance of data structure and distribution in CIF.
引用
收藏
页码:2384 / 2389
页数:6
相关论文
共 50 条
  • [31] A new pipeline for generating instruction dataset via RAG and self fine-tuning
    Sung, Chih-Wei
    Lee, Yu-Kai
    Tsai, Yin-Te
    2024 IEEE 48TH ANNUAL COMPUTERS, SOFTWARE, AND APPLICATIONS CONFERENCE, COMPSAC 2024, 2024, : 2308 - 2312
  • [32] Enhancing multiple-choice question answering through sequential fine-tuning and Curriculum Learning strategies
    Yigit, Gulsum
    Amasyali, Mehmet Fatih
    KNOWLEDGE AND INFORMATION SYSTEMS, 2023, 65 (11) : 5025 - 5042
  • [33] Enhancing multiple-choice question answering through sequential fine-tuning and Curriculum Learning strategies
    Gulsum Yigit
    Mehmet Fatih Amasyali
    Knowledge and Information Systems, 2023, 65 : 5025 - 5042
  • [34] Enhancing Chinese Essay Discourse Logic Evaluation Through Optimized Fine-Tuning of Large Language Models
    Song, Jinwang
    Song, Yanxin
    Zhou, Guangyu
    Fu, Wenhui
    Zhang, Kunli
    Zan, Hongying
    NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING, PT V, NLPCC 2024, 2025, 15363 : 342 - 352
  • [35] Active Learning for Effectively Fine-Tuning Transfer Learning to Downstream Task
    Abul Bashar, Md
    Nayak, Richi
    ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY, 2021, 12 (02)
  • [36] FashionGPT: LLM instruction fine-tuning with multiple LoRA-adapter fusion
    Gao, Dehong
    Ma, Yufei
    Liu, Sen
    Song, Mengfei
    Jin, Linbo
    Jiang, Wen
    Wang, Xin
    Ning, Wei
    Yu, Shanqing
    Xuan, Qi
    Cai, Xiaoyan
    Yang, Libin
    KNOWLEDGE-BASED SYSTEMS, 2024, 299
  • [37] Domain-Aware Fine-Tuning: Enhancing Neural Network Adaptability
    Ha, Seokhyeon
    Jeong, Sunbeom
    Lee, Jungwoo
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 11, 2024, : 12261 - 12269
  • [38] Enhancing Transformers with Gradient Boosted Decision Trees for NLI Fine-Tuning
    Minixhofer, Benjamin
    Gritta, Milan
    Iacobacci, Ignacio
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL-IJCNLP 2021, 2021, : 303 - 313
  • [39] Dithienopyrrolothiophene (DTPT) based acceptors for fine-tuning molecular packing and enhancing ternary organic photovoltaic performance
    Jiang, Bing-Huang
    Afraj, Shakil N.
    Ezhumalai, Yamuna
    Chang, Chun-Yen
    Yang, Yun-Hsuan
    Su, Yu-Wei
    Abdelhady, Ahmed L.
    Li, Yu-Qi
    Shi, Zhong-En
    Liu, Cheng-Liang
    Chen, Ming-Chou
    Kao, Hsien-Ming
    Chen, Chih-Ping
    JOURNAL OF MATERIALS CHEMISTRY C, 2024, 12 (44) : 17966 - 17976
  • [40] Mechanically responsive crystals: tuning flexibilty through fine-tuning intermolecular interactions
    Dakovic, M.
    Pisacic, M.
    Misura, O.
    ACTA CRYSTALLOGRAPHICA A-FOUNDATION AND ADVANCES, 2022, 78 : E191 - E191