Optimizing Fine-Tuning in Quantized Language Models: An In-Depth Analysis of Key Variables

被引:0
|
作者
Shen, Ao [1 ]
Lai, Zhiquan [1 ]
Li, Dongsheng [1 ]
Hu, Xiaoyu [2 ]
机构
[1] Natl Univ Def Technol, Natl Key Lab Parallel & Distributed Comp, Changsha 410073, Peoples R China
[2] Acad Mil Sci, Strateg Assessments & Consultat Inst, Beijing 100091, Peoples R China
来源
CMC-COMPUTERS MATERIALS & CONTINUA | 2025年 / 82卷 / 01期
基金
国家重点研发计划; 中国国家自然科学基金;
关键词
Large-scale Language Model; Parameter-Efficient Fine-Tuning; parameter quantization; key variable; trainable; parameters; experimental analysis;
D O I
10.32604/cmc.2024.057491
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Large-scale Language Models (LLMs) have achieved significant breakthroughs in Natural Language Processing (NLP), driven by the pre-training and fine-tuning paradigm. While this approach allows models to specialize in specific tasks with reduced training costs, the substantial memory requirements during fine-tuning present a barrier to broader deployment. Parameter-Efficient Fine-Tuning (PEFT) techniques, such as Low-Rank Adaptation (LoRA), and parameter quantization methods have emerged as solutions to address these challenges by optimizing memory usage and computational efficiency. Among these, QLoRA, which combines PEFT and quantization, has demonstrated notable success in reducing memory footprints during fine-tuning, prompting the development of various QLoRA variants. Despite these advancements, the quantitative impact of key variables on the fine-tuning performance of quantized LLMs remains underexplored. This study presents a comprehensive analysis of these key variables, focusing on their influence across different layer types and depths within LLM architectures. Our investigation uncovers several critical findings: (1) Larger layers, such as MLP layers, can maintain performance despite reductions in adapter rank, while smaller layers, like self-attention layers, are more sensitive to such changes; (2) The effectiveness of balancing factors depends more on specific values rather than layer type or depth; (3) In quantization-aware fine-tuning, larger layers can effectively utilize smaller adapters, whereas smaller layers struggle to do so. These insights suggest that layer type is a more significant determinant of fine-tuning success than layer depth when optimizing quantized LLMs. Moreover, for the same discount of trainable parameters, reducing the trainable parameters in a larger layer is more effective in preserving fine-tuning accuracy than in a smaller one. This study provides valuable guidance for more efficient fine-tuning strategies and opens avenues for further research into optimizing LLM fine-tuning in resource-constrained environments.
引用
收藏
页码:307 / 325
页数:19
相关论文
共 50 条
  • [41] Fine-tuning constraints on supergravity models
    Bastero-Gil, M
    Kane, GL
    King, SF
    PHYSICS LETTERS B, 2000, 474 (1-2) : 103 - 112
  • [42] OptimalMEE: Optimizing Large Language Models for Medical Event Extraction Through Fine-Tuning and Post-hoc Verification
    Sun, Yaoqian
    Wu, Dan
    Chen, Zikang
    Cai, Hailing
    An, Jiye
    ARTIFICIAL INTELLIGENCE IN MEDICINE, PT I, AIME 2024, 2024, 14844 : 303 - 311
  • [43] Memory-Efficient Fine-Tuning for Quantized Diffusion Model
    Ryu, Hyogon
    Lim, Seohyun
    Shim, Hyunjung
    COMPUTER VISION - ECCV 2024, PT XVI, 2025, 15074 : 356 - 372
  • [44] Parameter-efficient fine-tuning of large language models using semantic knowledge tuning
    Prottasha, Nusrat Jahan
    Mahmud, Asif
    Sobuj, Md. Shohanur Islam
    Bhat, Prakash
    Kowsher, Md
    Yousefi, Niloofar
    Garibay, Ozlem Ozmen
    SCIENTIFIC REPORTS, 2024, 14 (01):
  • [45] Gender-tuning: Empowering Fine-tuning for Debiasing Pre-trained Language Models
    Ghanbarzadeh, Somayeh
    Huang, Yan
    Palangi, Hamid
    Moreno, Radames Cruz
    Khanpour, Hamed
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, 2023, : 5448 - 5458
  • [46] Analysis of fine-tuning measures in models with extended Higgs sectors
    Boer, Daniel
    Peeters, Ruud
    Zeinstra, Sybrand
    NUCLEAR PHYSICS B, 2019, 946
  • [47] Safety Fine-Tuning at (Almost) No Cost: A Baseline for Vision Large Language Models
    Zong, Yongshuo
    Bohdal, Ondrej
    Yu, Tingyang
    Yang, Yongxin
    Hospedales, Timothy
    arXiv, 1600,
  • [48] Fine-tuning Language Models for Joint Rewriting and Completion of Code with Potential Bugs
    Wang, Dingmin
    Zhao, Jinman
    Pei, Hengzhi
    Tana, Samson
    Zha, Sheng
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 15854 - 15868
  • [49] Parameter-efficient fine-tuning in large language models: a survey of methodologies
    Luping Wang
    Sheng Chen
    Linnan Jiang
    Shu Pan
    Runze Cai
    Sen Yang
    Fei Yang
    Artificial Intelligence Review, 58 (8)
  • [50] Recall and Learn: Fine-tuning Deep Pretrained Language Models with Less Forgetting
    Chen, Sanyuan
    Hou, Yutai
    Cui, Yiming
    Che, Wanxiang
    Liu, Ting
    Yu, Xiangzhan
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 7870 - 7881