Optimizing Fine-Tuning in Quantized Language Models: An In-Depth Analysis of Key Variables

被引:0
|
作者
Shen, Ao [1 ]
Lai, Zhiquan [1 ]
Li, Dongsheng [1 ]
Hu, Xiaoyu [2 ]
机构
[1] Natl Univ Def Technol, Natl Key Lab Parallel & Distributed Comp, Changsha 410073, Peoples R China
[2] Acad Mil Sci, Strateg Assessments & Consultat Inst, Beijing 100091, Peoples R China
来源
CMC-COMPUTERS MATERIALS & CONTINUA | 2025年 / 82卷 / 01期
基金
国家重点研发计划; 中国国家自然科学基金;
关键词
Large-scale Language Model; Parameter-Efficient Fine-Tuning; parameter quantization; key variable; trainable; parameters; experimental analysis;
D O I
10.32604/cmc.2024.057491
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Large-scale Language Models (LLMs) have achieved significant breakthroughs in Natural Language Processing (NLP), driven by the pre-training and fine-tuning paradigm. While this approach allows models to specialize in specific tasks with reduced training costs, the substantial memory requirements during fine-tuning present a barrier to broader deployment. Parameter-Efficient Fine-Tuning (PEFT) techniques, such as Low-Rank Adaptation (LoRA), and parameter quantization methods have emerged as solutions to address these challenges by optimizing memory usage and computational efficiency. Among these, QLoRA, which combines PEFT and quantization, has demonstrated notable success in reducing memory footprints during fine-tuning, prompting the development of various QLoRA variants. Despite these advancements, the quantitative impact of key variables on the fine-tuning performance of quantized LLMs remains underexplored. This study presents a comprehensive analysis of these key variables, focusing on their influence across different layer types and depths within LLM architectures. Our investigation uncovers several critical findings: (1) Larger layers, such as MLP layers, can maintain performance despite reductions in adapter rank, while smaller layers, like self-attention layers, are more sensitive to such changes; (2) The effectiveness of balancing factors depends more on specific values rather than layer type or depth; (3) In quantization-aware fine-tuning, larger layers can effectively utilize smaller adapters, whereas smaller layers struggle to do so. These insights suggest that layer type is a more significant determinant of fine-tuning success than layer depth when optimizing quantized LLMs. Moreover, for the same discount of trainable parameters, reducing the trainable parameters in a larger layer is more effective in preserving fine-tuning accuracy than in a smaller one. This study provides valuable guidance for more efficient fine-tuning strategies and opens avenues for further research into optimizing LLM fine-tuning in resource-constrained environments.
引用
收藏
页码:307 / 325
页数:19
相关论文
共 50 条
  • [31] Fine-Tuning Large Enterprise Language Models via Ontological Reasoning
    Baldazzi, Teodoro
    Bellomarini, Luigi
    Ceri, Stefano
    Colombo, Andrea
    Gentili, Andrea
    Sallinger, Emanuel
    RULES AND REASONING, RULEML+RR 2023, 2023, 14244 : 86 - 94
  • [32] Fine-Tuning Pre-Trained Language Models with Gaze Supervision
    Deng, Shuwen
    Prasse, Paul
    Reich, David R.
    Scheffer, Tobias
    Jager, Lena A.
    PROCEEDINGS OF THE 62ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 2: SHORT PAPERS, 2024, : 217 - 224
  • [33] Repeatability of Fine-Tuning Large Language Models Illustrated Using QLoRA
    Alahmari, Saeed S.
    Hall, Lawrence O.
    Mouton, Peter R.
    Goldgof, Dmitry B.
    IEEE ACCESS, 2024, 12 : 153221 - 153231
  • [34] Fine-tuning large language models for rare disease concept normalization
    Wang, Andy
    Liu, Cong
    Yang, Jingye
    Weng, Chunhua
    JOURNAL OF THE AMERICAN MEDICAL INFORMATICS ASSOCIATION, 2024, 31 (09) : 2076 - 2083
  • [35] Robust Fine-Tuning of Vision-Language Models for Domain Generalization
    Vogt-Lowell, Kevin
    Lee, Noah
    Tsiligkaridis, Theodoros
    Vaillant, Marc
    2023 IEEE HIGH PERFORMANCE EXTREME COMPUTING CONFERENCE, HPEC, 2023,
  • [36] Fine-tuning the circadian system with light treatment for Parkinson's disease: an in-depth, critical review
    Willis, Gregory L.
    Armstrong, Stuart M.
    REVIEWS IN THE NEUROSCIENCES, 2024, 35 (01) : 57 - 84
  • [37] Fine-tuning the circadian system with light treatment for Parkinson's disease: an in-depth, critical review
    Willis, Gregory L.
    Armstrong, Stuart M.
    REVIEWS IN THE NEUROSCIENCES, 2023,
  • [38] Lexicon-based fine-tuning of multilingual language models for low-resource language sentiment analysis
    Dhananjaya, Vinura
    Ranathunga, Surangika
    Jayasena, Sanath
    CAAI TRANSACTIONS ON INTELLIGENCE TECHNOLOGY, 2024, 9 (05) : 1116 - 1125
  • [39] Multi-phase Fine-Tuning: A New Fine-Tuning Approach for Sign Language Recognition
    Sarhan, Noha
    Lauri, Mikko
    Frintrop, Simone
    KUNSTLICHE INTELLIGENZ, 2022, 36 (01): : 91 - 98
  • [40] Multi-phase Fine-Tuning: A New Fine-Tuning Approach for Sign Language Recognition
    Noha Sarhan
    Mikko Lauri
    Simone Frintrop
    KI - Künstliche Intelligenz, 2022, 36 : 91 - 98