Memory-Efficient Fine-Tuning for Quantized Diffusion Model

被引:0
|
作者
Ryu, Hyogon [1 ]
Lim, Seohyun [1 ]
Shim, Hyunjung [1 ]
机构
[1] Korea Adv Inst Sci & Technol, Kim Jaechul Grad Sch AI, Seoul, South Korea
来源
基金
新加坡国家研究基金会;
关键词
Quantization; Diffusion Model; Transfer Learning;
D O I
10.1007/978-3-031-72640-8_20
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The emergence of billion-parameter diffusion models such as Stable Diffusion XL, Imagen, and DALL-E 3 has significantly propelled the domain of generative AI. However, their large-scale architecture presents challenges in fine-tuning and deployment due to high resource demands and slow inference speed. This paper explores the relatively unexplored yet promising realm of fine-tuning quantized diffusion models. Our analysis revealed that the baseline neglects the distinct patterns in model weights and the different roles throughout time steps when finetuning the diffusion model. To address these limitations, we introduce a novel memory-efficient fine-tuning method specifically designed for quantized diffusion models, dubbed TuneQDM. Our approach introduces quantization scales as separable functions to consider inter-channel weight patterns. Then, it optimizes these scales in a timestep-specific manner for effective reflection of the role of each time step. TuneQDM achieves performance on par with its full-precision counterpart while simultaneously offering significant memory efficiency. Experimental results demonstrate that our method consistently outperforms the baseline in both single-/multi-subject generations, exhibiting high subject fidelity and prompt fidelity comparable to the full precision model.
引用
收藏
页码:356 / 372
页数:17
相关论文
共 50 条
  • [21] Fine-tuning Image Transformers using Learnable Memory
    Sandler, Mark
    Zhmoginov, Andrey
    Vladymyrov, Max
    Jackson, Andrew
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, : 12145 - 12154
  • [22] Parameter-Efficient Fine-Tuning Large Speech Model Based on LoRA
    Ou, Ling
    Feng, Gen
    PROCEEDINGS OF THE 2024 27 TH INTERNATIONAL CONFERENCE ON COMPUTER SUPPORTED COOPERATIVE WORK IN DESIGN, CSCWD 2024, 2024, : 36 - 41
  • [23] BEFM: A balanced and efficient fine-tuning model in class-incremental learning
    Liu, Lize
    Ji, Jian
    Zhao, Lei
    KNOWLEDGE-BASED SYSTEMS, 2025, 315
  • [24] AgglutiFiT: Efficient Low-Resource Agglutinative Language Model Fine-Tuning
    Li, Zhe
    Li, Xiuhong
    Sheng, Jiabao
    Slamu, Wushour
    IEEE ACCESS, 2020, 8 : 148489 - 148499
  • [25] How fine can fine-tuning be? Learning efficient language models
    Radiya-Dixit, Evani
    Wang, Xin
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 108, 2020, 108 : 2435 - 2442
  • [26] FINE-TUNING FINE CHEMICALS
    ROYSE, S
    EUROPEAN CHEMICAL NEWS, 1995, 64 (1693): : 28 - &
  • [27] Augmented Neural Fine-Tuning for Efficient Backdoor Purification
    Karim, Nazmul
    Al Arafat, Abdullah
    Khalid, Umar
    Guo, Zhishan
    Rahnavard, Nazanin
    COMPUTER VISION - ECCV 2024, PT LXXX, 2025, 15138 : 401 - 418
  • [28] A fine-tuning enhanced RAG system with quantized influence measure as AI judge
    Rangan, Keshav
    Yin, Yiqiao
    SCIENTIFIC REPORTS, 2024, 14 (01):
  • [29] AN INFLATIONARY MODEL WITHOUT FINE-TUNING OR SUPERSYMMETRY
    SHER, M
    PHYSICS LETTERS B, 1983, 124 (05) : 329 - 332
  • [30] FedDD: A Federated Fine-Tuning Diffusion Model for Low-count PET Denoising
    Zhou, Yinchi
    Xie, Huidong
    Zhou, Bo
    Wang, Hanzhong
    Li, Biao
    Rominger, Axel
    Shi, Kuangyu
    Liu, Chi
    JOURNAL OF NUCLEAR MEDICINE, 2024, 65