Memory-Efficient Fine-Tuning for Quantized Diffusion Model

被引:0
|
作者
Ryu, Hyogon [1 ]
Lim, Seohyun [1 ]
Shim, Hyunjung [1 ]
机构
[1] Korea Adv Inst Sci & Technol, Kim Jaechul Grad Sch AI, Seoul, South Korea
来源
基金
新加坡国家研究基金会;
关键词
Quantization; Diffusion Model; Transfer Learning;
D O I
10.1007/978-3-031-72640-8_20
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The emergence of billion-parameter diffusion models such as Stable Diffusion XL, Imagen, and DALL-E 3 has significantly propelled the domain of generative AI. However, their large-scale architecture presents challenges in fine-tuning and deployment due to high resource demands and slow inference speed. This paper explores the relatively unexplored yet promising realm of fine-tuning quantized diffusion models. Our analysis revealed that the baseline neglects the distinct patterns in model weights and the different roles throughout time steps when finetuning the diffusion model. To address these limitations, we introduce a novel memory-efficient fine-tuning method specifically designed for quantized diffusion models, dubbed TuneQDM. Our approach introduces quantization scales as separable functions to consider inter-channel weight patterns. Then, it optimizes these scales in a timestep-specific manner for effective reflection of the role of each time step. TuneQDM achieves performance on par with its full-precision counterpart while simultaneously offering significant memory efficiency. Experimental results demonstrate that our method consistently outperforms the baseline in both single-/multi-subject generations, exhibiting high subject fidelity and prompt fidelity comparable to the full precision model.
引用
收藏
页码:356 / 372
页数:17
相关论文
共 50 条
  • [1] MEFT: Memory-Efficient Fine-Tuning through Sparse Adapter
    Hao, Jitai
    Sun, Weiwei
    Xin, Xin
    Meng, Qi
    Chen, Zhumin
    Ren, Pengjie
    Ren, Zhaochun
    PROCEEDINGS OF THE 62ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1: LONG PAPERS, 2024, : 2375 - 2388
  • [2] Quantized Side Tuning: Fast and Memory-Efficient Tuning of Quantized Large Language Models
    Zhang, Zhengxin
    Zhao, Dan
    Miao, Xupeng
    Oliaro, Gabriele
    Zhang, Zhihao
    Li, Qing
    Jiang, Yong
    Jia, Zhihao
    PROCEEDINGS OF THE 62ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1: LONG PAPERS, 2024, : 1 - 17
  • [3] Memory-Efficient Fine-Tuning of Compressed Large Language Models via sub-4-bit Integer Quantization
    Kim, Jeonghoon
    Lee, Jung Hyun
    Kim, Sungdong
    Park, Joonsuk
    Yoo, Kang Min
    Kwon, Se Jung
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [4] Make Pre-trained Model Reversible: From Parameter to Memory Efficient Fine-Tuning
    Liao, Baohao
    Tan, Shaomu
    Monz, Christof
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [5] Efficient Index Learning via Model Reuse and Fine-tuning
    Liu, Guanli
    Qi, Jianzhong
    Kulik, Lars
    Soga, Kazuya
    Borovica-Gajic, Renata
    Rubinstein, Benjamin I. P.
    2023 IEEE 39TH INTERNATIONAL CONFERENCE ON DATA ENGINEERING WORKSHOPS, ICDEW, 2023, : 60 - 66
  • [6] Fine-tuning Pipeline for Hand Image Generation Using Diffusion Model
    Bai, Bingyuan
    Xie, Haoran
    Miyata, Kazunori
    2024 NICOGRAPH INTERNATIONAL, NICOINT 2024, 2024, : 58 - 63
  • [7] Towards Adaptive Prefix Tuning for Parameter-Efficient Language Model Fine-tuning
    Zhang, Zhen-Ru
    Tan, Chuanqi
    Xu, Haiyang
    Wang, Chengyu
    Huang, Jun
    Huang, Songfang
    61ST CONFERENCE OF THE THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 2, 2023, : 1239 - 1248
  • [8] MultiFiT: Efficient Multi-lingual Language Model Fine-tuning
    Eisenschlos, Julian
    Ruder, Sebastian
    Czapla, Piotr
    Kardas, Marcin
    Gugger, Sylvain
    Howard, Jeremy
    2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019): PROCEEDINGS OF THE CONFERENCE, 2019, : 5702 - 5707
  • [9] On the Effectiveness of Parameter-Efficient Fine-Tuning
    Fu, Zihao
    Yang, Haoran
    So, Anthony Man-Cho
    Lam, Wai
    Bing, Lidong
    Collier, Nigel
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 11, 2023, : 12799 - 12807
  • [10] PockEngine: Sparse and Efficient Fine-tuning in a Pocket
    Zhu, Ligeng
    Hu, Lanxiang
    Lin, Ji
    Wang, Wei-Chen
    Chen, Wei-Ming
    Gan, Chuang
    Han, Song
    56TH IEEE/ACM INTERNATIONAL SYMPOSIUM ON MICROARCHITECTURE, MICRO 2023, 2023, : 1381 - 1394