Selective privacy-preserving framework for large language models fine-tuning

被引:0
|
作者
Wang, Teng [1 ]
Zhai, Lindong [1 ]
Yang, Tengfei [1 ]
Luo, Zhucheng [2 ]
Liu, Shuanggen [1 ]
机构
[1] Xian Univ Posts & Telecommun, Sch Cyberspace Secur, Xian 710121, Shaanxi, Peoples R China
[2] Sun Yat Sen Univ, Affiliated Hosp 3, Informat Ctr, Guangzhou 510630, Peoples R China
基金
中国国家自然科学基金;
关键词
Large language models; Fine-tuning; Local differential privacy; Selective privacy protection; DIFFERENTIAL PRIVACY;
D O I
10.1016/j.ins.2024.121000
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Fine-tuning pre -trained large language models (LLMs) helps various downstream tasks, but brings serious privacy leaks when relying on large amounts of data for training. Differentially private stochastic gradient descent (DPSGD) has been designed to introduce noise during model updates to prevent privacy leaks. Nevertheless, fine-tuning LLMs via DPSGD limits the model utility since heavy perturbations are introduced on large high -dimensional gradients. Besides, existing privacy -preserving mechanisms directly perturb all tokens of the input sentences, which are too pessimistic to achieve good model performance. Therefore, this paper researches a selective privacy -preserving framework for fine-tuning LLMs. We propose a first -of -its -kind privacy notion called selective sequence local differential privacy (S-SeqLDP), which provides guarantees of indistinguishability only for the secret part of the sequences. Furthermore, we design a novel framework called SLDP-FT that enables S-SeqLDP-compliant large language model fine-tuning by perturbing the forward -pass embeddings with selective noises. We innovatively investigate the privacy forward weight that determines the noise magnitude of achieving selective privacy protection. Extensive experiments on three tasks demonstrate that our SLDP-FT achieves better model accuracy than state-of-the-art techniques when providing the same level of privacy protection.
引用
收藏
页数:14
相关论文
共 50 条
  • [31] Safety Fine-Tuning at (Almost) No Cost: A Baseline for Vision Large Language Models
    Zong, Yongshuo
    Bohdal, Ondrej
    Yu, Tingyang
    Yang, Yongxin
    Hospedales, Timothy
    Proceedings of Machine Learning Research, 2024, 235 : 62867 - 62891
  • [32] Parameter-efficient fine-tuning of large language models using semantic knowledge tuning
    Prottasha, Nusrat Jahan
    Mahmud, Asif
    Sobuj, Md. Shohanur Islam
    Bhat, Prakash
    Kowsher, Md
    Yousefi, Niloofar
    Garibay, Ozlem Ozmen
    SCIENTIFIC REPORTS, 2024, 14 (01):
  • [33] Enhancing Chinese comprehension and reasoning for large language models: an efficient LoRA fine-tuning and tree of thoughts framework
    Chen, Songlin
    Wang, Weicheng
    Chen, Xiaoliang
    Zhang, Maolin
    Lu, Peng
    Li, Xianyong
    Du, Yajun
    JOURNAL OF SUPERCOMPUTING, 2025, 81 (01):
  • [34] DP-Forward: Fine-tuning and Inference on Language Models with Differential Privacy in Forward Pass
    Du, Minxin
    Yue, Xiang
    Chow, Sherman S. M.
    Wang, Tianhao
    Huang, Chenyu
    Sun, Huan
    PROCEEDINGS OF THE 2023 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, CCS 2023, 2023, : 2665 - 2679
  • [35] Local large language models for privacy-preserving accelerated review of historic echocardiogram reports
    Vaid, Akhil
    Duong, Son Q.
    Lampert, Joshua
    Kovatch, Patricia
    Freeman, Robert
    Argulian, Edgar
    Croft, Lori
    Lerakis, Stamatios
    Goldman, Martin
    Khera, Rohan
    Nadkarni, Girish N.
    JOURNAL OF THE AMERICAN MEDICAL INFORMATICS ASSOCIATION, 2024,
  • [36] Comprehensive Review of Large Language Model Fine-Tuning
    Zhang, Qintong
    Wang, Yuchao
    Wang, Hexi
    Wang, Junxin
    Chen, Hai
    Computer Engineering and Applications, 2024, 60 (17) : 17 - 33
  • [37] Detection of suicidality from medical text using privacy-preserving large language models
    Wiest, Isabella Catharina
    Verhees, Falk Gerrik
    Ferber, Dyke
    Zhu, Jiefu
    Bauer, Michael
    Lewitzka, Ute
    Pfennig, Andrea
    Mikolas, Pavol
    Kather, Jakob Nikolas
    BRITISH JOURNAL OF PSYCHIATRY, 2024, 225 (06) : 532 - 537
  • [38] CONVFIT: Conversational Fine-Tuning of Pretrained Language Models
    Vulic, Ivan
    Su, Pei-Hao
    Coope, Sam
    Gerz, Daniela
    Budzianowski, Pawel
    Casanueva, Inigo
    Mrksic, Nikola
    Wen, Tsung-Hsien
    2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 1151 - 1168
  • [39] Improve Performance of Fine-tuning Language Models with Prompting
    Yang, Zijian Gyozo
    Ligeti-Nagy, Noenn
    INFOCOMMUNICATIONS JOURNAL, 2023, 15 : 62 - 68
  • [40] Fine-tuning language models to recognize semantic relations
    Roussinov, Dmitri
    Sharoff, Serge
    Puchnina, Nadezhda
    LANGUAGE RESOURCES AND EVALUATION, 2023, 57 (04) : 1463 - 1486