Large language models (LLMs) have achieved remarkable success in natural language processing (NLP) tasks. However, fine-tuning LLMs using private datasets raises significant privacy concerns, as models can inadvertently memorize sensitive information. Differentially Private Stochastic Gradient Descent (DP-SGD) provides a mathematically rigorous solution but suffers from high computational overhead, slow convergence, and excessive privacy budget consumption, making it impractical for large-scale models. To address these challenges, we propose an efficient differentially private fine-tuning method that combines Quantized Low-Rank Adaptation (QLoRA) and Prefix Tuning. QLoRA employs 4-bit NormalFloat quantization and low-rank adaptation, significantly reducing memory consumption and improving computational efficiency. Prefix Tuning optimizes a small set of prefix vectors without modifying the model’s main parameters, further reducing the impact of DP noise. Additionally, we introduce a hybrid adaptive gradient clipping strategy, which applies sample-wise adaptive clipping for Prefix Tuning and group-wise clipping for QLoRA, effectively balancing privacy protection and model utility. We evaluate our approach on GPT-2 using benchmark datasets including E2E NLG Challenge, XSum, SST-2, and DART, measuring performance using BLEU, ROUGE, and F1-score. Results demonstrate that QLoRA + Prefix Tuning achieves up to 75% memory reduction while maintaining over 95% of the original model performance under a moderate privacy budget (ε=3), outperforming traditional DP fine-tuning methods. Our work provides a practical and scalable solution for privacy-preserving LLM fine-tuning in resource-constrained environments.
Da YuSaurabh NaikArtūrs BačkursSivakanth GopiHuseyin A. InanGautam KamathJanardhan KulkarniYin Tat LeeAndre ManoelLukas WutschitzSergey YekhaninHuishuai Zhang
Saeed AlahmariLawrence HallPeter R. MoutonDmitry B. Goldgof