全部 标题 作者
关键词 摘要

OALib Journal期刊
ISSN: 2333-9721
费用:99美元

查看量下载量

相关文章

更多...

Whispered Tuning: Data Privacy Preservation in Fine-Tuning LLMs through Differential Privacy

DOI: 10.4236/jsea.2024.171001, PP. 1-22

Keywords: NLP, Differential Privacy, Adversarial Attacks, Informed Decisions

Full-Text   Cite this paper   Add to My Lib

Abstract:

The proliferation of Large Language Models (LLMs) across various sectors underscored the urgency of addressing potential privacy breaches. Vulnerabilities, such as prompt injection attacks and other adversarial tactics, could make these models inadvertently disclose their training data. Such disclosures could compromise personal identifiable information, posing significant privacy risks. In this paper, we proposed a novel multi-faceted approach called Whispered Tuning to address privacy leaks in large language models (LLMs). We integrated a PII redaction model, differential privacy techniques, and an output filter into the LLM fine-tuning process to enhance confidentiality. Additionally, we introduced novel ideas like the Epsilon Dial for adjustable privacy budgeting for differentiated Training Phases per data handler role. Through empirical validation, including attacks on non-private models, we demonstrated the robustness of our proposed solution SecureNLP in safeguarding privacy without compromising utility. This pioneering methodology significantly fortified LLMs against privacy infringements, enabling responsible adoption across sectors.

References

[1]  Touvron, H., Martin, L., Stone, K., Albert, P., Almahairi, A., Babaei, Y., Bashlykov, N., Batra, S., Bhargava, P., Bhosale, S., et al. (2023) Llama 2: Open Foundation and Fine-Tuned Chat Models. arXiv preprint arXiv:2307.09288.
[2]  Liu, Y., Deng, G.L., Li, Y.K., Wang, K.L., Zhang, T.W., Liu, Y.P., Wang, H.Y., Zheng, Y. and Liu, Y. (2023) Prompt Injection Attack against LLM-Integrated Applications. arXiv preprint arXiv:2306.05499.
[3]  Shi, W., Shea, R., Chen, S., Zhang, C., Jia, R. and Yu, Z. (2022) Just Fine-Tune Twice: Selective Differential Privacy for Large Language Models. In: Goldberg, Y., Kozareva, Z. and Zhang, Y., Eds., Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, Association for Computational Linguistics, Abu Dhabi, 6327-6340.
https://doi.org/10.18653/v1/2022.emnlp-main.425
[4]  Klymenko, O., Meisenbacher, S. and Matthes, F. (2022) Differential Privacy in Natural Language Processing: The Story So Far. In: Feyisetan, O., Ghanavati, S., Thaine, P., Habernal, I. and Mireshghallah, F., Eds., Proceedings of the Fourth Workshop on Privacy in Natural Language Processing, Association for Computational Linguistics, Seattle, 1-11.
https://doi.org/10.18653/v1/2022.privatenlp-1.1
[5]  Behnia, R., Ebrahimi, M.R., Pacheco, J. and Padmanabhan, B. (2022) EW-Tune: A Framework for Privately Fine-Tuning Large Language Models with Differential Privacy. 2022 IEEE International Conference on Data Mining Workshops (ICDMW), Orlando, FL, 28 November-1 December 2022, 560-566.
https://doi.org/10.1109/ICDMW58026.2022.00078
[6]  Li, Y., Tan, Z. and Liu, Y. (2023) Privacy-Preserving Prompt Tuning for Large Language Model Services. arXiv preprint arXiv:2305.06212.
[7]  Plant, R., Giuffrida, V. and Gkatzia, D. (2022) You Are What You Write: Preserving Privacy in the Era of Large Language Models. arXiv preprint arXiv:2204.09391.
https://doi.org/10.2139/ssrn.4417900
[8]  Kandpal, N., Wallace, E. and Raffel, C. (2022) Deduplicating Training Data Mitigates Privacy Risks in Language Models. International Conference on Machine Learning, Baltimore, Maryland, July 17-23 2022, 10697-10707.
[9]  Yu, D., Kamath, G., Kulkarni, J., Yin, J., Liu, T.-Y. and Zhang, H.S. (2022) Per-Instance Privacy Accounting for Differentially Private Stochastic Gradient Descent. arXiv preprint arXiv:2206.02617.
[10]  Mireshghallah, F., Uniyal, A., Wang, T.H., Evans, D. and Berg-Kirkpatrick, T. (2022) Memorization in NLP Fine-Tuning Methods. arXiv Preprint arXiv:2205.12506.
[11]  Li, X.C., Liu, D.G., Hashimoto, T.B., Inan, H.A., Kulkarni, J., Lee, Y.-T. and Guha Thakurta, A. (2022) When Does Differentially Private Learning Not Suffer in High Dimensions? Advances in Neural Information Processing Systems, 35, 28616-28630.
[12]  Cherrueau, R.-A., Douence, R. and Südholt, M. (2015) A Language for the Composition of Privacy-Enforcement Techniques. 2015 IEEE Trustcom/BigDataSE/ISPA, Helsinki, 20-22 August 2015, 1037-1044.
https://doi.org/10.1109/Trustcom.2015.480
[13]  Sanh, V., Debut, L., Chaumond, J. and Wolf, T. (2019) DistilBERT, A Distilled Version of BERT: Smaller, Faster, Cheaper and Lighter. arXiv preprint arXiv:1910.01108.
[14]  Devlin, J., Chang, M.-W., Lee, K. and Toutanova, K. (2018) Bert: Pre-Training of Deep Bidirectional Transformers for Language Understanding. arXiv Preprint arXiv:1810.04805.

Full-Text

comments powered by Disqus

Contact Us

service@oalib.com

QQ:3279437679

WhatsApp +8615387084133

WeChat 1538708413