%0 Journal Article %T Protecting LLMs against Privacy Attacks While Preserving Utility %A Gunika Dhingra %A Saumil Sood %A Zeba Mohsin Wase %A Arshdeep Bahga %A Vijay K. Madisetti %J Journal of Information Security %P 448-473 %@ 2153-1242 %D 2024 %I Scientific Research Publishing %R 10.4236/jis.2024.154026 %X The recent interest in the deployment of Generative AI applications that use large language models (LLMs) has brought to the forefront significant privacy concerns, notably the leakage of Personally Identifiable Information (PII) and other confidential or protected information that may have been memorized during training, specifically during a fine-tuning or customization process. This inadvertent leakage of sensitive information typically occurs when the models are subjected to black-box attacks. To address the growing concerns of safeguarding private and sensitive information while simultaneously preserving its utility, we analyze the performance of Targeted Catastrophic Forgetting (TCF). TCF involves preserving targeted pieces of sensitive information within datasets through an iterative pipeline which significantly reduces the likelihood of such information being leaked or reproduced by the model during black-box attacks, such as the autocompletion attack in our case. The experiments conducted using TCF evidently demonstrate its capability to reduce the extraction of PII while still preserving the context and utility of the target application. %K Large Language Models %K PII Leakage %K Privacy %K Memorization %K Membership Inference Attack (MIA) %K Defenses %K Generative Adversarial Networks (GANs) %K Synthetic Data %U http://www.scirp.org/journal/PaperInformation.aspx?PaperID=136070