Table of Contents
Fetching ...

Enhancing Data Privacy in Large Language Models through Private Association Editing

Davide Venditti, Elena Sofia Ruzzetti, Giancarlo A. Xompero, Cristina Giannone, Andrea Favalli, Raniero Romagnoli, Fabio Massimo Zanzotto

TL;DR

Private Association Editing (PAE) is designed to effectively remove Personally Identifiable Information (PII) without retraining the model, and experimental results demonstrate the effectiveness of PAE with respect to alternative baseline methods.

Abstract

Large language models (LLMs) require a significant redesign in solutions to preserve privacy in data-intensive applications due to their text-generation capabilities. Indeed, LLMs tend to memorize and emit private information when maliciously prompted. In this paper, we introduce Private Association Editing (PAE) as a novel defense approach for private data leakage. PAE is designed to effectively remove Personally Identifiable Information (PII) without retraining the model. Experimental results demonstrate the effectiveness of PAE with respect to alternative baseline methods. We believe PAE will serve as a critical tool in the ongoing effort to protect data privacy in LLMs, encouraging the development of safer models for real-world applications.

Enhancing Data Privacy in Large Language Models through Private Association Editing

TL;DR

Private Association Editing (PAE) is designed to effectively remove Personally Identifiable Information (PII) without retraining the model, and experimental results demonstrate the effectiveness of PAE with respect to alternative baseline methods.

Abstract

Large language models (LLMs) require a significant redesign in solutions to preserve privacy in data-intensive applications due to their text-generation capabilities. Indeed, LLMs tend to memorize and emit private information when maliciously prompted. In this paper, we introduce Private Association Editing (PAE) as a novel defense approach for private data leakage. PAE is designed to effectively remove Personally Identifiable Information (PII) without retraining the model. Experimental results demonstrate the effectiveness of PAE with respect to alternative baseline methods. We believe PAE will serve as a critical tool in the ongoing effort to protect data privacy in LLMs, encouraging the development of safer models for real-world applications.

Paper Structure

This paper contains 27 sections, 4 equations, 3 figures, 7 tables.

Figures (3)

  • Figure 1: Preserving privacy for LLMs by using Private Association Editing
  • Figure 2: Memorization Attack against sequentially post-edit models. The smaller the batch size $k$, the larger the number of sequential updates necessary to edit all the private email addresses leaked by the original model.
  • Figure 3: The post-edit model is increasingly different from the pre-edit model as $\lambda$ increases. As discussed in Section \ref{['sec:consistency']}, this is an indication of a diminished utility of the model.