Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Differentially Private Decoding in Large Language Models (2205.13621v2)

Published 26 May 2022 in cs.CL and cs.LG

Abstract: Recent large-scale NLP systems use a pre-trained LLM on massive and diverse corpora as a headstart. In practice, the pre-trained model is adapted to a wide array of tasks via fine-tuning on task-specific datasets. LLMs, while effective, have been shown to memorize instances of training data thereby potentially revealing private information processed during pre-training. The potential leakage might further propagate to the downstream tasks for which LLMs are fine-tuned. On the other hand, privacy-preserving algorithms usually involve retraining from scratch, which is prohibitively expensive for LLMs. In this work, we propose a simple, easy to interpret, and computationally lightweight perturbation mechanism to be applied to an already trained model at the decoding stage. Our perturbation mechanism is model-agnostic and can be used in conjunction with any LLM. We provide theoretical analysis showing that the proposed mechanism is differentially private, and experimental results showing a privacy-utility trade-off.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Jimit Majmudar (9 papers)
  2. Christophe Dupuy (15 papers)
  3. Charith Peris (21 papers)
  4. Sami Smaili (1 paper)
  5. Rahul Gupta (146 papers)
  6. Richard Zemel (82 papers)
Citations (24)