Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Memorization in NLP Fine-tuning Methods (2205.12506v2)

Published 25 May 2022 in cs.CL and cs.LG

Abstract: LLMs are shown to present privacy risks through memorization of training data, and several recent works have studied such risks for the pre-training phase. Little attention, however, has been given to the fine-tuning phase and it is not well understood how different fine-tuning methods (such as fine-tuning the full model, the model head, and adapter) compare in terms of memorization risk. This presents increasing concern as the "pre-train and fine-tune" paradigm proliferates. In this paper, we empirically study memorization of fine-tuning methods using membership inference and extraction attacks, and show that their susceptibility to attacks is very different. We observe that fine-tuning the head of the model has the highest susceptibility to attacks, whereas fine-tuning smaller adapters appears to be less vulnerable to known extraction attacks.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Fatemehsadat Mireshghallah (26 papers)
  2. Archit Uniyal (3 papers)
  3. Tianhao Wang (98 papers)
  4. David Evans (63 papers)
  5. Taylor Berg-Kirkpatrick (106 papers)
Citations (33)