Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Exploring Memorization in Fine-tuned Language Models (2310.06714v2)

Published 10 Oct 2023 in cs.AI, cs.CL, and cs.LG

Abstract: LLMs have shown great capabilities in various tasks but also exhibited memorization of training data, raising tremendous privacy and copyright concerns. While prior works have studied memorization during pre-training, the exploration of memorization during fine-tuning is rather limited. Compared to pre-training, fine-tuning typically involves more sensitive data and diverse objectives, thus may bring distinct privacy risks and unique memorization behaviors. In this work, we conduct the first comprehensive analysis to explore LLMs' (LMs) memorization during fine-tuning across tasks. Our studies with open-sourced and our own fine-tuned LMs across various tasks indicate that memorization presents a strong disparity among different fine-tuning tasks. We provide an intuitive explanation of this task disparity via sparse coding theory and unveil a strong correlation between memorization and attention score distribution.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (10)
  1. Shenglai Zeng (19 papers)
  2. Yaxin Li (27 papers)
  3. Jie Ren (329 papers)
  4. Yiding Liu (30 papers)
  5. Han Xu (92 papers)
  6. Pengfei He (36 papers)
  7. Yue Xing (47 papers)
  8. Shuaiqiang Wang (68 papers)
  9. Jiliang Tang (204 papers)
  10. Dawei Yin (165 papers)
Citations (17)
X Twitter Logo Streamline Icon: https://streamlinehq.com