Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

OpenUnlearning: Accelerating LLM Unlearning via Unified Benchmarking of Methods and Metrics (2506.12618v1)

Published 14 Jun 2025 in cs.CL

Abstract: Robust unlearning is crucial for safely deploying LLMs in environments where data privacy, model safety, and regulatory compliance must be ensured. Yet the task is inherently challenging, partly due to difficulties in reliably measuring whether unlearning has truly occurred. Moreover, fragmentation in current methodologies and inconsistent evaluation metrics hinder comparative analysis and reproducibility. To unify and accelerate research efforts, we introduce OpenUnlearning, a standardized and extensible framework designed explicitly for benchmarking both LLM unlearning methods and metrics. OpenUnlearning integrates 9 unlearning algorithms and 16 diverse evaluations across 3 leading benchmarks (TOFU, MUSE, and WMDP) and also enables analyses of forgetting behaviors across 450+ checkpoints we publicly release. Leveraging OpenUnlearning, we propose a novel meta-evaluation benchmark focused specifically on assessing the faithfulness and robustness of evaluation metrics themselves. We also benchmark diverse unlearning methods and provide a comparative analysis against an extensive evaluation suite. Overall, we establish a clear, community-driven pathway toward rigorous development in LLM unlearning research.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Vineeth Dorna (4 papers)
  2. Anmol Mekala (6 papers)
  3. Wenlong Zhao (18 papers)
  4. Andrew McCallum (132 papers)
  5. Zachary C. Lipton (137 papers)
  6. J. Zico Kolter (151 papers)
  7. Pratyush Maini (19 papers)