Papers
Topics
Authors
Recent
Search
2000 character limit reached

On the Impossibility of Retrain Equivalence in Machine Unlearning

Published 18 Oct 2025 in cs.LG | (2510.16629v2)

Abstract: Machine unlearning seeks to selectively remove the "influence" of specific training data on a model's outputs. The ideal goal is Retrain Equivalence--behavior identical to a model trained from scratch on only the retained data. This goal was formulated for models trained on i.i.d. data batches, but modern pipelines often involve multi-stage training, with each stage having a distinct data distribution and objective. Examples include LLM fine-tuning for alignment, reasoning ability, etc. Our study shows via theory and experiments that this shift to multi-stage training introduces a fundamental barrier for machine unlearning. The theory indicates that the outcome of local unlearning--methods that only use gradients computed on the forget set--is path-dependent. That is, a model's behavior during unlearning is influenced by the order of its training stages during learning, making it impossible for path-oblivious algorithms to universally achieve Retrain Equivalence. We empirically demonstrate the same phenomenon in LLM post-training across Llama and Qwen models (1B to 14B) with gradient ascent, NPO, and SimNPO local unlearning algorithms. Models fine-tuned via different orderings of identical training stages diverge in behavior during unlearning, with the degradation in GSM8K accuracy after unlearning varying by over 20% across paths. We also observe that some learning paths consistently produce models that unlearn slowly. During unlearning, whether the probability mass gets squeezed into paraphrasing or alternative concepts is also path-dependent. These results consistently show that Retrain Equivalence is an ill-posed target for local unlearning algorithms, so long as the target models are trained in stages. In situations where access to models' training histories is hard, the current work calls for rethinking the definition and desiderata of machine unlearning.

Summary

  • The paper establishes that retrain equivalence in machine unlearning is unachievable due to inherent path-dependent training effects.
  • It utilizes overparameterized linear regression models to mathematically demonstrate exponential divergence caused by training order variations.
  • Empirical evaluations on LLMs reveal that sequential data exposure, especially the recency effect, significantly hinders local unlearning accuracy.

On the Impossibility of Retrain Equivalence in Machine Unlearning

Introduction

The paper "On the Impossibility of Retrain Equivalence in Machine Unlearning" addresses the challenging problem of machine unlearning, specifically targeting the notion of Retrain Equivalence (RE). This concept refers to a model's ability to yield identical behavior to one retrained from scratch on a subset of data that excludes the "forget set." The study extends traditional assumptions centered on i.i.d. datasets, which modern multi-stage training methodologies challenge. Through theoretical foundations and empirical evidence using LLMs, the paper elucidates why achieving RE is fundamentally infeasible in local unlearning algorithms absent path dependency consideration.

Theoretical Insights

At the core of the research is the realization that multi-stage training significantly impacts unlearning algorithms. In staged training, models encounter diverse datasets sequentially, which means subsequent model behavior, when subjected to unlearning tasks, is inextricably linked with the specific sequence of prior training data exposure. The work mathematically demonstrates path-dependent divergence using overparameterized linear regression models. These models, trained on identical datasets but differing in order, exhibit exponential divergence in predictions after the same local unlearning rule is applied. Thus, it is theoretically impossible for a path-oblivious, local unlearning algorithm to universally achieve RE. Figure 1

Figure 1: History dependence of gradient ascent unlearning. Each panel demonstrates the distinctive unlearning process due to the order of dataset exposure in the training sequence.

Experimental Evaluation

The empirical section demonstrates practical implications by analyzing the unlearning behavior of LLMs from the Llama and Qwen model families. These models were fine-tuned with varying dataset sequences and subjected to identical unlearning protocols using gradient ascent and variations such as NPO and SimNPO. The experiments consistently affirmed path dependence, with substantial variances in accuracy degradation post-unlearning, evidencing the theoretical predictions that compensation for model behavior can not be path-oblivious. Figure 2

Figure 2: Change in forget quality and retained utility in three models, showcasing path-dependent IEEE divergences during unlearning.

Path Dependency and the Recency Effect

A novel phenomenon identified is the "recency effect," where unlearning proceeds most sluggishly when the forget set is the final exposure in the training sequence. This effect suggests that models optimize parameters profoundly during their last-exposed dataset segment, causing subsequent unlearning to misalign with previous learning stages. Figure 3

Figure 3: The recency effect persists across different learning rates, indicating a standard neural network property when subjected to unlearning algorithms.

Implications and Future Directions

This work illuminates the inherent conflict between achieving RE and the path dependency imposed by staged training in modern LLM architectures. Important implications include the call for reevaluating unlearning objectives beyond simple metrics that overlook these complexities. The research intertwines theoretical limitations with practical challenges, urging significant rethinking in deploying machine unlearning algorithms, especially in regulatory contexts where data erasure and model updates impart legal and ethical obligations.

Conclusion

In summary, the paper decisively unravels the impracticality of achieving Retrain Equivalence in machine unlearning through local algorithms lacking historical consideration. This finding has profound implications for designing and evaluating unlearning methods, advocating for nuanced approaches that transcend current evaluative measures restricted by singular retrain baseline ideals. By bridging theoretical insights with empirical evidence, the study sets a platform for future explorations into robust and contextually aware unlearning algorithms.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 1 tweet with 3 likes about this paper.