What governs catastrophic forgetting and why RL vs. SFT differ
Determine the underlying mechanism that governs catastrophic forgetting in foundation models and explain why supervised fine-tuning (SFT) and on-policy reinforcement learning (RL) exhibit different forgetting behavior despite achieving similar new-task performance.
References
Consequently, it remains unclear what truly governs forgetting or why different training algorithms behave so differently.
— RL's Razor: Why Online Reinforcement Learning Forgets Less
(2509.04259 - Shenfeld et al., 4 Sep 2025) in Section 1, Introduction