SMFA: Sculpted Memory Forgetting Adapter
- The paper presents SMFA's novel method to confine unlearning to specific memory regions using loss-driven adapters and dynamic masks for precise concept erasure.
- It leverages domain-specific techniques in text-to-image diffusion and multimodal language models, applying concept-aware losses and LoRA-style adapters for controlled updates.
- Benchmark results show SMFA achieves significant forgetting efficiency with high retention of unrelated knowledge, outperforming previous unlearning baselines.
The Sculpted Memory Forgetting Adapter (SMFA) is a specialized unlearning framework for controlling knowledge retention and deletion in large-scale neural models. In text-to-image diffusion models and multimodal LLMs (MLLMs), SMFA enables selective “erasure” of sensitive or undesirable concepts while rigorously preserving unrelated functional capabilities. The framework was independently presented for both diffusion-based image synthesis and MLLMs, with modality-specific instantiations. Both share a central paradigm: confining model updates to memory regions responsible for the targeted concepts by combining loss-driven adapters and precision masking mechanisms (Li et al., 12 Apr 2025, Zeng et al., 25 Nov 2025).
1. Architectural Principles
SMFA is an adapter-based unlearning method, but the adapter’s construction and application are tailored to the model domain.
In T2I diffusion settings, SMFA augments the pre-trained U-Net with a dynamic, sparsity-inducing gradient mask and a composite loss function:
- Dynamic Mask (): Gates U-Net parameters, restricting updates to weights implicated by the current unlearning objective.
- Concept-Aware Loss (): Guides unlearning, aligning forgotten concepts to user-defined superclasses and regularizing via knowledge distillation.
In MLLMs, SMFA operates by fine-tuning a frozen base model () with a Memory Forgetting Adapter () that is further sculpted using a Retaining Anchor ():
- MFA: is trained to transform all forgotten responses (on ) to explicit refusals (from a large label bank), while preserving few-shot retained samples ().
- Masking: The update is masked to zero-out weights where the forgetting and retaining anchors conflict in sign and magnitude, thus localizing memory changes.
Both implementations eschew wholesale model retraining in favor of controlled, local adaptation (Li et al., 12 Apr 2025, Zeng et al., 25 Nov 2025).
2. Dynamic Masking and Gradient Control
The SMFA for diffusion models introduces a three-stage dynamic mask pipeline:
- Gradient Masking: Each update computes the loss gradient , multiplies it elementwise by the mask , and updates weights only where the mask is active.
- Accumulated Gradient Statistics: A running sum of absolute gradients is maintained to inform mask updates.
- Scheduled Mask Reallocation: Every steps, a fraction of active and inactive mask entries are swapped based on statistics, enforcing sustained sparsity and adaptability.
In MLLMs, masking is formulated over learned LoRA adapters:
- Directional Conflict Mask (): if , otherwise 0.
- Relative Magnitude Mask (): if , with .
- Final Mask (): .
- Sculpted Adapter: is obtained by zeroing out masked entries: .
This ensures updates effect forgetting only in memory regions implicated by the “forget” task, mitigating collateral damage to unrelated knowledge (Li et al., 12 Apr 2025, Zeng et al., 25 Nov 2025).
3. Concept-Aware and Selective Loss Functions
For diffusion models, SMFA employs a three-part objective:
- Unlearning Loss (): Penalizes the U-Net when predictions for the target (to-be-forgotten) concept do not match those for its mapped superclass :
- Superclass Alignment (): Ensures the U-Net’s output for prompts aligns with the true underlying noise, maintaining output semantic structure.
- Knowledge-Distillation Regularization (): Locks in prior unlearned concepts by aligning current predictions to those of a snapshot teacher on all previously erased prompts.
The total loss is
with typical , .
In MLLMs, SMFA minimizes standard cross-entropy over refusal-labeled forget sets () and few-shot retained examples (). The regularization is imposed via hard masking of adapter updates rather than as a differentiable term (Li et al., 12 Apr 2025, Zeng et al., 25 Nov 2025).
4. Optimization and Training Protocol
The optimization for multi-concept forgetting is sequential and relies on model snapshots:
- Initialization: Pre-trained model parameters are loaded; running statistics () for the mask are optionally warmed up.
- Per-Concept Forgetting:
- For each concept , copy the current model as the “teacher” for distillation regularization.
- Fine-tune on and batches, compute all loss terms, update using masked gradients.
- Update mask every steps by reallocating based on .
- Discard teacher after concept completion.
- Output: Final model reflects all target forgets, with unrelated concepts and functionalities preserved.
In MLLMs, LoRA adapters are learned for both forget and retain anchors, only modifying the adapter weights () with the base model fixed. Masking is applied post-hoc to using the directional/magnitude scheme, yielding the sculpted adapter. The process is compatible with various model sizes (e.g., Qwen2.5-VL-7B, LLaVA-OneVision-7B) and forget ratios up to 15% of finetuning data. Hyperparameters such as the mask trade-off (default ) are empirically validated (Zeng et al., 25 Nov 2025).
5. Benchmarking, Evaluation, and Performance
In MLLMs, selective unlearning efficacy is quantified using the S-MLLMUn Bench. Key evaluation axes include:
- Forgetting (, ): ROUGE-L and Fact Score for the forget set, where lower is better.
- Retention (, , ): ROUGE-L, Fact, and Meaningful scores for retained and general understanding sets, where higher is better.
SMFA achieves low (strong erasure) with minimal decreases in and image understanding scores, outperforming GA, KL, MANU, and IDK baselines. Ablation experiments indicate both directional and magnitude-based mask criteria are required for robust trade-off. Increasing mask hyperparameter deepens forgetting up to a point, after which retention degrades. The forgetting occurs precisely via refusals (“I’m not the right source for that”), not arbitrary collapse. Even with few-shot anchors (), SMFA maintains coverage across memory and vision tasks (Zeng et al., 25 Nov 2025).
For diffusion models, unlearning is measured by output fidelity, forgetting effectiveness, and semantic integrity post-unlearning. SMFA yields improved results over previous unlearning techniques, especially in multi-concept scenarios, due to targeted weight changes and explicit alignment losses preventing drift or collapse (Li et al., 12 Apr 2025).
6. Module Interactions, Limitations, and Practical Guidance
- Adapter–Mask Synergy: In both modalities, the “forgetting” adapter direction is reined in by the retention anchor or dynamic mask, which localizes erasure and prevents spillage into unrelated capabilities.
- Loss Granularity: The multi-component loss in diffusion settings aligns forgotten concepts () toward semantically meaningful superclasses (), unlike naïve unlearning, which often yields degenerate or semantically empty results.
- Regularization and Memory Locking: Knowledge distillation regularization in diffusion and hard-masked updates in MLLMs safeguard previously unlearned concepts against sequential catastrophic re-learning.
- Hyperparameter Sensitivity: Forget vs retain balance is tunable (e.g., via in MLLMs, in diffusion). Extreme values can induce over-forgetting or retention failures; practically, moderate values yield best trade-offs.
- Implementation Strategies: In MLLMs, LoRA-style adapters inserted in all linear layers attain capacity-efficiency balance. In diffusion, 50% sparsity and periodic mask reallocation are empirically optimal.
- Scalability: In both domains, SMFA’s localized adapter paradigm admits efficient post-hoc, modular updates to deployed large-scale models without end-to-end retraining.
7. Context and Relation to Broader Unlearning Research
SMFA addresses long-standing weaknesses in model unlearning: instability, residual memory, over-forgetting, and generation collapse. Compared to optimization-based global unlearning (e.g., full-model retraining, GA Difference, KL Minimization), SMFA confines weight changes spatially and semantically, ensuring targeted unlearning with bounded side effects. The two independently developed SMFA variants establish a modular, scalable template for future selective unlearning work in both generative and multimodal neural architectures (Li et al., 12 Apr 2025, Zeng et al., 25 Nov 2025).