Info-Theoretic Framework for Attribute Unlearning
- The paper introduces an information-theoretic objective that minimizes mutual information between learned representations and sensitive attributes while preserving task-relevant utility.
- It employs surrogate losses, such as variational bounds, MMD, adversarial and contrastive methods, to efficiently estimate and control the attribute's statistical footprint.
- Practical schemes like LEGO and MaSS demonstrate strong empirical trade-offs, significantly reducing attribute inference without substantial loss in main-task performance.
Selective removal of sensitive or nuisance attributes from learned representations, known as attribute unlearning, has emerged as a fundamental requirement for privacy, fairness, and compliance in machine learning systems. The information-theoretic framework for attribute unlearning formalizes this objective as the selective reduction of statistical dependence—typically measured via Shannon mutual information—between target representations and the attributes to be forgotten, while ensuring maximal preservation of utility-relevant information. This paradigm provides mathematically rigorous objectives, optimization schemes, and guarantees, enabling principled design of algorithms that operate in settings as diverse as recommender systems, federated models, deep feature spaces, and multi-modal neural architectures.
1. Formal Information-Theoretic Objectives for Attribute Unlearning
Attribute unlearning is cast as an optimization over the encoder or data transformation parameters to maximize the retained information about task-relevant signals and inputs, while minimizing the mutual information with sensitive attributes. Let denote the original input, the main task label, the attribute to be unlearned, and the learned representation.
The canonical information-theoretic objective is: where is a utility functional such as (task-fidelity) or (input preservation), and quantifies the attribute's residual footprint. In soft-constraint form, this becomes: for a Lagrange parameter (Guo et al., 2022, Xu et al., 8 Feb 2025).
In frameworks supporting multi-attribute unlearning (e.g., MaSS), the problem further generalizes to: where is the transformed sample, are sensitive attributes, utility attributes, denotes unannotated generic information, and are per-attribute budget parameters (Chen et al., 2024). Attribute unlearning in federated or distributed models defines analogous criteria with respect to the information Fisher-score of model parameters about attributes or client data (Balordi et al., 26 Aug 2025).
2. Surrogate Losses and Mutual Information Estimation
Direct estimation of mutual information between high-dimensional representations and discrete/categorical attributes is generally intractable. Practical frameworks employ variational upper bounds or proxy divergences.
- Upper bounds via Variational Classifiers: The vCLUB bound, for attribute and embedding , is (Yu et al., 23 Oct 2025):
for a learned variational classifier .
- Distributional Divergence (MMD): In kernel-based approaches (e.g., post-training attribute unlearning), distinguishability is measured by the squared Maximum Mean Discrepancy between attribute-conditional embedding distributions (Chen et al., 2024, Li et al., 2023):
where is the distribution of embeddings conditioned on attribute class , and is an RKHS.
- Adversarial and Contrastive Surrogates: Adversarial classifiers (cross-entropy losses) and contrastive InfoNCE-type objectives are used to penalize mutual information with certain attributes while maximizing predictive information for utility attributes (Chen et al., 2024).
- Jacobian-Norm Minimization: In local information-gain frameworks, the squared norm of the network Jacobian along the attribute axes is minimized, suppressing the transfer of information about the forgotten subspace (Foster et al., 2024):
serving as a first-order proxy for .
3. Unified Optimization and Algorithmic Strategies
Multiple algorithmic instantiations arise for optimizing surrogate losses subject to information constraints:
- Training-phase MI regularization: MI-based regularizers are integrated into the representation learning process, with closed-form bounds (e.g., InfoFiltra) supporting stepwise surrogate minimization (Guo et al., 2022).
- Post-training Editing: Models can be edited post hoc, by optimizing only the embedding matrix or feature layers using distributional or adversarial losses, combined with functional/parameter-space regularization to protect utility (Li et al., 2023, Chen et al., 2024).
- Two-step Procedures and Combinatorial Unlearning: Multi-attribute settings exploit parallelizable attribute-wise calibration steps, followed by a convex combination search to align embeddings to minimize information leakage on all targeted attributes (Yu et al., 23 Oct 2025).
A summary of MI estimation and surrogate tasks appears below:
| MI estimation | Optimization phase | Approach |
|---|---|---|
| vCLUB bound | Training/post-train | Variational MI |
| MMD (& barycenter) | Post-train | RKHS divergence |
| Adversarial CE | Training/post-train | Classifier-based |
| InfoNCE contrastive | Both | Mutual info proxy |
| Jacobian norm | Post-train | Output sensitivity |
4. Theoretical Guarantees and Trade-off Bounds
Rigorous bounds characterize Pareto trade-offs between utility retention and attribute removal:
- Surrogate Loss Tightness: For InfoFiltra, it is shown that the optimized upper-bound loss satisfies (Guo et al., 2022):
and under additional conditions, the slack is controlled by:
- Feasibility Bounds: In selective suppression with utility lower bounds and sensitive-attribute upper bounds, the existence of a feasible mapping requires (Chen et al., 2024):
and for unannotated utility, the achievable retained information is upper-bounded by .
- Distributional Pareto Frontier: In distributional unlearning, the Gaussian Pareto frontier of minimum deletions versus preservation is analytically given by (Allouah et al., 20 Jul 2025):
with .
- Federated Unlearning: Per-parameter Target Information Score (TIS) quantifies the impact of unlearning:
Resetting high-TIS parameters followed by minimal retraining provably increases the adversary's error in inferring attribute presence (Balordi et al., 26 Aug 2025).
5. Practical Schemes and Empirical Evidence
Frameworks instantiate the above theory in various modalities and tasks:
- LEGO: Employs parallelizable embedding calibration and a flexible convex-combination layer to support simultaneous/dynamic multi-attribute removal. It minimizes variational MI upper bounds subject to -proximity constraints, with theoretical approximation-ratio guarantees. Empirically, LEGO achieves a point drop in attribute-inference attacker BAcc and NDCG@10 loss (Yu et al., 23 Oct 2025).
- MaSS: Introduces adversarial cross-entropy on labeled attributes, contrastive InfoNCE for unannotated utility, and normalizes performance via "Normalized Accuracy Gain"; it outperforms alternatives across audio, image, and sensor datasets (Chen et al., 2024).
- PoT-AU/D2D-FR: Distributional metric unlearning using MMD for class separation enables post hoc attribute removal at high efficiency (seconds vs retrain), with attacker BAcc reduced to chance levels and only $1$– drop in HR@10/NDCG@10 (Chen et al., 2024, Li et al., 2023).
- Jacobian Smoothing: Just-in-Time attribute unlearning minimizes local information gain along attribute axes, yielding fast, output-level invariance to target attributes (Foster et al., 2024).
- Wasserstein Barycenter: Feature-unlearning via optimal transport of attribute-conditional distributions to the barycenter gives a unified, closed-form solution for multiple objectives and admits efficient computation with neural OT maps (Xu et al., 8 Feb 2025).
- Federated Unlearning: Hessian-diagonal scoring and targeted re-initialization generalizes attribute unlearning to federated and distributed settings (Balordi et al., 26 Aug 2025).
Empirical evaluations consistently show strong reduction in attribute inference, little degradation to main task metrics, and vastly improved computational and deletion efficiency relative to blind retraining or adversarial in-training unlearning.
6. Assessment, Metrics, and Generalization
Mutual-information–based metrics serve both as unlearning objectives and post hoc assessment tools:
- Information Difference Index (IDI): Quantifies retained mutual information about forgotten attributes in intermediate layers; indicates full erasure, no removal (Jeon et al., 2024). IDI leverages InfoNCE-based estimation and robustly separates truly unlearned models from those only masking outputs.
- Trade-off analyses: Many frameworks present empirical utility–privacy Pareto frontiers, exhibiting minimal main-task loss up to near-complete attribute suppression (Yu et al., 23 Oct 2025, Chen et al., 2024, Allouah et al., 20 Jul 2025).
- Extension to Arbitrary Attributes: All frameworks support multi-class, continuous, and even arbitrarily structured attribute unlearning by adapting the MI estimators, proxy losses, and regularization strengths accordingly (Chen et al., 2024, Yu et al., 23 Oct 2025, Xu et al., 8 Feb 2025).
7. Open Problems and Significance
While information-theoretic frameworks for attribute unlearning deliver closed-form objectives, surrogate bounds, and post hoc metrics, some limitations persist. These include the computational cost of high-dimensional MI estimation, the explicit-knowledge assumption about sensitive attributes for labelled suppression, and the non-convexity of deep representation mapping spaces. Nonetheless, these frameworks supply the first rigorous foundations for selective attribute removal, unifying disparate strategies under the powerful machinery of mutual information, and enabling efficient, theoretically justified implementations across a range of modern machine learning architectures (Guo et al., 2022, Yu et al., 23 Oct 2025, Chen et al., 2024, Chen et al., 2024, Xu et al., 8 Feb 2025, Allouah et al., 20 Jul 2025, Jeon et al., 2024, Balordi et al., 26 Aug 2025).