Optimized Coreset Selection
- Optimized coreset selection is a technique that builds compact, representative data subsets to yield full-data performance with reduced computational cost.
- It leverages methods such as hardness-based, training-free submodular selection, and adaptive pruning to enhance robustness against noise and adversarial attacks.
- State-of-the-art approaches like EasyCore, SubZeroCore, HyperCore, and FAST offer formal guarantees, improved scalability, and significant accuracy gains in diverse learning tasks.
Optimized coreset selection refers to the design and application of algorithms that construct compact, representative subsets (“coresets”) of large datasets, such that training on the coreset efficiently yields model performance comparable to full-data training—even under adverse, noisy, or adversarial regimes. Recent advances formalize, analyze, and empirically validate coreset selection methods with strong theoretical guarantees and state-of-the-art empirical performance across robustness, scalability, and label-noise resilience, leading to new paradigms for efficient data-centric learning (Ramesh et al., 13 Oct 2025, Moser et al., 26 Sep 2025, Moser et al., 26 Sep 2025).
1. Hardness-based Coreset Selection and Adversarial Robustness
A principal recent breakthrough is the explicit linking of sample-wise “hardness”—quantified via the Average Input Gradient Norm (AIGN)—to adversarial vulnerability and model robustness (Ramesh et al., 13 Oct 2025). The AIGN of sample for a model trained over epochs is defined as
where is the per-sample loss. Sorting samples by low AIGN yields “easy” examples, which lie farthest from the decision boundary, exhibit high adversarial robustness, and induce wider margins when used preferentially in training.
The EasyCore algorithm proceeds by (a) accumulating per-sample input gradients over epochs, (b) computing AIGN per sample, and (c) statically selecting the lowest-AIGN fraction (commonly to ). Training on this coreset, under both standard and adversarial pipelines (e.g., PGD, TRADES), consistently yields up to percentage points in adversarial accuracy compared to uniform, gradient-matching, or k-center baselines. Theoretical analysis attributes the gains to increased decision boundary margins and reduced boundary curvature, with high-AIGN example removal freeing model capacity to learn generalizable, robust features (Ramesh et al., 13 Oct 2025).
2. Submodular, Training-Free, and Density-Aware Selection
Optimized coreset methods have advanced beyond gradient or label-centric criteria to entirely training-free selection. SubZeroCore achieves this by combining submodular coverage and density weighting into a single monotone objective (Moser et al., 26 Sep 2025): where is a density score for candidate , computed using a closed-form expectation over -nearest neighbor radii, tuned via a single hyperparameter controlling density–coverage trade-off.
SubZeroCore is fully label and training agnostic, requiring only embedding similarities; its greedy maximization yields a -approximation to the optimal subset. Empirically, SubZeroCore matches or surpasses gradient-based coreset methods at high pruning ratios (), with superior runtime (e.g., $1$ min for CIFAR-10 at vs. $10$–$50$ min for training-based signals) and robustness under label noise injections. These properties make SubZeroCore highly scalable and robust for large, noisy, or web-scale data (Moser et al., 26 Sep 2025).
3. Robust and Adaptive Coreset Selection Under Annotation Noise
HyperCore introduces class-wise hypersphere models for robust coreset selection in environments with noisy or ambiguous labels (Moser et al., 26 Sep 2025). For each class , a small neural network maps samples to , and in-class samples (label ) are clustered near the origin inside a hypersphere, while outliers (mislabeled, corrupted, or out-of-class points) lie outside.
Pruning is performed adaptively by maximizing Youden’s statistic over the conformity distance distribution within and outside each class: where TPR is the true positive rate for in-class points at radius , and FPR is the false positive rate for outliers. The optimal retention threshold is set to maximize . This process discards ambiguous and noisy examples while keeping the informative core.
Experiments demonstrate HyperCore’s superiority under both moderate and severe label noise, low-data regimes, and imbalanced conditions: at retention, accuracy gains over best baselines reach points (Moser et al., 26 Sep 2025). The method is highly parallelizable and efficiently scalable.
4. Distribution Matching, Topological Constraints, and Frequency-Domain Methods
FAST leverages graph spectral theory and distributional matching in the frequency domain for DNN-free optimized coreset selection (Cui et al., 22 Nov 2025). The task is formulated as minimizing the characteristic function distance (CFD) between the empirical distribution of the full dataset and candidate coreset : where is the empirical characteristic function evaluated at frequency .
To overcome the “vanishing phase gradient” in mid/high frequency regimes, FAST introduces an attenuated phase-decoupled CFD loss, with targeted penalty in regions where only the phase conveys essential distributional information. Topology is preserved by extracting Laplacian eigenvectors from a multi-scale graph on , and the selection is constrained to align in manifold space by Hungarian matching and Laplacian regularization. Progressive Discrepancy-Aware Sampling schedules frequency sampling from low to high to match global structure prior to local refinement.
Benchmarks show FAST achieves up to accuracy gains over SOTA DNN-free methods, CPU speedup, and a reduction in power consumption, with cross-model generalization loss (Cui et al., 22 Nov 2025).
5. Coreset Selection for Structured and Non-classification Tasks
Optimized coreset selection has been adapted to object detection, continual learning, and dynamic systems identification. The CSOD framework for object detection aggregates classwise and imagewise RoI features, then applies a submodular facility-location objective balancing representativeness and diversity (Lee et al., 14 Apr 2024). Greedy maximization yields a performance boost of points in AP over random selection on Pascal VOC, with scalability to BDD100k and COCO2017.
Online coreset selection in dynamic systems uses geometric penetration criteria and polyhedral volume reduction guarantees for system identification, maintaining vanishing selection rates ( over timesteps) while ensuring convergence of the feasible parameter set (Li et al., 28 Jun 2025).
For rehearsal-based continual learning, OCS maximizes gradient similarity scores online for current adaptation and memory affinity, yielding strong empirical results across standard, imbalanced, and noisy benchmarks (Yoon et al., 2021).
6. Theoretical Guarantees and Empirical Performance
State-of-the-art optimized coreset selection methods exhibit formal approximation guarantees and efficiency bounds. SubZeroCore and CSOD achieve provable approximations to their objectives via greedy selection (Moser et al., 26 Sep 2025, Lee et al., 14 Apr 2024). Distribution-matching methods (Cui et al., 22 Nov 2025, Kokot et al., 28 Apr 2025) show poly-logarithmic or exponential sample size reduction for lossless compression under smooth divergences and RKHS spectral decay, with matching error rates to random sampling.
Hardness-based selection (Ramesh et al., 13 Oct 2025) demonstrates statistically significant robustness gains under adversarial training, outperforming model-dependent, gradient-based, and dynamic coreset methods by points adv-accuracy (CIFAR-100, PGD-20). Robust selection methods such as HyperCore adaptively prune noisy data and outperform fixed-ratio selectors under rising label noise (Moser et al., 26 Sep 2025). FAST and SubZeroCore offer massive reductions in compute and power costs while preserving or improving downstream accuracy (Cui et al., 22 Nov 2025, Moser et al., 26 Sep 2025).
7. Integration and Deployment Considerations
Optimized coreset selection algorithms are highly modular. EasyCore requires only a single pre-training gradient accumulation run for AIGN computation; its scores are transferable across architectures and training protocols (Ramesh et al., 13 Oct 2025). SubZeroCore can operate on raw embeddings from any pretrained model, needing no labels or network training, and is suitable for large-scale noisy or unlabeled data (Moser et al., 26 Sep 2025). HyperCore trains small MLPs per class and requires no tuning of coreset size, automatically adjusting via Youden’s statistic (Moser et al., 26 Sep 2025). FAST only needs Laplacian graph construction and frequency library sampling, enabling efficient execution on edge devices (Cui et al., 22 Nov 2025).
Empirical guidance recommends coreset fractions in the range –, with class-balancing and feature-space diversity constraints applied for high-imbalance or multi-modal settings. All leading approaches exceed or match baseline performance across classification, detection, continual learning, and regression tasks, with the added benefit of substantial computational savings.
Optimized coreset selection now offers rigorous, scalable, and robust frameworks for reducing data and computational burdens in deep learning, with theoretical guarantees, superior noisiness resilience, and ease of integration into diverse data-centric pipelines (Ramesh et al., 13 Oct 2025, Moser et al., 26 Sep 2025, Moser et al., 26 Sep 2025, Cui et al., 22 Nov 2025).