Papers
Topics
Authors
Recent
2000 character limit reached

Adaptive Subset Selection

Updated 5 December 2025
  • Adaptive subset selection is a method for dynamically choosing data or features based on model state and empirical criteria, improving sampling efficiency and generalization.
  • It employs techniques like importance sampling, adaptive weighting, and bandit-guided strategies to iteratively update selection policies in response to observed outcomes.
  • This approach is applied in domains such as active learning, neural architecture search, sparse recovery, and experimental design to reduce computational cost and enhance performance.

Adaptive subset selection is a methodological paradigm in statistical learning, optimization, and machine learning wherein data points, variables, parameters, features, or scenarios are selected dynamically in response to model state, criteria, or observed outcomes, typically with the goal of improving sample efficiency, generalization, computational efficiency, or adaptability to underlying structure. Adaptive mechanisms contrast with static or one-shot selection by iteratively updating selection policies in response to ongoing partial model fits, empirical losses, structural scores, or external validation. The field encompasses algorithmic, theoretical, and empirical studies spanning batch learning, active learning, sparse recovery, neural architecture search, experimental design, optimization under uncertainty, parameter-efficient fine-tuning, and more. Core facets include importance sampling, entropy-driven uncertainty, adaptive sampling, dynamic weighting, bandit-guided selection, low-dimensional subspace pursuit, and correlational leveraging.

1. Formal Problem Definitions and General Principles

Adaptive subset selection typically addresses the cardinality-constrained maximization or risk-minimization problem over subsets of a ground set P\mathcal{P} (e.g., data pool, feature space, parameter blocks, or scenarios). In batch streaming or online contexts, the selector must choose subsets SrS_r of examples at each round based on properties of previously trained models or adaptive scores, subject to fixed seed, batch, or total subset size constraints. The subset selection algorithm can be formalized as: minSP, Sk L(S,θ)\min_{S \subseteq \mathcal{P},\ |S| \leq k}\ L(S,\theta) or in bi-objective settings,

minA{1,,K}(LA,  Aw)\min_{A \subseteq \{1,\dots,K\}} \left( L_A, \; \frac{|A|}{|w|} \right)

where L(S,θ)L(S, \theta) denotes loss (potentially with adaptive weighting), and LAL_A is the minimal achievable loss with only parameter blocks in AA updated (Xu et al., 18 May 2025).

Key adaptive protocols involve importance-weighted sampling (IWeS) (Citovsky et al., 2023), disagreement-based entropy selection, adaptive probability vectors for scenarios (Miyagi et al., 2022), or updating inclusion probabilities for variables and features (Staerk et al., 2019), always conditioning future selection decisions on model or selection state.

2. Algorithmic Frameworks and Adaptive Mechanisms

Multiple adaptive subset selection algorithms have been developed:

  • Importance Weighted Subset Selection (IWeS): In each batch round, examples are sampled using probabilities determined by entropy or model disagreement. Selected examples receive adaptive importance weights w=min{1/p(x,y),u}w=\min\{1/p(x,y),u\}, capped to control variance. Empirical loss is minimized over selected examples with these weights. The algorithm distinguishes between "IWeS-dis" (labels available, uses disagreement between two models) and "IWeS-ent" (active-learning variant, label free, uses normalized entropy). Retraining occurs only after batch completion (Citovsky et al., 2023).
  • Adaptive Sampling for Differential Submodularity: For objectives with (weak) submodularity, the Dash algorithm exploits an adaptive batchwise filtering-sampling protocol. Marginals for batch addition are evaluated in parallel, leveraging differential submodularity constants α\alpha to guarantee approximation. The filtering phase reduces the pool size geometrically, and batch sampling proceeds with guaranteed bounds. The approach achieves O(logn)O(\log n) adaptivity (Qian et al., 2019).
  • Adaptive Subspace Variable Selection (AdaSub): The algorithm proceeds by iteratively sampling low-dimensional variable subspaces, solving submodel selection problems (via penalized information criteria), and updating inclusion probabilities rjr_j for future rounds. This mechanism adaptively concentrates probability on influential variables as determined by their frequent selection in optimal submodels (Staerk et al., 2019).
  • Bandit-Guided Submodular Curriculum (ONLINESUBMOD): Adaptive subset selection is cast into a multi-armed bandit framework, where each arm is a (submodular) utility function (e.g., facility-location, log-determinant, disparity-sum). The policy adaptively explores/exploits over arms via an online greedy maximization, with rewards driven by validation loss reduction. Regret bounds are established for per-round and cumulative regret with respect to ideal arm (Chanda et al., 28 Nov 2025).
  • Adaptive Scenario Subset Selection (AS3): In worst-case optimization, scenario support is tracked via adaptive probability vectors p(t)p^{(t)}. Scenarios most relevant to local optima are assigned higher selection probability and updated based on observed candidate solutions in a stochastic search (CMA-ES), reducing simulation cost (Miyagi et al., 2022).

3. Theoretical Guarantees and Analytical Results

Multiple rigorous results underpin adaptive subset selection:

  • Generalization Bounds: For importance-weighted adaptive sampling (IWeS-V), with bounded loss and finite hypothesis class, the generalization gap (to optimal in-class predictor) is O(log(T/δ)/T)O(\sqrt{{\log(T/\delta)}/{T}}). Uniform martingale concentration is key. Sampling rate bounds leverage the (subset-selection) disagreement coefficient θS\theta_S, establishing label efficiency under bounded loss (Citovsky et al., 2023).
  • Approximation Guarantees for Differential Submodularity: The Dash algorithm matches or closely trails greedy in objective value, with parallel runtime scaling logarithmically in nn. The (1exp(α2)ε)(1-\exp(-\alpha^2)-\varepsilon) approximation ratio is achieved with O(logn)O(\log n) adaptive rounds, far faster than the classical greedy sequence (Qian et al., 2019).
  • Consistency in Variable Selection (AdaSub): Under the Ordered Importance Property, inclusion probabilities for truly relevant variables converge to 1, those for spurious variables converge to 0, with probability 1 as tt \to \infty. Thus, AdaSub asymptotically recovers the optimal variable subset under the selected criterion (Staerk et al., 2019).
  • Bandit Regret Bounds: In ONLINESUBMOD, expected per-round regret decays as O(1/t+logt/t)O(1/t+\log t/t) under fractional exploration and utility gap assumptions. This establishes no-regret online adaptive curriculum selection (Chanda et al., 28 Nov 2025).

4. Domain-Specific Instantiations and Practical Contexts

Adaptive subset selection algorithms are adapted to various domains:

  • Batch Streaming and Active Learning: IWeS applies to classical batch selection, streaming selection, and pool-based active learning. The entropy-driven variant remains competitive when labels are unavailable and achieves empirical superiority in multi-label tasks (OpenImages v6) (Citovsky et al., 2023).
  • Neural Architecture Search (NAS): Adaptive subset selection (GLISTER-NAS) accelerates NAS algorithms by recurrently optimizing per-step subset selection with respect to meta-objective (validation loss), using greedy submodular optimization and bi-level approximations. Substantial accuracy is preserved at reduced computational cost (C et al., 2022).
  • Scenario Selection in Optimization: AS3-CMA-ES adaptively selects support scenarios for min-max optimization, significantly reducing simulation counts and improving restart efficiency in applications such as CO2_2 sequestration well placement (Miyagi et al., 2022).
  • Sparse Recovery and Feature Selection: In high-dimensional regression and experimental design, adaptive sampling algorithms with differential submodularity enable feature selection and optimal experimental design in parallel, matching greedy quality but with greatly accelerated runtime (Qian et al., 2019).
  • Curriculum Learning: Bandit-guided curriculum selects data points in minibatches for efficient training in both supervised classification and large-scale LLM fine-tuning, outperforming static and bi-level selection in speed and accuracy (Chanda et al., 28 Nov 2025).

5. Limitations, Computational Complexity, and Open Questions

Noteworthy limitations and open directions include:

  • Computational Overheads: Importance-weighted selection with double model retraining (IWeS-dis) increases per-round cost; label-free variants offer moderate savings but may sacrifice information. AdaSub and adaptive greedy approaches may require many model fits under high-dimensionality, though warm-starting and batched selection can offset these costs (Citovsky et al., 2023, Staerk et al., 2019).
  • Variance Control and Diversity: Weight cap uu in IWeS, or entropy regularization in SubSelNet-style neural selectors, mitigate variance instability but add hyperparameter tuning. Explicit diversity modeling (e.g., submodular selection in curriculum, facility-location or logdet scores in ONLINESUBMOD) may be necessary for large-batch settings (Citovsky et al., 2023, Chanda et al., 28 Nov 2025, Jain et al., 18 Sep 2024).
  • Theoretical Extensions: Extensions to batch-streaming regimes with delayed feedback, sharper analysis for label-free variants (IWeS-ent), and more comprehensive disagreement geometry studies remain open (Citovsky et al., 2023).
  • Non-additive Effects and Amplification: Greedy influence-based heuristics for subset selection may fail to capture non-additive group effects. Adaptive greedy or higher-order approximations can partially remedy, at increased computational cost (Hu et al., 25 Sep 2024).

6. Comparative Performance and Empirical Evaluation

Empirical benchmarks show adaptive subset selection algorithms consistently match or outperform static, random, or uncertainty-based baselines across tasks:

Algorithm Context Key Benchmark Accuracy / Utility Speedup / Cost Reduction
IWeS-dis/ent Image, tabular CIFAR-10/100, SVHN Superior or equal to BADGE, Uncertainty, Coreset N/A; batch training only
ONLINESUBMOD Curriculum CIFAR-100, LLM MMLU Highest test accuracy, lowest regret 6×\times (classification), faster convergence (LLM)
GLISTER-NAS NAS CIFAR-10/100, ImageNet Full-data performance at 10-20% subset 10×\times faster search
AdaSub Feature selection Simulated, PCR data Better than stability selection, matches EBIC optimal Efficient for p2000p\le 2000
AS3-CMA-ES Optimization Well placement, synthetic Higher median solution (well placement), matches oracle k-scenarios 5×\times fewer simulations

On large public datasets (CIFAR, SVHN, OpenImages), importance-weighted and bandit-adaptive selection algorithms reliably achieve near-full accuracy with 10–20% data, and adaptive scenario selection dramatically improves sample efficiency for worst-case optimization problems (Citovsky et al., 2023, C et al., 2022, Chanda et al., 28 Nov 2025, Miyagi et al., 2022).

Adaptive subset selection is closely related to active learning, submodular maximization, experimental design, compressed sensing, coresets, and curriculum learning. The use of bandit policies, neural surrogates (SubSelNet), Pareto-optimized knapsack selection for PEFT, and scalable parallel adaptive sampling expands the applicability of adaptive subset selection to large-scale models, resource-constrained training, AutoML, and parameter-efficient fine-tuning (Jain et al., 18 Sep 2024, Xu et al., 18 May 2025). Future research directions include more nuanced diversity modeling, integration with federated and continual learning, robust adaptive selection in non-i.i.d. regimes, and more efficient group effect modeling in influential sample selection (Hu et al., 25 Sep 2024).

Slide Deck Streamline Icon: https://streamlinehq.com

Whiteboard

Forward Email Streamline Icon: https://streamlinehq.com

Follow Topic

Get notified by email when new papers are published related to Adaptive Subset Selection.