Papers
Topics
Authors
Recent
2000 character limit reached

Automated Data Selection in ML

Updated 11 December 2025
  • Automated data selection is an algorithmic process that identifies optimal data subsets based on relevance, diversity, and representativeness.
  • It employs methods such as representation-based scoring, greedy optimization, and reinforcement learning to enhance computational efficiency and model performance.
  • Applications span speech recognition, computer vision, and scientific data curation while addressing challenges like domain-shift and noisy data.

Automated data selection refers to algorithmic frameworks and methods that systematically select informative, representative, or otherwise optimal subsets of data from larger pools, according to explicit criteria, for downstream tasks such as model training, validation, evaluation, or analysis. Its scope spans sample, feature, and variable selection and is motivated by both computational efficiency (reducing training cost) and improvements in model generalization, domain adaptation, and scientific analysis. Automated data selection is central to large-scale machine learning, data-driven science, and high-dimensional inference.

1. Key Principles and Theoretical Foundations

Automated data selection relies on the premise that using all available data can lead to suboptimal or even degraded performance compared to carefully curated subsets. This assertion is empirically validated in domains such as speech recognition, where in-domain data is critical for robust model adaptation and out-of-domain data can introduce distributional mismatch, increasing error rates (Mortaza et al., 2019). Automated data selection aims to identify data that maximizes task-relevant information, covers the desired distribution, or aligns with a specific downstream objective (e.g., minimizing word error rate, maximizing held-out task accuracy, or matching metadata targets). The major guiding principles include:

  • Relevance: Select data most similar or informative with respect to a target domain or query set.
  • Diversity: Avoid redundancy by selecting data that is sufficiently diverse or orthogonal in feature space.
  • Representativeness: Ensure the selected subset's distribution matches the target or expected distribution in a well-defined feature or metadata space.
  • Robustness: Discard data points with high likelihood of corruption, label noise, or outlier behavior.

The process typically involves defining a scoring or similarity function over the data pool and optimizing a selection objective, often constrained by computational or annotation budget.

2. Methodologies and Algorithms

Automated data selection comprises a diverse methodological landscape, spanning unsupervised, supervised, and reinforcement learning paradigms. Key approaches include:

a. Representation-Based Selection

Data points are represented as feature vectors (latent representations, embeddings, Dirichlet posteriors, etc.), and similarity metrics (e.g., cosine similarity, Euclidean distance) are used to select points close to a desired prototype or sparse basis. Notable examples include:

  • aLDA-based selection: Each utterance is mapped to a high-dimensional Dirichlet posterior vector via latent Dirichlet allocation over quantized "acoustic words." Cosine distance to domain prototypes is used to select samples closest to the target distribution (Mortaza et al., 2019).
  • RDS+ for instruction tuning: Last-layer hidden states from pretrained LLMs are weighted, pooled representations; round-robin matching is performed to maximize alignment with a target query set (Ivison et al., 3 Mar 2025).

Subset selection is posed as an optimization problem, e.g., minimizing the L1L_1 distance between selected and target metadata distributions (Trinh et al., 16 Jul 2024), or maximizing similarity-based utility functions subject to cardinality constraints.

c. Ensemble and Hybrid Selection

Multiple feature- or variable-selection techniques (filter, wrapper, embedded) are combined in an ensemble, with their outputs merged via intersection, union, or majority heuristics to stabilize and optimize subset selection (see Table 1) (Nakashima et al., 2019).

Table 1. Aggregation Heuristics in Ensemble Selection (Nakashima et al., 2019)

Heuristic Operation Typical Output Size
Union mSm(t)\bigcup_m S_m^{(t)} Largest
Intersection mSm(t)\bigcap_m S_m^{(t)} Smallest
Quorum Features selected by q\ge q methods Intermediate

d. Reinforcement Learning and Adaptive Schedules

Formulate data selection as a Markov Decision Process where filtering decisions are actions, and rewards depend on training trajectories or convergence. The Neural Data Filter (NDF) learns policies for SGD mini-batch selection that optimize downstream rewards (e.g., speed of convergence, validation accuracy) (Fan et al., 2017).

e. Gradient-Based Coresets

Methods such as AUTOMATA select subsets whose (possibly weighted) aggregate gradient closely approximates the full-data gradient, using techniques like orthogonal matching pursuit, facilitating compute-efficient hyperparameter tuning and meta-learning (Killamsetty et al., 2022).

f. Multi-Modal and Information-Theoretic Selection

Frameworks such as CLIP-powered data selection leverage joint image–text representations to score samples by semantic alignment and diversity, and solve the multi-objective subset selection problem via small-scale continuous optimization (Yang et al., 15 Oct 2024).

g. Online Feature Selection and Negotiation

Systems such as MOANOFS aggregate multiple online learners, negotiating feature choices based on trust and multi-objective optimization (accuracy, speed, confidence); feature selection decisions are made online as instances arrive (BenSaid et al., 2018).

3. Domain-Specific Applications

Speech Technology

Latent Dirichlet Allocation-based selection aligns acoustic data distribution with a small in-domain set, substantially reducing word error rates compared to using all data or random selection (Mortaza et al., 2019).

Dynamic proxy subset selection frameworks such as ASP and AUTOMATA provide substantial speedups (2×2\times30×30\times) for NAS and HPO, with negligible loss in architecture or hyperparameter ranking fidelity. These systems use mixtures of uncertainty, loss, or gradient metrics, applied epoch-wise or per configuration (Yao et al., 2023, Killamsetty et al., 2022).

Computer Vision and Multimodal Tasks

FreeSel exploits pretrained vision transformers for single-pass, pattern-level semantic sampling, providing 530×530\times faster selection than iterative active learning with state-of-the-art performance (Xie et al., 2023). CLIP-powered optimization leverages joint image–text spaces for robust sample selection even under high label noise or corruption (Yang et al., 15 Oct 2024).

Scientific Data Curation

LOTUS automates satellite data culling using ensemble classifiers, temporal aggregation, and run-length postprocessing to match or exceed expert-level accuracy in domain-specific classification tasks (Stricklin et al., 13 Mar 2024).

Safety Validation

Metadata-driven selection aligns selected subsets' empirical distribution to expert-specified targets, as in autonomous vehicle scenario validation where precise category-level quotas are enforced for highly reliable and auditable validation protocols (Trinh et al., 16 Jul 2024).

Feature and Variable Selection

Automated pipelines (filter, wrapper, embedded, ensemble) select minimal, non-redundant feature sets for tasks such as anomaly detection and clinical subpopulation analysis, preserving predictive or scan statistic performance with sharp gains in computational efficiency (Nakashima et al., 2019, Wanjiru et al., 2021).

4. Empirical Performance and Practical Considerations

Empirical findings consistently underscore several points:

  • Subset selection outperforms full-dataset training in many settings by reducing domain mismatch, label noise, or computational redundancy (Mortaza et al., 2019, Yao et al., 2023).
  • Selection scale is critical: Methods that outperform random at small pool sizes often degrade or underperform at million-scale selection (e.g., Top-PPL, IFD, gradient-influence), while representation-based methods like RDS+ scale robustly (Ivison et al., 3 Mar 2025).
  • Efficiency: Contemporary frameworks achieve order-of-magnitude speedups over active learning or random search, with selection modules adding negligible overhead—typically O(N)O(N)O(NlogN)O(N \log N) in the number of samples (Xie et al., 2023, Yang et al., 15 Oct 2024, Trinh et al., 16 Jul 2024).
  • Robustness to Noise/Corruption: Multi-modal and alignment-based selection frameworks effectively suppress the inclusion of noisy or irrelevant data, reducing post-selection error rates (Yang et al., 15 Oct 2024).
  • Dynamic/Adaptive Selection: Systems such as NDF and ASP alternate between exploration (random or high-uncertainty samples) and exploitation (high-gradient, hard examples), adapting the selection policy as model learning progresses (Fan et al., 2017, Yao et al., 2023).
  • Multi-objective and Constraint Handling: Practical implementations typically incorporate constraints on subset size, budget, diversity, or metadata quotas for regulatory compliance or experimental design (Trinh et al., 16 Jul 2024, BenSaid et al., 2018).

5. Challenges, Limitations, and Open Questions

Despite rapid advances, substantive limitations persist:

  • Sensitivity to data/label noise: Some selection heuristics, especially those relying on loss or gradient magnitude, are susceptible to adversarial or highly noisy samples.
  • Domain-shift and transferability: Pretrained-model-based selection can suffer when faced with out-of-distribution data; adapters and dynamic adaptation can mitigate, but not eliminate, this risk (Yang et al., 15 Oct 2024).
  • Automated confounder selection risks: In high-dimensional causal inference, data-driven variable selection can inadvertently select endogenous “bad controls,” introducing substantial bias unless causal structure is explicitly modeled (Hünermund et al., 2021).
  • Selection-size tuning: Many schemas require careful tuning of selection ratios or threshold hyperparameters; over-aggressive culling can undermine downstream generalizability, while undersampling blunts efficiency gains (Ivison et al., 3 Mar 2025).
  • Scalability of optimization: Large-scale similarity or gradient-matching entails computational challenges; greedy approximations, batch selection, or single-pass inference serve as mitigations (Killamsetty et al., 2022, Xie et al., 2023).
  • Automation vs. domain knowledge: Fully automated selection—especially for features/covariates—cannot substitute for substantive domain expertise in causal or safety-critical applications (Hünermund et al., 2021, Nakashima et al., 2019).

6. Future Directions

Research in automated data selection continues to develop:

  • Compositional Selection: Task-adaptive weighting and joint modeling of mixed objectives (e.g., safety, diversity, fairness) within end-to-end optimization frameworks.
  • Efficient Large-Pool Search: Scalable sublinear or quantized selection methods for hundred-million-example or streaming data regimes.
  • Federated and Continual Selection: Extension to decentralized/federated settings (local coreset selection, federated HPO) and continual learners (Killamsetty et al., 2022).
  • Integrated Uncertainty and Bias Control: Filtering for data quality, safety, bias, and toxicity directly as part of the selection mechanism (Ivison et al., 3 Mar 2025).
  • Causal-Aware Selection: Incorporating causal graphs and d-separation criteria into selection, especially for confounder control in high-dimensional causal inference (Hünermund et al., 2021).

Automated data selection frameworks are now integral to data-efficient, reliable large-scale learning, scientific discovery, and safety validation. Ongoing work seeks to unify adaptive, scalable, and robust data selection into modular pipelines deployable across a wide range of domains and modalities.

Whiteboard

Follow Topic

Get notified by email when new papers are published related to Automated Data Selection.