Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 169 tok/s
Gemini 2.5 Pro 53 tok/s Pro
GPT-5 Medium 31 tok/s Pro
GPT-5 High 38 tok/s Pro
GPT-4o 104 tok/s Pro
Kimi K2 191 tok/s Pro
GPT OSS 120B 433 tok/s Pro
Claude Sonnet 4.5 37 tok/s Pro
2000 character limit reached

Exemplar Selection and Use

Updated 7 November 2025
  • Exemplar selection is the process of choosing a small subset of representative data points from a larger dataset to maximize information utility under resource constraints.
  • It employs optimization techniques such as submodular maximization, greedy algorithms, and bandit approaches to balance sparsity and reconstruction fidelity.
  • In applications like in-context learning, federated learning, and computer vision, exemplar selection enhances accuracy, efficiency, and data diversity.

Exemplar selection is the process of identifying a small subset of representative instances—termed "exemplars"—from a larger dataset, with the goal of efficiently summarizing, reconstructing, or adapting to new tasks. Exemplar-based approaches are central in numerous machine learning domains, including in-context learning with LLMs, continual and federated learning, computer vision, and data compression. Techniques for exemplar selection differ according to problem structure, theoretical guarantees, and application-specific constraints, but all are fundamentally concerned with maximizing information utility under resource limitations.

1. Theoretical Foundations and Models

Exemplar selection is typically formalized as an optimization problem, where one seeks to select a subset X0X_0 (of size at most kk) that best represents a dataset XX, often under explicit constraints. In the context of structured high-dimensional data, the selection problem can be stated as

minX0kFλ(X0)\min_{|X_0| \leq k} F_\lambda(X_0)

where Fλ(X0)F_\lambda(X_0) is an application-dependent cost function. For union-of-subspaces models, a prominent formulation involves sparse self-representation, measuring the worst-case reconstruction cost: fλ(xj,X0):=mincc1+λ2xji:xiX0cixi22f_\lambda(x_j, X_0) := \min_{c} \|c\|_1 + \frac{\lambda}{2} \left\| x_j - \sum_{i: x_i \in X_0} c_i x_i \right\|_2^2

Fλ(X0):=supxjXfλ(xj,X0)F_\lambda(X_0) := \sup_{x_j \in X} f_\lambda(x_j, X_0)

Here, λ>1\lambda > 1 regulates the trade-off between 1\ell_1 sparsity and fidelity. For matrix or subspace-based methods, geometric coverage can be measured via the Minkowski functional of the subset, linking exemplar selection to convex and spherical covering in high dimensions.

Many methods for exemplar selection are NP-hard in their general form. Approximations using submodularity, greedy algorithms (e.g., Frank-Wolfe iterative selection), or bandit formulations with surrogate objective modeling are prevalent, providing tractable solutions with theoretical approximation guarantees (e.g., (11/e)(1-1/e)-optimality for submodular maximization).

2. Algorithmic Methodologies

A broad class of exemplar selection algorithms falls into several methodological categories:

Category Core Mechanism Key Features
Greedy/Sparse Coding Iteratively select worst/worst-represented points, e.g., Farthest First Search (FFS) (You et al., 2020), Frank-Wolfe (Cheng et al., 2018) Focus on "hard" or under-represented examples; computationally efficient
Convex Optimization Relax boolean selection to group-lasso or lasso problems (Cheng et al., 2018) Allows gradient-based, scalable solutions; amenable to kernelization
Information-theoretic Maximize (approximate) submodular utilities, e.g., kernelized or D-optimal design-based methods (Singh et al., 19 Sep 2025) Balances query-specific relevance and coverage/diversity
Bandit and Black-box Stochastic/linear bandits, Bayesian optimization with surrogate modeling (Wu et al., 25 May 2024, Purohit et al., 10 Jun 2025, Purohit et al., 6 Nov 2024, Luo et al., 1 Oct 2025) Efficient in limited feedback settings; enables multi-objective optimization
Clustering/Geometric Spectral clustering, k-means, PCA-median (PBES) (Nokhwal et al., 2023, Resani et al., 12 Sep 2024) Captures multi-modal structure; robust to outliers or variance
Probabilistic Models Determinantal Point Processes (DPPs) for diversity (Santosh et al., 23 Jan 2025, Zhang et al., 2016) Favors jointly relevant but non-redundant sets

Exemplar order has been shown to strongly affect downstream model performance in in-context learning settings, motivating ordering-aware selection and joint prompt-exemplar optimization frameworks (Wu et al., 25 May 2024). For prompt-based tasks, order can be optimized alongside exemplar content using neural bandit and surrogate methods.

3. Applications Across Learning Paradigms

Exemplar selection plays a pivotal role in several distinct problem classes:

  1. In-Context Learning (ICL) with LLMs:
    • Fixed or query-specific exemplar sets are used as prompt demonstrations to condition LLM outputs.
    • Methods such as EASE (Wu et al., 25 May 2024) optimize ordered exemplar sequences globally for a task, while methods like CASE (Purohit et al., 10 Jun 2025) solve a top-m arm bandit to sample-efficiently identify best demonstration sets.
    • Multi-objective formulations explicitly jointly optimize accuracy and calibration (expected calibration error (ECE)) using combinatorial Bayesian optimization (COM-BOM (Luo et al., 1 Oct 2025)).
  2. Continual and Federated Learning:
    • Exemplar buffers (memory) mitigate catastrophic forgetting in sequential task or non-IID distributed settings. Robust selection requires condensation that matches the gradient and feature statistics of streaming data (Sun et al., 25 Dec 2024), sometimes using generative models for inter-client augmentation.
    • Gradient or influence-based criteria (e.g., HESIT (Chen et al., 16 May 2024)) select data with maximal positive effect on future generalization, avoiding expensive influence function Hessian inversions.
  3. Computer Vision and Summarization:
    • Exemplar-based subset selection by spectral clustering, DPP, or summary-structure transfer enables efficient dataset summarization, representative selection, or label-efficient recognition (Zhang et al., 2016, Nokhwal et al., 2023, Resani et al., 12 Sep 2024).
    • In VQA/VQG, semantic (not just visual) nearest neighbors are integrated as supporting/opposing exemplars, with differential losses promoting both relevance and diversity (Patro et al., 2019).
  4. Data Selection for Unsupervised Learning:
    • In dictionary learning, saliency- or error-driven active selection accelerates learning by prioritizing high-information ("supercharged") samples, provided diversity across latent atoms is maintained (Tsuchida et al., 2014).

4. Performance, Robustness, and Theoretical Guarantees

Strong analytical guarantees distinguish modern exemplar selection algorithms:

  • Subspace Coverage: Under union-of-independent-subspaces models, FFS guarantees selecting a basis for each subspace if kdk \geq \sum_\ell d_\ell, enabling subspace-preserving reconstruction and perfect clustering/classification under model assumptions (You et al., 2020).
  • Approximate Submodularity: Information-theoretical and D-optimal objectives confer (11/e)(1-1/e)-approximation through greedily selected sets. Query-specific submodularities enable both local relevance and global coverage (Singh et al., 19 Sep 2025).
  • Sample Complexity: Bandit-based algorithms for top-m arm selection provide explicit bounds on the number of "pulls" (LLM calls) required for ϵ\epsilon-optimal identification among exponentially many candidate sets (Purohit et al., 10 Jun 2025).
  • Empirical Results: Principal benchmarks (EMNIST, GTSRB, MMLU-Pro, GSM8K, FinQA, VideoSummarization, etc.) consistently highlight that structure-aware, ordering-aware, and interaction-aware exemplar selection outperforms random, nearest neighbor, or single-instance-based approaches in accuracy, F-score, calibration, coverage, and efficiency.

5. Diversity, Ordering, and Interaction Effects

Maximizing the utility of exemplars requires careful control of redundancy, decay in marginal information, and prompt composition effects:

  • Diversity: Structure-induced diversity (via clustering, DPP, or optimal design regularization) reduces redundancy and increases coverage, especially critical in high-dimensional or multi-modal distributions (Singh et al., 19 Sep 2025, Purohit et al., 6 Nov 2024, Santosh et al., 23 Jan 2025).
  • Ordering Effects: For LLMs, prompt example order materially alters output distributions; methods that optimize over ordered permutations (rather than unordered sets) consistently realize large gains, particularly for difficult or out-of-distribution tasks (Wu et al., 25 May 2024).
  • Interaction Modeling: Static subset-scoring frameworks (e.g., EXPLORA (Purohit et al., 6 Nov 2024)) directly optimize over sets, taking exemplar interactions into account, instead of assuming independent marginal impacts of each exemplar.

6. Practical and Deployment Considerations

Effective real-world deployment of exemplar selection methodologies must address:

  • Efficiency & Scalability: Modern algorithms (e.g., EASE, CASE, COM-BOM) scale sublinearly in the number of candidate subsets due to bandit or surrogate-guided optimization, supporting tractable search even when nkn^k selection space is intractable.
  • Robustness to Imbalance/Non-IID: Methods that guarantee structural coverage (e.g., FFS (You et al., 2020), spectral clustering (Resani et al., 12 Sep 2024), inter-client generative augmentation (Sun et al., 25 Dec 2024)) mitigate loss of rare or minority information in imbalanced or distributed data.
  • Privacy: Condensed exemplars and generative memory buffers (e.g., in federated learning) support privacy preservation by avoiding retention of raw user data (Sun et al., 25 Dec 2024).
  • Generalization: Multi-level similarity integration (across surface, syntactic, semantic layers) and task/output-aware selection mechanisms provide strong generalization across tasks (Liu et al., 14 Jun 2024).

7. Outlook and Open Challenges

Recent evidence establishes exemplar selection as a multi-faceted optimization challenge, with ordering, diversity, and coverage all dictating downstream performance in high-stakes, resource-constrained, and rapidly adaptive environments. Open challenges include:

  • Scalable multi-objective optimization for large-scale settings, balancing accuracy, calibration, and fairness (Luo et al., 1 Oct 2025).
  • Automatic detection and adaptation to task/instance structure, e.g., structural alignment in semantic parsing (Li et al., 28 Aug 2025).
  • Integration with generative models for data-efficient exemplar synthesis and privacy.
  • Formal analysis of interaction effects among exemplars in dynamic prompt construction, bridging theoretical and empirical approaches.

Exemplar selection remains an active field, with ongoing innovation at the intersection of optimization, information theory, and the statistical structure of modern datasets.

Forward Email Streamline Icon: https://streamlinehq.com

Follow Topic

Get notified by email when new papers are published related to Exemplar Selection and Use.