Blend-ASC: Adaptive Sampling for LLM Self-Consistency
- Blend-ASC is a dynamic-sample-allocation algorithm that optimizes test-time self-consistency inference by adaptively assigning samples per question.
- It employs a blended ranking mechanism combining Beta-posterior estimates from adaptive SC and PPR-1v1 to effectively distribute a fixed prompt budget.
- Empirical evaluations on benchmarks like GSM8K and MATH show Blend-ASC achieves up to 6.9× sample efficiency compared to traditional methods.
Blend-ASC is a dynamic-sample-allocation algorithm for test-time self-consistency (SC) inference with LLMs, introduced to address the sample efficiency and budget allocation limitations inherent to classical self-consistency methods. SC itself refers to the inference protocol in which multiple chain-of-thought (CoT) responses are sampled for each question, with the most frequent answer taken as the output. Blend-ASC enables practitioners to fit the inference to any arbitrary sample budget by adaptively allocating samples across questions, thereby minimizing overall error within the specified computational constraints. Significantly, Blend-ASC is hyperparameter-free and empirically achieves state-of-the-art sample efficiency, requiring an average of 6.8× fewer samples than vanilla SC for matched accuracy on major LLM reasoning benchmarks such as GSM8K and MATH (Feng et al., 15 Nov 2025).
1. Motivation and Problem Formulation
Traditional self-consistency operates by a fixed allocation: each question receives independent CoT samples from the LLM, with the modal answer returned. The empirical mode (where denotes empirical frequency) serves as the prediction. This fixed-per-question budget is known to be wasteful: "easy" questions—those with a large margin between the top two answer probabilities—attain near-zero error with few samples, while "hard" small-margin questions dominate aggregate sampling requirements. Realistic deployments, however, impose a strict total inference sample budget (i.e., total number of LLM prompts), which classical SC cannot flexibly distribute. Blend-ASC directly addresses the objective of minimizing aggregate dataset-level error under a constraint , adaptively allocating samples per question (Feng et al., 15 Nov 2025).
2. Theoretical Foundations and Scaling Laws
The self-consistency protocol is mathematically framed as empirical mode estimation. For question whose answer distribution is mode-aligned (i.e., true answer is the mode), the error rate after samples satisfies:
where and are the top two probabilities, and is the margin. This exponential decay of error with respect to samples is empirically verified across model families and datasets: the margin accurately predicts the sample complexity for confidence in majority-vote (Feng et al., 15 Nov 2025). When evaluating over a dataset with varying , the aggregate error follows a power-law scaling:
for common distributions of margin values. Fixed-allocation SC thus exhibits sublinear convergence at the dataset level. The analysis further demonstrates that adaptive strategies, if granted access to (or estimation of) per-question margin, can surpass the scaling law and attain linear ($1/x$) convergence rates, a theoretical optimum for sample use (Feng et al., 15 Nov 2025).
3. Algorithmic Structure of Blend-ASC
Blend-ASC integrates two strands of confidence estimation—adaptive SC (ASC) and sequential mode estimation (PPR-1v1)—to arrive at a sample allocation strategy that blends their strengths across the low- and high-sample regimes, while operating under exact adherence to user-specified prompt budgets. At each incremental sample allocation , Blend-ASC performs:
- For each question , maintain sample counts and counts for the two most frequent answers.
- Compute two statistics:
- (adaptive SC Beta-posterior probability that mode is incorrect)
- (PPR-1v1 mode confidence; is number of classes seen)
- For each , compute blended priority , where each is the rank of across .
- Select question with minimal priority , draw an additional sample for .
- Enforce a hard upper limit: if any exceeds $16$ times the average per-question samples, halt allocation for that .
No tuning parameters are required: Blend-ASC is fully hyperparameter-free. Sampling proceeds until the global budget is exactly exhausted, at which point each over is determined by the majority vote of its drawn samples (Feng et al., 15 Nov 2025).
4. Key Formulas and Confidence Estimators
Blend-ASC relies on principled probabilistic estimates:
- Beta-posterior probability of incorrect mode:
which estimates the confidence that the empirical mode is spurious, used for prioritizing "hard" questions early.
- PPR-1v1 stopping proxy:
This martingale confidence bound ensures sample-optimality in the high-sample asymptote, acting as a virtual p-value for stopping.
A transition/blending mechanism interpolates between these as sampling proceeds, ensuring robust allocation both when little information is known (short runs) and in the regime where enough data secure high-confidence stopping.
5. Sample Efficiency and Empirical Performance
Extensive experiments across reasoning benchmarks (GSM8K, MATH, GPQA-Diamond) and LLMs (LLaMA-3B, Qwen-Math) show that Blend-ASC consistently achieves lower error at the same prompt count relative to SC and all existing adaptive/fixed-allocation baselines, irrespective of temperature (0.6–1.0). Average prompt savings to match fixed- SC are as follows (Feng et al., 15 Nov 2025):
| Setting | Fixed-Alloc | Adaptive SC | Blend-ASC |
|---|---|---|---|
| SC@64 | 4.6× | 5.9× | 6.8× |
| SC@128 | 4.6× | 5.0× | 6.9× |
On six model-dataset combinations, Blend-ASC error curves are below all comparators, with the blended allocation closely matching near-theoretical optimality for a broad spectrum of task hardness (Feng et al., 15 Nov 2025). The protocol operates without requiring threshold selection, window size tuning, or temperature reconfiguration—parameters often problematic in production.
6. Implementation, Integration, and Limitations
Blend-ASC is designed for straightforward integration: it requires only tracking answer counts per question and updating Beta-based statistics; sample allocation can be efficiently managed via a heap or priority queue. The practitioner specifies the overall sample budget ; no per-question or global hyperparameters are set. Empirical stopping can utilize blended priorities to terminate early if a predetermined confidence is attained per question.
While the 16×-oversampling cutoff for any given question is a heuristic rather than a theoretically minimal bound, and PPR-1v1 can be pessimistic at very low sample counts, these practical aspects are minor in standard use cases. Blend-ASC is applicable without modification to any SC-compatible domain.
7. Context and Differentiation from Related Work
Blend-ASC generalizes and improves on several prior self-consistency allocation strategies. Classical SC and fixed-alloc schemes fail to exploit inter-question heterogeneity in answer margin. Adaptive SC [Aggarwal et al.] prioritizes uncertain questions, improving low-sample regimes but does not achieve optimal scaling asymptotically. The PPR-1v1 martingale approach achieves theoretical optimality at higher sample counts, but can underperform early, and requires parameterization. Blend-ASC uniquely covers both short- and long-sample domains with a unified, nonparametric mechanism and is distinguished by its provable error scaling and empirical efficiency. All theoretical rates and design choices are validated empirically across multiple LLMs and real-world benchmarks (Feng et al., 15 Nov 2025).
Sponsored by Paperpile, the PDF & BibTeX manager trusted by top AI labs.
Get 30 days free