Papers
Topics
Authors
Recent
2000 character limit reached

SmallML: Efficient Conformal Prediction

Updated 19 November 2025
  • SmallML is a framework for constructing prediction sets that guarantee exact finite-sample marginal coverage while minimizing set size.
  • It utilizes innovative methods like SAPS and adaptive thresholding to optimize nonconformity scores and improve model calibration.
  • Empirical results on benchmarks like ImageNet demonstrate significant reductions in average set sizes, enhancing efficiency over classical approaches.

SmallML (Small-Set Marginal Learning) refers to the suite of methods, theory, and algorithmic design principles surrounding minimization of conformal prediction set size (often called "efficiency") subject to exact finite-sample marginal coverage constraints. The central object of paper is the prediction set C(x)C(x): given a pretrained classifier or regressor, and under a minimal exchangeability or i.i.d. assumption on calibration and test data, construct C(x)C(x) such that %%%%2%%%% for a user-specified miscoverage rate α\alpha, while maintaining C(x)|C(x)| as small as possible on average. Recent advances, particularly in deep learning and large-scale benchmarks, have focused on refined score constructions, optimization of nonconformity functions, adaptive thresholding, and incorporation of model calibration and probabilistic structure—all towards achieving set sizes orders of magnitude smaller than those produced by prior methods.

1. Principles of Marginal Validity and Set Efficiency

The conformal prediction framework requires only the exchangeability of calibration and test data, making no assumptions about the accuracy or calibration of the underlying model. For a pre-trained model π^:XΔK1\hat{\pi}: \mathcal X \to \Delta^{K-1} (class probabilities), a held-out calibration set {(xi,yi)}i=1n\{(x_i, y_i)\}_{i=1}^n, and a user-chosen miscoverage level α\alpha, the classical workflow computes non-conformity scores S(xi,yi)S(x_i, y_i), sets a threshold τ\tau as the empirical (1α)(1-\alpha)-quantile, and defines C(xn+1)={y:S(xn+1,y)τ}C(x_{n+1}) = \{ y : S(x_{n+1}, y) \le \tau \} (Huang et al., 2023). By construction,

Pr[YC(X)]1α.\Pr[Y \in C(X)] \ge 1-\alpha.

All subsequent developments—whether relying on softmax scores, instance difficulty, higher-order uncertainty, or distributional estimates—are constrained by this foundational guarantee. The challenge addressed by SmallML is to design SS and calibrate τ\tau such that C(x)C(x) is as small as possible for typical xx.

2. Sorted Adaptive Prediction Sets (SAPS) and Label Ranking

The Sorted Adaptive Prediction Sets (SAPS) framework illustrates a key innovation in SmallML: minimize dependence on miscalibrated probability tails by discarding all but the maximum predicted probability. For KK-class deep classifiers, SAPS defines for each (x,y)(x, y) and random uUniform[0,1]u \sim \mathrm{Uniform}[0,1]: SSAPS(x,y,u;π^)={uπ^max(x)o(y,π^(x))=1 π^max(x)+(o(y,π^(x))2+u)λo(y,π^(x))2S_{\mathrm{SAPS}}(x, y, u; \hat{\pi}) = \begin{cases} u \cdot \hat{\pi}_{\max}(x) & o(y, \hat{\pi}(x)) = 1 \ \hat{\pi}_{\max}(x) + (o(y, \hat{\pi}(x)) - 2 + u)\lambda & o(y, \hat{\pi}(x)) \geq 2 \end{cases} where o(y,π^(x))o(y, \hat{\pi}(x)) is the rank of label yy in descending sorted predictions, and λ>0\lambda>0 is a user-tuned hyperparameter (Huang et al., 2023). Only the maximum softmax value is retained; all other probabilistic information is replaced by λ\lambda. The conformal set is determined by thresholding SSAPSS_{\mathrm{SAPS}} at the calibrated τ\tau. This design yields:

  • Average set sizes 1/7\sim1/7 those of classical APS on ImageNet (APS: 20.95, SAPS: 2.98 at α=0.1\alpha=0.1).
  • Consistent marginal coverage at the prescribed 1α1-\alpha level.
  • Smaller, instance-adaptive sets—"easy" examples with high confidence receive smaller C(x)C(x), while "hard" instances expand accordingly.
  • Uniformly improved conditional coverage (ESCV) versus RAPS and APS.

Experiments on ImageNet, CIFAR-100, and CIFAR-10 consistently find SAPS to outperform alternatives in size and conditional error, with stable performance over reasonable calibration set sizes and hyperparameter choices.

3. Theoretical Guarantees and Coverage Properties

SmallML methods such as SAPS preserve exact marginal validity by standard conformal theory: if the set construction is "nested" (set size increases monotonically as the acceptance threshold grows laxer) and the threshold τ\tau is chosen as the empirical (1α)(1-\alpha)-quantile of S(xi,yi)S(x_i, y_i), then for all distributions: [ \Pr[Y \in C_{1-\alpha}(X)] \geq

Definition Search Book Streamline Icon: https://streamlinehq.com
References (1)

Whiteboard

Follow Topic

Get notified by email when new papers are published related to SmallML.