Papers
Topics
Authors
Recent
2000 character limit reached

Conformal-Calibrated Interval (CCI)

Updated 30 November 2025
  • Conformal-Calibrated Interval (CCI) is a framework that constructs set-valued predictors ensuring valid uncertainty quantification using conformal prediction techniques.
  • It employs a calibration procedure with an uncertainty-scaling function and conformal quantiles to achieve finite-sample marginal coverage even in non-i.i.d. or misspecified settings.
  • CCI integrates an abstention mechanism based on interval endpoints to support uncertainty-aware decision-making in high-stakes applications like clinical diagnostics.

A Conformal-Calibrated Interval (CCI) is a statistically principled interval or set-valued predictor constructed to provide valid uncertainty quantification for predictions from machine learning models. CCIs generalize classical conformal prediction to accommodate flexible underlying predictors, explicit uncertainty-scaling, and group-conditional or composite targets, with finite-sample marginal (or group/multitarget) coverage guarantees even for complex, misspecified, or non-i.i.d. data regimes. The CCI framework is central in modern uncertainty-aware decision support, calibration of probabilistic classifiers, and risk-controlling sets in high-stakes applications, as exemplified by its deployment in clinical smart home systems for early detection of urinary tract infection flare-ups (Ugwu et al., 23 Nov 2025).

1. Formal Definition and Construction

Let X\mathcal X be a feature space and Y{0,1}Y\in\{0,1\} a binary outcome. Suppose a pre-trained probabilistic classifier f:X[0,1]f:\mathcal X\to[0,1] outputs p=f(x)=P^(Y=1x)p=f(x)=\widehat P(Y=1\mid x). To improve interpretability and enable coverage calibration, define a transformed label: Y=0.25+0.5Y{0.25,0.75}Y' = 0.25 + 0.5 Y \in \{0.25, 0.75\} On a calibration set {(xi,yi)}i=1n\{(x_i, y_i)\}_{i=1}^n, compute: pi=f(xi),yi=0.25+0.5yip_i = f(x_i), \quad y_i' = 0.25 + 0.5 y_i with per-instance nonconformity score

S(pi,yi)=(yipi)2σ(pi)S(p_i, y_i') = \frac{(y_i' - p_i)^2}{\sigma(p_i)}

where σ(p)\sigma(p) is a positive-valued, user-chosen or theoretically justified uncertainty-scaling function. The conformal quantile is

q^=min{q:1ni=1n1{S(pi,yi)q}1α}\hat q = \min\left\{q : \frac{1}{n}\sum_{i=1}^n \mathbf{1}\{S(p_i, y_i') \leq q\} \geq 1-\alpha \right\}

For a new input xx, compute p=f(x)p=f(x) and use the class-anchor

ytest={0.25p<0.5 0.75p0.5y'_{\rm test} = \begin{cases} 0.25 & p < 0.5 \ 0.75 & p \geq 0.5 \end{cases}

The Conformal-Calibrated Interval is

C(x)={p[0,1]|S(p,ytest)q^}C(x) = \left\{\, p'\in[0,1] \,\middle|\, S(p', y'_{\rm test}) \leq \hat q \right\}

which, due to the construction, forms a closed interval

pytestq^σ(p)|p' - y'_{\rm test}| \leq \sqrt{\hat q \cdot \sigma(p')}

and numerically, the endpoints are

L(x)=inf{ppytestq^σ(p)},U(x)=sup{ppytestq^σ(p)}L(x) = \inf\{p' \mid |p' - y'_{\rm test}| \leq \sqrt{\hat q\,\sigma(p')}\}, \quad U(x) = \sup\{p' \mid |p' - y'_{\rm test}| \leq \sqrt{\hat q\,\sigma(p')}\}

The reported CCI for a test xx is [L(x),U(x)][L(x), U(x)] (Ugwu et al., 23 Nov 2025).

2. Theoretical Guarantees and Statistical Rationale

Under exchangeability (the joint law of calibration plus test is invariant to order), the quantile q^\hat q ensures that for the calibration and new test example

Pr{S(f(Xtest),Ytest)q^}1α\Pr\left\{ S\big(f(X_{\mathrm{test}}), Y'_{\mathrm{test}}\big) \leq \hat q \right\} \geq 1-\alpha

This yields a marginal coverage guarantee: Pr{YtestC(Xtest)}1α\Pr\left\{ Y'_{\mathrm{test}} \in C(X_{\mathrm{test}}) \right\} \geq 1-\alpha As Y{0.25,0.75}Y'\in\{0.25,0.75\}, this means the true label-anchor for the test case lies in the CCI with the prescribed probability. The scaling function σ(p)\sigma(p) modulates the uncertainty inflation (the default linear form, $1 + (1 - |p - 0.5|)$, penalizes near-decision-boundary calibration), but the coverage guarantee holds for any positive σ\sigma (Ugwu et al., 23 Nov 2025).

3. Calibration Procedure and Algorithm

The CCI construction is a special instance of full conformal prediction and can be implemented, as in (Ugwu et al., 23 Nov 2025), by the following pipeline:

1
2
3
4
5
6
7
8
9
10
11
12
13
Input: D_train, D_cal, classifier f, miscoverage α
1. Train f on D_train.
2. For each (x_i, y_i) in D_cal:
   p_i  f(x_i)
   y_i' ← 0.25 + 0.5 · y_i
   σ_i  σ(p_i)
   s_i  (y_i' - p_i)^2 / σ_i
3. q̂  ((n+1)(1-α))-th smallest of {s_i}
4. For a new test x:
   p̂  f(x)
   y' ← 0.25 + 0.5 · 1[p̂ ≥ 0.5]
   Solve |p - y'| ≤ sqrt(q̂ · σ(p)) over p ∈ [0,1]
   Output interval [L, U]
There is no requirement for regularization beyond the choice of σ(p)\sigma(p). Additional smoothing or randomized quantile tie-breaking can be used if desired (Ugwu et al., 23 Nov 2025).

4. Abstention and Decision Rule

Beyond outputting CIs, the CCI enables an abstention mechanism. Compute:

  • If L(x)0.5L(x) \geq 0.5: the interval lies entirely above $0.5$, so predict "UTI".
  • If U(x)<0.5U(x) < 0.5: the interval lies entirely below $0.5$, so predict "No UTI".
  • Otherwise: abstain (i.e., "I don't know").

This abstention capability provides an uncertainty-aware decision threshold, which is critical in clinical and safety-critical scenarios (Ugwu et al., 23 Nov 2025).

5. Trade-offs, Hyperparameters, and Extensions

  • Significance level α\alpha: Lower α\alpha yields wider intervals and higher coverage.
  • Uncertainty-scaling σ(p)\sigma(p): Default is linear, but any positive function is permitted.
  • Label-anchor mapping: Fixed at {0.25,0.75}\{0.25, 0.75\} but can be adapted if needed.
  • Base classifier ff: CCI is agnostic to ff: logistic regression, neural nets, and other probabilistic classifiers can all be wrapped, yielding varying interval sharpness but identical coverage.
  • Width control: Interval width is explicitly reportable and should be averaged empirically across test points for performance assessment (Ugwu et al., 23 Nov 2025).

6. Empirical Evaluation and Practical Impact

In real-world clinical deployment, the CCI-calibrated system achieved:

  • Outperformance of baselines in recall and other classification metrics.
  • The lowest observed abstention proportion and interval width among tested methods.
  • Validation from a survey of 42 nurses, who found CCI outputs valuable for guiding clinical decision-making in early UTI detection.

This result demonstrates the critical practical utility of CCIs in operational decision support systems where both accurate prediction and explicit, actionable uncertainty quantification are essential (Ugwu et al., 23 Nov 2025).

7. Broader Context and Relationship to Conformal Approaches

The CCI as instantiated above is a conformal extension that applies to general probabilistic classifiers for classification with abstention. It is rooted in conformal prediction theory, which provides finite-sample, model-agnostic marginal and conditional coverage guarantees via exchangeability-based score quantile calibration. The use of a tunable uncertainty-scaling function, anchor mapping, and abstention logic distinguish the CCI framework from classical conformal prediction, enabling reliable, sharp, and uncertainty-aware predictions in demanding applications (Ugwu et al., 23 Nov 2025).

Definition Search Book Streamline Icon: https://streamlinehq.com
References (1)
Slide Deck Streamline Icon: https://streamlinehq.com

Whiteboard

Forward Email Streamline Icon: https://streamlinehq.com

Follow Topic

Get notified by email when new papers are published related to Conformal-Calibrated Interval (CCI).