Papers
Topics
Authors
Recent
2000 character limit reached

Coefficient of Prescriptiveness

Updated 24 November 2025
  • Coefficient of prescriptiveness is defined as the normalized improvement in cost reduction achieved by a contextual policy over a static reference relative to an optimal clairvoyant benchmark.
  • It decomposes the expected cost into the minimal achievable cost and the regret, linking to classical metrics like the R² metric in predictive models.
  • Its robust formulation via convex reformulations and LP-oracle methods enables effective optimization under distribution shifts while ensuring policy stability.

The coefficient of prescriptiveness, also known as the prescriptiveness competitive ratio (PCR), is a universal, unitless performance measure that quantifies the value of contextual (data-driven) decision-making relative to a static reference and an anticipative (clairvoyant) benchmark. Introduced to evaluate the practical prescriptive power of side information in stochastic optimization, the PCR expresses the proportion of the maximum attainable reduction in expected cost (regret gap) achieved by a given contextual policy. This framework has theoretical guarantees on admissible values, direct ties to classical statistical metrics, and supports robust optimization formulations that address model uncertainty and data distribution shifts (Poursoltani et al., 2023).

1. Formal Definition and Properties

Let h(x,ξ)h(x, \xi) be a cost function where xXx \in X is a feasible decision and ξ\xi is a random vector with distribution FF. Decisions are selected according to a policy x(ζ):RnζXx(\zeta): \mathbb{R}^{n_\zeta} \to X, with ζRnζ\zeta \in \mathbb{R}^{n_\zeta} representing observed side information. The clairvoyant (fully anticipative) solution is

xargminxXEF[h(x,ξ)]x^* \in \arg\min_{x' \in X} \mathbb{E}_F[h(x', \xi)]

and a static reference decision xˉX\bar{x} \in X (e.g., the solution ignoring ζ\zeta).

The prescriptiveness competitive ratio of policy x()x(\cdot) relative to xˉ\bar{x} is

PCRF(x(),xˉ)=1EF[h(x(ζ),ξ)]EF[h(x,ξ)]EF[h(xˉ,ξ)]EF[h(x,ξ)]\mathrm{PCR}_F(x(\cdot),\, \bar{x}) = 1 - \frac{\mathbb{E}_F[h(x(\zeta), \xi)] - \mathbb{E}_F[h(x^*, \xi)]}{\mathbb{E}_F[h(\bar{x}, \xi)] - \mathbb{E}_F[h(x^*, \xi)]}

By construction, PCR is unitless and provides the fraction of the improvement available over xˉ\bar{x} (compared to clairvoyance) delivered by x()x(\cdot). When xˉX\bar{x} \in X, it holds (Lemma 1) that PCRF(x(),xˉ)[0,1]\mathrm{PCR}_F(x(\cdot),\bar{x}) \in [0,1] for any x()x(\cdot): a value of 1 corresponds to optimality (clairvoyant policy), and 0 to parity with the static rule.

2. Interpretation and Special Cases

The expected cost decomposition for a contextual policy is

EF[h(x(ζ),ξ)]=EF[h(x,ξ)]+r(x()),\mathbb{E}_F[h(x(\zeta),\xi)] = \mathbb{E}_F[h(x^*,\xi)] + r(x(\cdot)),

where r(x())r(x(\cdot)) denotes the “regret” of x()x(\cdot) versus xx^*. The PCR compares the regret of the contextual rule to the regret incurred by xˉ\bar{x}, acting as a normalized regret measure. PCR is monotonically decreasing in the regret of x()x(\cdot) and increasing in the regret of xˉ\bar{x}.

When the expected cost is replaced by empirical averages (empirical distribution F^\hat{F}), the PCR reduces to the standard definition used in practice. In the case h(x,ξ)=(xξ)2h(x,\xi) = (x-\xi)^2, x,ζx,\zeta one-dimensional, the PCR coincides with the R2R^2 metric for predictive models.

3. Distributionally Robust Prescriptiveness Optimization

In data-driven settings, the true FF may not be known and could shift from an empirical estimate F^\hat{F}. Consequently, policies seek to maximize the worst-case PCR over an ambiguity set D\mathcal{D} of distributions:

PCRD(x(),xˉ)=infFDPCRF(x(),xˉ)\mathrm{PCR}_{\mathcal{D}}(x(\cdot), \bar{x}) = \inf_{F \in \mathcal{D}} \mathrm{PCR}_F(x(\cdot), \bar{x})

The distributionally robust prescriptiveness optimization problem (DR–PCR) seeks

maxx()HPCRD(x(),xˉ),\max_{x(\cdot) \in \mathcal{H}} \mathrm{PCR}_{\mathcal{D}}(x(\cdot), \bar{x}),

for some class H\mathcal{H} of policies. Lemma 1 shows the optimal DR–PCR objective is in [0,1][0,1] if xˉH\bar{x} \in \mathcal{H}. Lemma 2 states that when D={F}\mathcal{D} = \{F\} and H\mathcal{H} is unrestricted, the contextual stochastic optimization (CSO) solution maximizes worst-case PCR.

4. Tractable Reformulation and Algorithmic Solution

By introducing a variable γ\gamma (interpreted as the worst-case PCR), Proposition 1 yields an equivalent convex reformulation:

maxx(),γ    γ subject to supFDEF[h(x(ζ),ξ)[(1γ)h(xˉ,ξ)+γh(x,ξ)]]0, 0γ1.\begin{align*} &\max_{x(\cdot),\, \gamma} \;\; \gamma \ &\text{subject to } \sup_{F\in \mathcal{D}}\, \mathbb{E}_F\Bigl[h(x(\zeta),\xi) - [(1-\gamma)h(\bar{x},\xi)+\gamma h(x^*,\xi)]\Bigr] \le 0, \ &0 \le \gamma \le 1. \end{align*}

For convex D\mathcal{D} (e.g., polyhedral or CVaR-type ambiguity sets), this is a convex program.

For a popular choice of D\mathcal{D} (the “nested CVaR” set), the robust objective decomposes for each ζ\zeta-scenario ω\omega into LPs parameterized by γ\gamma, and overall feasibility is certified by a convex, non-decreasing function ψ(γ)\psi(\gamma):

  • For candidate γ\gamma, per-scenario LPs ϕω(γ)\phi_\omega(\gamma) are solved.
  • Bisection is used to find the maximal γ[0,1]\gamma^* \in [0,1] such that ψ(γ)0\psi(\gamma^*) \leq 0.
  • Algorithmic complexity is O(log(1/ϵ))O(\log(1/\epsilon)) bisection iterations, each requiring Ωζ|\Omega_\zeta| LP solves; the procedure is polynomial-time for polyhedral XX and linear hh.

5. Illustrative Application: Contextual Shortest Path

As a case study, the contextual shortest path problem involves a network G=(V,A)G = (V, A) with V=45|V| = 45 nodes and A=97|A| = 97 arcs; arc costs ξR97\xi \in \mathbb{R}^{97} are uncertain. Decisions x{0,1}97x \in \{0,1\}^{97} route flow from origin oo to destination dd to minimize xξx^\top \xi. Side information ζR200\zeta \in \mathbb{R}^{200} is correlated with ξ\xi via a multivariate normal. A random-forest model trained on Ntrain=400N_{\text{train}}=400 samples generates empirical conditional distributions F^ξζ\hat{F}_{\xi|\zeta}, yielding tree-leaf–weighted scenarios for any new ζ\zeta.

Four methods are compared, with parameter α\alpha validated against Nval=400N_{\mathrm{val}} = 400:

  1. CSO (classical contextual SAA, α=0\alpha=0)
  2. DRCSO (nested-CVaR robustification of expected-cost)
  3. DRCRO (nested-CVaR robustification of “ex-post regret”)
  4. DR–PCR (distributionally robust maximization of worst-case PCR, α[0,1]\alpha \in [0,1])

A distribution shift test constructs an out-of-sample test set (Ntest=1000N_{\text{test}}=1000) with increasing perturbation in the mean of ξ\xi (m{0%,20%,30%,40%,50%,60%}m \in \{0\%,20\%,30\%,40\%,50\%,60\%\}). The metric evaluated is out-of-sample PCRF(x(),xˉ)\mathrm{PCR}_F(x^*(\cdot),\bar x). The results (cf. Figure 1) show that for m=0m=0, all methods yield similar PCR 0.45\approx 0.45, but for larger mm, PCR for CSO/DRCSO/DRCRO quickly falls below zero while DR–PCR maintains PCR0\mathrm{PCR} \geq 0 up to m=60%m=60\%, indicating notable robustness to distribution shift.

Method PCR when m=0m=0 PCR as mm \uparrow60%
CSO 0.45\approx 0.45 Negative
DRCSO 0.45\approx 0.45 Negative
DRCRO 0.45\approx 0.45 Negative
DR–PCR 0.45\approx 0.45 0\geq 0

6. Theoretical and Practical Insights

The worst-case PCR objective serves as a regularizer, biasing contextual policies toward the static reference xˉ\bar{x} in cases of unreliable side information ζ\zeta, and penalizing over-reaction in scenarios of distribution shift. The trade-off between conservatism and robustness is controlled by the choice of ambiguity set D\mathcal{D} (e.g., nested CVaR, Wasserstein, moment-based). Calibration of α\alpha should be performed on a hold-out set mimicking anticipated distribution shifts.

The bisection plus LP-oracle framework generalizes to any setting where the subproblem ϕω(γ)\phi_\omega(\gamma) is a tractable convex program. Restricting H\mathcal{H}, such as to affine policies, reduces the dimensionality of the underlying optimization. The PCR is a standard-comparable, interpretable metric on [0,1][0,1]; a realized value γ=p\gamma^* = p certifies that at least fraction pp of the maximal gain over the static policy is achieved even under the worst-case distributional stress within D\mathcal{D}.

7. Significance and Interpretations

The coefficient of prescriptiveness provides a rigorous, application-invariant measure for quantifying the utility of contextualization in decision-making under uncertainty. Its robust formulation directly addresses the challenges of estimation error and distributional shift, which are central to the deployment of data-driven optimization policies. The ability to guarantee a quantifiable proportion of clairvoyance-improvable cost under worst-case conditions supports both theoretical analysis and practical decision support (Poursoltani et al., 2023).

Definition Search Book Streamline Icon: https://streamlinehq.com
References (1)

Whiteboard

Topic to Video (Beta)

Follow Topic

Get notified by email when new papers are published related to Coefficient of Prescriptiveness.