Papers
Topics
Authors
Recent
Search
2000 character limit reached

Confidence-Evidence Ratio (Ψ)

Updated 29 December 2025
  • Confidence-Evidence Ratio (Ψ) is a metric that quantifies the relationship between subjective confidence and additive evidence in probabilistic inference frameworks.
  • It is derived from Bayesian and axiomatic learning principles, converting fractional confidence into additive evidence using log-based transformations with context-dependent parameters.
  • The ratio is pivotal in assessing dataset consistency, informing cosmological tension tests, and unifying approaches from Kalman filtering to Dempster–Shafer theory.

The Confidence-Evidence Ratio, denoted Ψ\Psi (and alias RR in certain contexts), formalizes the relationship between a quantitative measure of confidence and an associated amount of evidence in probabilistic inference and belief-updating frameworks. It arises from both Bayesian statistics and axiomatic theories of learning with confidence, where it links the multiplicative structure of belief updates to an additive quantification of evidence or trust. Ψ\Psi serves both as a diagnostic in model comparison—especially for data-set consistency/tension in cosmology—and as a foundational object in abstract learning frameworks, mediating the flow from subjective confidence measures to structural update dynamics.

1. Definitions and Mathematical Formulation

In Bayesian inference, the Confidence-Evidence Ratio Ψ\Psi offers a probabilistically meaningful measure of how much "extra" confidence one should place in the joint compatibility of two datasets. Given Bayesian evidences ZA\mathcal{Z}_A, ZB\mathcal{Z}_B for datasets AA and BB, and their joint evidence ZAB\mathcal{Z}_{AB} under a shared prior π(θ)\pi(\theta), the ratio is defined as: RR0 This admits several equivalent formulations, notably in terms of posterior distributions RR1: RR2 In axiomatically-motivated frameworks, such as learning with confidence, RR3 emerges as a conversion parameter between fractional confidence RR4 and additive evidence RR5: RR6 The parameter RR7 is the Confidence-Evidence Ratio and is context-dependent, typically taking the value 1 in standard Bayesian updates and Kalman filtering settings (Richardson, 14 Aug 2025).

2. Axiomatic Foundations and Canonical Domains

A rigorous foundation for RR8 arises from a set of axioms for confidence-based updates, central to the work "Learning with Confidence" (Richardson, 14 Aug 2025). The relevant space is characterized as follows:

  • RR9: space of belief states (priors)
  • Ψ\Psi0: space of possible observations
  • Ψ\Psi1: domain of confidence values, supporting binary associative operation Ψ\Psi2 (composition), partial order, and identity elements Ψ\Psi3 (no confidence) and Ψ\Psi4 (full confidence).

The axioms CF0–CF4 (zero confidence, idempotence, smoothness, associativity, and non-cyclicity) guarantee that any such update admits a reparameterization into an additive flow: Ψ\Psi5 For the canonical continuous domains—(a) fractional Ψ\Psi6 with Ψ\Psi7-combination, and (b) additive Ψ\Psi8 with ordinary addition—an isomorphism exists via Ψ\Psi9, with Ψ\Psi0 as a scale parameter: Ψ\Psi1 Ψ\Psi2 acts as the conversion factor between fractional trust and linearized (additive) evidence.

3. Prior Dependence and the Role of KL Divergences

A salient property of the Bayesian version of Ψ\Psi3 is its explicit dependence on the prior volume Ψ\Psi4. As shown in "Quantifying tensions in cosmological parameters" (Handley et al., 2019), for uniform (top-hat) priors,

Ψ\Psi5

and for Gaussian posteriors within a box prior,

Ψ\Psi6

Thus, shrinking the prior volume always drives down Ψ\Psi7 (and Ψ\Psi8), which can artificially mask incompatibility between datasets. The recommendation is that if any physically reasonable prior width yields Ψ\Psi9, the datasets should be deemed to be in tension.

To correct for this, the prior-independent modification uses Kullback–Leibler divergences ZA\mathcal{Z}_A0: ZA\mathcal{Z}_A1 The suspiciousness statistic ZA\mathcal{Z}_A2 is then defined as: ZA\mathcal{Z}_A3 ZA\mathcal{Z}_A4 isolates genuine model tension, free of prior volume artifacts, and supports calibrated statistical interpretation.

4. Interpretations in Learning and Inference

Across learning theory and Bayesian updating, ZA\mathcal{Z}_A5 governs how different types of confidence feedback map onto additive evidence accumulation. In belief flows induced by learning rules,

  • The vector-field representation interprets the path ZA\mathcal{Z}_A6 as an integral curve.
  • In a loss-based (gradient) model, the relationship between infinitesimal update rates and total evidence follows directly from ZA\mathcal{Z}_A7: ZA\mathcal{Z}_A8 In the Bayesian case, with ZA\mathcal{Z}_A9, the soft-update via likelihood ratio ZB\mathcal{Z}_B0 raised to the ZB\mathcal{Z}_B1-th power recovers ordinary Bayes rule for ZB\mathcal{Z}_B2. For Kalman filtering, the confidence-evidence mapping reproduces the Kalman gain structure, with

ZB\mathcal{Z}_B3

A similar structure incorporates Shafer's weight of evidence and Dempster–Shafer theory, reinforcing the centrality of ZB\mathcal{Z}_B4 (Richardson, 14 Aug 2025).

5. Practical Computation and Statistical Usage

The computation of ZB\mathcal{Z}_B5 in data-set cross-consistency testing is rigorous and systematic (Handley et al., 2019). The recommended workflow is:

  1. Compute the raw ratio ZB\mathcal{Z}_B6 using the joint and marginal evidences for the datasets under consistent priors.
  2. Explore the dependence of ZB\mathcal{Z}_B7 on prior width; search for any physically reasonable priors with ZB\mathcal{Z}_B8 as a necessary test for tension.
  3. Calculate ZB\mathcal{Z}_B9 for a prior-invariant assessment, and report the tension probability AA0 based on a calibrated AA1-distribution:

AA2

where AA3 is the effective dimension (from Bayesian model dimensionality).

The table below summarizes the qualitative interpretation:

AA4/AA5 AA6 value Tension classification
AA7, AA8 AA9 No tension
Some BB0, BB1 BB2 Moderate tension
All BB3, BB4 BB5 Strong tension

6. Special Cases and Connections

  • Cosmological parameter tension: BB6 and variants are routinely applied to test compatibility between major data sets (e.g., Planck, DES, BOSS, SHOES). Empirical results indicate graduated degrees of tension depending on the datasets and priors, supporting fine-grained model assessment (Handley et al., 2019).
  • Kalman filter and Dempster–Shafer weight: In dynamic and epistemic uncertainty models, BB7 recovers classical weights of evidence and trust scaling, unifying several learning paradigms under a shared conversion law (Richardson, 14 Aug 2025).

BB8 provides a mathematically principled and operationally effective interface between subjective trust, additive evidence, and model/data consistency. Correct usage requires careful attention to the prior dependence (in Bayesian settings) and the correct interpretation of BB9 in abstract learning scenarios. Reporting both the raw Bayesian confidence ratio ZAB\mathcal{Z}_{AB}0 and its prior-independent counterpart ZAB\mathcal{Z}_{AB}1 (with effective dimension ZAB\mathcal{Z}_{AB}2) is advocated for transparency and reproducibility, especially in inference problems involving high-stakes scientific consistency tests. This dual reporting framework is now a proposed standard in cosmological data analysis (Handley et al., 2019), and the axiomatic translation underscores its universality in learning and evidence aggregation (Richardson, 14 Aug 2025).

Definition Search Book Streamline Icon: https://streamlinehq.com
References (2)

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Confidence-Evidence Ratio ($\Psi$).