Papers
Topics
Authors
Recent
2000 character limit reached

A Bayesian Perspective on Evidence for Evolving Dark Energy (2511.10631v1)

Published 13 Nov 2025 in astro-ph.CO and astro-ph.IM

Abstract: The DESI collaboration reports a significant preference for a dynamic dark energy model ($w_0w_a$CDM) over the cosmological constant ($Λ$CDM) when their data are combined with other frontier cosmological probes. We present a direct Bayesian model comparison using nested sampling to compute the Bayesian evidence, revealing a contrasting conclusion: for the key combination of the DESI DR2 BAO and the Planck CMB data, we find the Bayesian evidence modestly favours $Λ$CDM (log-Bayes factor $\ln B = -0.57{\scriptstyle\pm0.26}$), in contrast to the collaboration's 3.1$σ$ frequentist significance in favoring $w_0w_a$CDM. Extending this analysis to also combine with the DES-Y5 supernova catalogue, our Bayesian analysis reaches a significance of $3.07{\scriptstyle\pm0.10}\,σ$ in favour of $w_0w_a$CDM. By performing a comprehensive tension analysis, employing five complementary metrics, we pinpoint the origin: a significant ($\approx 2.95σ$), low-dimensional tension between DESI DR2 and DES-Y5 that is present only within the $Λ$CDM framework. The $w_0w_a$CDM model is preferred precisely because its additional parameters act to resolve this specific dataset conflict. The convergence of our findings with independent geometric analyses suggests that the preference for dynamic dark energy is primarily driven by the resolution of inter-dataset tensions, warranting a cautious interpretation of its statistical significance.

Summary

  • The paper demonstrates that Bayesian model selection via nested sampling reveals modest evidence for dynamical dark energy primarily when DES-Y5 data is included.
  • The authors employ PolyChord with Cobaya and CAMB to compute log-Bayes factors, integrating Occam’s razor to penalize unnecessary model complexity.
  • The study highlights that alleviating inter-dataset tensions, especially between DESI BAO and supernova data, drives the preference for the evolving dark energy model.

A Bayesian Re-Assessment of Evidence for Evolving Dark Energy

Introduction

The reported preference for dynamical dark energy models, specifically w0waw_0w_aCDM, over the cosmological constant Λ\LambdaCDM in recent cosmological data analyses—such as those led by DESI—has motivated intense scrutiny regarding model selection methodologies. This work deploys a direct Bayesian model comparison via nested sampling to critically evaluate the statistical evidence for evolving dark energy. Contrary to frequentist hypothesis tests which often favor w0waw_0w_aCDM at high significance when combining baryon acoustic oscillation (BAO), cosmic microwave background (CMB), and Type Ia supernova data, the Bayesian approach reveals a more nuanced landscape where the evidence for new physics is highly sensitive to dataset combinations and the underlying inter-dataset tensions.

Methodology: Bayesian Evidence and Tension Metrics

The analysis leverages the PolyChord nested sampling algorithm for robust Bayesian computation, interfaced via Cobaya and CAMB. The parameter space for w0waw_0w_aCDM is bounded following DESI priors, with a physical lower limit on neutrino mass aligned with oscillation data. Bayesian model selection is performed through direct computation of the marginal likelihoods Z\mathcal{Z} of competing cosmological models; the log-Bayes factor lnB\ln B quantifies the relative support for w0waw_0w_aCDM versus Λ\LambdaCDM. This method intrinsically implements Occam’s razor, penalizing unnecessary model complexity via prior volume integration.

Significance is conservatively translated from lnB\ln B to frequentist sigma via upper bounds on Bayes factors and established pp-value relations. Statistical tension between datasets is quantified with the unimpeded evidence framework, encompassing metrics such as evidence ratios, suspiciousness (quantifying direct likelihood conflict), and Bayesian dimensionality (characterizing the parameter subspace where conflicts reside).

Results: Model Preference Is Dataset-Dependent

Bayesian model comparison yields results that diverge substantially from DESI's headline frequentist significances. Individually, DESI DR2 BAO favors Λ\LambdaCDM (lnB=1.47±0.11\ln B = -1.47 \pm 0.11). When paired with Planck CMB data—a combination frequently cited as key evidence for evolving dark energy—the Bayesian evidence continues to modestly prefer Λ\LambdaCDM with lnB=0.57±0.26\ln B = -0.57 \pm 0.26, contrasting sharply with the 3.1σ\sigma frequentist preference for w0waw_0w_aCDM. Inclusion of Pantheon+ supernova data further strengthens Bayesian support for Λ\LambdaCDM.

The preference for w0waw_0w_aCDM materializes only when the DES-Y5 supernova catalogue is added. Both pairwise and triplet combinations involving DES-Y5 show positive log-Bayes factors and reach Bayesian significances up to 3.07±0.10σ3.07 \pm 0.10\,\sigma, though still weaker than the frequentist claims. The origin and dimensionality of model preference are illuminated by dedicated tension metrics, revealing that the w0waw_0w_aCDM model is favored due to its ability to mitigate tension between DESI BAO and DES-Y5 data—not because of intrinsic support from BAO or CMB alone. Figure 1

Figure 1

Figure 1: Posterior constraints in w0waw_0w_aCDM parameter space for DESI DR2 alone and in combination with diverse external probes. Visible are significant shifts in w0w_0, waw_a, and broader cosmological constraints, illustrating how dataset choices impact inferred dark energy dynamics.

Tension Analysis and the Mechanism of Model Selection

The inter-dataset conflict between DESI DR2 BAO and DES-Y5 supernovae is quantified at 2.95±0.04σ2.95 \pm 0.04 \,\sigma within Λ\LambdaCDM, manifesting as a low-dimensional, local likelihood tension (dG1d_G \approx 1), with strongly negative suspiciousness and evidence ratio metrics. This tension is efficiently diffused in w0waw_0w_aCDM (σ\sigma drops to 1.56±0.03σ1.56 \pm 0.03\,\sigma; dG>3d_G > 3), highlighting the role of additional dark energy degrees of freedom in accommodating inter-survey inconsistencies.

Triplet combinations (DESI DR2 + CMB + DES-Y5) exhibit even more severe likelihood conflicts within Λ\LambdaCDM (logS=5.56±0.09\log S = -5.56 \pm 0.09), but again, w0waw_0w_aCDM expands the parameter space to mitigate these tensions. Other supernova catalogues, such as Pantheon+, demonstrate only mild tension, and do not drive model preference toward dynamical dark energy.

Methodological Discrepancies and Their Roots

The persistent magnitude of discrepancy between Bayesian and frequentist outcomes—especially in combinations involving DESI DR2—cannot be categorically attributed to philosophical differences between paradigms. Instead, questions arise regarding the reliability of asymptotic formulae in frequentist significance calculation, given the finite dimensionality and compression of the DESI BAO dataset. As in particle physics, careful Monte Carlo validation of test statistics’ distributions is standard, and similar nested sampling approaches can bridge frequentist-Bayesian computations. Independent geometric analyses corroborate the Bayesian results, suggesting that published frequentist significances overstate the case for evolving dark energy.

Implications and Future Directions

This paper demonstrates that reported evidence for dynamical dark energy is contingent upon dataset selection, and is primarily an artifact of the w0waw_0w_aCDM model’s ability to dilute statistical tension between specific probes. The finding that the DESI BAO and CMB data alone remain consistent with Λ\LambdaCDM (with corroboration from geometric approaches) underscores the necessity for cautious interpretation of headline significances in cosmological model selection. Adopted Bayesian workflows offer robustness against overfitting and provide high-dimensional diagnostics for the origins of inter-dataset disagreement.

Further research must focus on the detailed characterization of supernova systematics—particularly in the DES-Y5 catalogue—and rigorous calibration of frequentist statistics for cosmological datasets of modest effective dimensionality. Publicly available nested sampling chains and tension analysis tools (e.g., the unimpeded framework) set a benchmark for reproducibility and transparency in cosmological inference.

Conclusion

A Bayesian nested sampling model comparison yields only modest or non-significant evidence for dynamical dark energy when analyzing DESI BAO and CMB data, contradicting strong frequentist results. Apparent support for w0waw_0w_aCDM arises uniquely when incorporating the DES-Y5 supernova dataset, and is diagnosed as the resolution of stark inter-dataset tension rather than a discovery of new physical dynamics. These results highlight the imperative for careful model selection, principled tension diagnostics, and validation of inferential machinery in future cosmological surveys.

Ai Generate Text Spark Streamline Icon: https://streamlinehq.com

Explain it Like I'm 14

A simple guide to “A Bayesian Perspective on Evidence for Evolving Dark Energy”

1. What is this paper about?

This paper asks: Does dark energy change over time, or is it a constant? Dark energy is a mysterious “push” making the universe expand faster. The standard model of the universe, called ΛCDM (lambda-CDM), treats dark energy as a constant. A newer idea, often written as w0waCDM, lets dark energy change with time using two extra “knobs” (parameters) called w0 and wa.

A recent big survey (DESI) suggested there’s strong evidence that dark energy might be changing. This paper double‑checks that claim using a different statistical approach and finds a more cautious answer.

2. What questions are they asking?

The authors focus on three plain‑English questions:

  • When we combine today’s best space data sets, which model does the data prefer: constant dark energy (ΛCDM) or changing dark energy (w0waCDM)?
  • Do different ways of judging evidence (two “referees” called frequentist and Bayesian) agree?
  • If some data prefer changing dark energy, is that because the universe truly behaves that way, or because certain data sets don’t quite agree with each other and the extra “knobs” in the changing model smooth over that disagreement?

3. How did they paper it?

They compared models using several types of sky measurements:

  • BAO (Baryon Acoustic Oscillations) from DESI: Think of ancient ripples in the distribution of galaxies—like “sound waves” frozen in space—that act as a giant ruler for distances.
  • CMB (Cosmic Microwave Background) from Planck: The universe’s baby picture, a faint glow left from shortly after the Big Bang.
  • Type Ia supernovae from several catalogs (Pantheon+, Union3, DES‑Y5): Exploding stars used as “standard candles” to measure how fast the universe expands.

They used a Bayesian method to compare models:

  • Frequentist vs. Bayesian: Two fair ways to judge evidence. A frequentist test asks, “How surprising are the data if this model is true?” A Bayesian test asks, “How much would you bet on each model, given both the data and how flexible the model is?”
  • Occam’s razor (Bayesian version): If two models fit the data similarly well, prefer the simpler one with fewer knobs. The “changing dark energy” model has more knobs, so it must earn its complexity.
  • “Bayesian evidence” and “Bayes factor”: Think of these like betting odds. If the odds favor one model, it’s preferred. They compute these odds with a technique called nested sampling, which is like carefully searching a landscape of possibilities to find where the data fit best, while counting how much model flexibility you needed to get there.

They also did a “tension analysis”:

  • Tension means two data sets pull the answer in different directions—like two witnesses telling slightly different stories. If adding extra knobs fixes the disagreement, the model might look “better,” but it could be solving the disagreement rather than revealing new physics.

4. What did they find, and why does it matter?

Key results, in simple terms:

  • DESI BAO + CMB: Using the Bayesian approach, the simpler constant‑dark‑energy model (ΛCDM) is slightly preferred. This disagrees with the frequentist result, which reported a roughly “3 sigma” preference for changing dark energy. (“Sigma” is a way to measure surprise; 3 sigma is notable but not a discovery.)
  • Adding supernova data depends on which catalog you use:
    • With Pantheon+, the Bayesian test favors ΛCDM (constant dark energy).
    • With DES‑Y5, the Bayesian test prefers changing dark energy, and the preference gets strong when combined with DESI + CMB.
  • Why the difference? The authors show that DESI and DES‑Y5 don’t fully agree with each other if you assume ΛCDM. The extra knobs in the changing‑dark‑energy model help “soak up” that disagreement, making it look better. In other words, the model wins because it patches a mismatch between those two data sets, not because every data set independently shouts “dark energy is changing!”

Why this matters:

  • If the preference for changing dark energy mainly comes from a disagreement between data sets, we should be careful before calling it evidence of new physics.
  • The Bayesian method’s built‑in simplicity penalty is doing its job: it only rewards the more complicated model when it truly earns it.

5. What’s the big picture?

This paper urges caution. Claims that dark energy is changing may be influenced by how certain data sets (especially DESI BAO and the DES‑Y5 supernovae) fit together under the standard model. The more flexible model can hide that mismatch, which can look like a “signal” for evolving dark energy.

Implications:

  • We should double‑check supernova catalogs and how different data are combined.
  • Different statistical “referees” can disagree, so it’s wise to look at both and understand why.
  • Before announcing new physics, make sure the evidence isn’t mostly driven by data tension. Future surveys and careful cross‑checks will be key to settling this question.
Ai Generate Text Spark Streamline Icon: https://streamlinehq.com

Knowledge Gaps

The paper leaves several concrete gaps and unresolved questions that future work can address:

  • Quantify the robustness of the reported Bayes factors to prior choices, including systematic sweeps over the w0w_0, waw_a priors (ranges, shapes, reparameterizations), the w0+wa<0w_0+w_a<0 boundary, and the neutrino mass lower bound (0 vs 0.06 eV), and report how these choices change lnB\ln B for each dataset combination.
  • Localize the low-dimensional DESI–DES-Y5 tension to specific parameters and observables by identifying which combinations (e.g., DM/rdD_M/r_d at particular redshifts, H(z)H(z), rdr_d, MBM_B, H0H_0, or w0,waw_0,w_a directions) drive the conflict, and which redshift bins in DES-Y5 and DESI contribute most strongly.
  • Perform targeted systematics tests on DES-Y5 that could generate the observed tension: reprocess with alternative light-curve fitters and calibration pipelines, vary selection cuts, host-mass step treatments, color law and dust models, Malmquist corrections, and redshift-dependent standardization, and propagate these variants into the Bayesian evidence and tension metrics.
  • Assess cross-covariances and shared systematics between probes (Planck CMB, DESI BAO, and SNe) rather than assuming independence in likelihood products; construct and use joint covariance models or nuisance-parameter linking to see how correlated systematics affect Bayes factors and tension diagnostics.
  • Replace or augment the compressed 13-component DESI BAO data vector with the full-shape likelihood and alternative summary statistics to test whether compression-induced information loss or approximations bias the evidence and tension conclusions.
  • Validate nested sampling evidence estimates across samplers and settings by cross-checking with MultiNest, Dynesty, UltraNest, and dynamic nested sampling, and by varying PolyChord hyperparameters (live points, slice sampling settings, stopping criteria) to ensure numerical stability of lnZ\ln \mathcal{Z} and tension metrics.
  • Calibrate the frequentist test statistics used by DESI via Monte Carlo pseudo-experiments tailored to the actual likelihoods and parameter boundaries (including non-Gaussianities and constrained parameter spaces), and compare simulation-based p-values against asymptotic formulae to quantify any breakdowns.
  • Provide a simulation-based mapping between Bayes factors and “sigma” levels (beyond the Sellke bound) for the specific cosmological likelihoods considered, or standardize reporting in terms of odds/Bayes factors without sigma conversion to avoid miscalibration.
  • Explore alternative model extensions that could also alleviate the DESI–DES-Y5 tension and compare their Bayesian evidences: wwCDM (constant ww), curvature (Ωk\Omega_k), varying NeffN_{\mathrm{eff}}, early dark energy, free ALA_L, interacting dark energy, modified gravity, or flexible w(z)w(z) models beyond CPL, and determine which specific physical extensions best resolve the conflict.
  • Test sensitivity to the choice of CMB likelihood and data subsets by repeating the analysis with Planck Plik/Commander, including/excluding CMB lensing, and adding ACT/SPT constraints, and quantify the impact on Bayes factors and tension metrics.
  • Harmonize supernova catalogs (Pantheon+, Union3, DES-Y5) under a common calibration and standardization framework to isolate catalog-driven differences, and perform leave-one-subsample-out analyses within DES-Y5 to identify internal contributors to tension.
  • Investigate whether the wavenumber-dependent modeling and reconstruction systematics in DESI BAO (e.g., non-linear corrections, bias modeling, reconstruction choices) could produce shifts that mimic w0waw_0w_aCDM preferences when combined with DES-Y5.
  • Decompose the reported “low-dimensional” tension using the Bayesian dimensionality metric into interpretable physical directions (e.g., sound horizon scale, absolute SN magnitude, late-time expansion rate), and verify the dimension count and its stability across prior choices and samplers.
  • Quantify the Occam penalty versus fit improvement by reporting Kullback–Leibler divergences and information gains for each dataset combination, clarifying whether w0waw_0w_aCDM’s preference with DES-Y5 is dominated by tension diffusion or genuine likelihood improvement.
  • Provide posterior predictive checks and residual diagnostics for each dataset under both Λ\LambdaCDM and w0waw_0w_aCDM to assess model adequacy beyond parameter-level tension metrics.
  • Justify and calibrate the “look-elsewhere” threshold (2.88σ) used to flag significant tensions within the unimpeded framework, and test the sensitivity of tension classifications to this choice on synthetic datasets.
  • Report the impact of including DESI RSD (growth-rate, fσ8f\sigma_8) alongside BAO distances on the Bayesian model comparison, as growth information can break degeneracies relevant to w(z)w(z).
  • Examine the role of absolute distance calibration in SNe (e.g., Cepheid/TRGB anchors and H0H_0 priors) on the combined evidence with CMB+BAO, explicitly testing how different H0H_0 priors or anchors move the DESI–DES-Y5 tension.
  • Evaluate whether the preferred w0,waw_0,w_a values inferred with DES-Y5 correspond to theoretically viable dark energy models (e.g., thawing/freezing quintessence, avoidance of extreme phantom regimes), and incorporate physics-informed priors to test robustness.
  • Replicate the frequentist analysis with the authors’ own pipeline to isolate the origin of the DESI collaboration’s higher significances (e.g., test statistic definition, nuisance profiling, boundary effects), enabling a like-for-like comparison with the Bayesian workflow.
Ai Generate Text Spark Streamline Icon: https://streamlinehq.com

Practical Applications

Immediate Applications

The following bullet points summarize practical, deployable applications that leverage this paper’s findings, methods, and tools, organized by sector where relevant. Each item includes key assumptions or dependencies.

  • Cosmology research workflows (academia; software)
    • Adopt a dual-framework reporting standard for model comparison results that includes both Bayesian evidence (log-Bayes factor, betting odds) and frequentist significances, especially for claims of “new physics.”
    • Integrate nested sampling (PolyChord via Cobaya) and tension diagnostics from the unimpeded framework into standard cosmology pipelines to routinely quantify dataset compatibility (evidence ratio R, suspiciousness S, dimensionality d_G).
    • Replace asymptotic significance translations with Monte Carlo pseudo-experiments when data vectors are small or approximations may fail, as highlighted for the DESI DR2 compressed BAO vector.
    • Dependencies: Access to DESI/Planck/SN likelihoods; computational resources for nested sampling; community buy-in for dual reporting; careful prior specification (e.g., neutrino mass bounds affect evidence).
  • Data-integration QA for multi-probe cosmology (academia; software)
    • Build a “tension dashboard” for probe combinations (BAO+CMB+SN) that flags low-dimensional conflicts like the DESI DR2–DES-Y5 tension under ΛCDM and monitors how extended models diffuse tensions.
    • Use dashboard outputs to decide which catalogues to include (e.g., prefer Pantheon+/Union3 when tensions persist) and to trigger systematic audits of specific data products (e.g., SN standardization and calibration).
    • Dependencies: Availability of the unimpeded software/database; access to probe likelihoods; domain expertise to interpret tension diagnostics; governance to act on flagged tensions.
  • Reinterpretation and communication of evolving dark energy claims (academia; policy)
    • Update public statements and internal memos to reflect that DESI+Planck evidence modestly favors ΛCDM under Bayesian analysis, and that preferences for w0waw_0w_aCDM are driven by specific inter-dataset tensions (primarily DES-Y5).
    • Adopt “betting odds” language for Bayes factors and avoid over-reliance on single-catalogue-driven significances.
    • Dependencies: Institutional acceptance of Bayesian communication standards; training in evidence-based phrasing; coordination with collaborations and journals.
  • Peer review and journal policy enhancements (policy; academia)
    • Encourage or require submissions to report Bayes factors alongside frequentist significances, include sensitivity analyses to priors, and validate asymptotic approximations with pseudo-experiments where feasible.
    • Dependencies: Editorial policy changes; reviewer expertise; computational budget for validation runs.
  • Ready-to-use software assets (software; academia)
    • Immediate adoption of unimpeded public nested-sampling databases and scripts for model comparison and tension analysis; use anesthetic for posterior visualization and Cobaya for likelihood orchestration.
    • Share reproducible analysis artifacts (chains, priors, configuration files) to standardize cross-group comparisons.
    • Dependencies: Familiarity with Python ecosystems; containerization for portability; storage for chain databases.
  • Cross-domain data-fusion QA pilots (healthcare, finance, robotics, energy; software)
    • Apply tension metrics (e.g., suspiciousness, evidence ratio) to integrated datasets in other sectors to detect when added model flexibility merely absorbs inter-source conflicts rather than improving truthfulness.
    • Healthcare: Combine EHR, trial outcomes, and registries; flag low-dimensional tensions that point to calibration or cohort biases.
    • Finance: Fuse macro indicators and alternative data; identify conflicts where complex models improve fit by masking incompatible signals.
    • Robotics and autonomous systems: Sensor fusion pipelines that distinguish true multimodal synergy from conflicts that extra parameters “paper over.”
    • Energy forecasting: Integrate grid, weather, and market signals; detect source-level inconsistencies before ensemble modeling.
    • Dependencies: Domain-specific likelihood modeling; access to high-quality data; translation of cosmology metrics to sector-specific inference; stakeholder training.
  • A/B testing and product analytics (software; industry)
    • Use Bayesian evidence and odds rather than only p-values to compare feature variants; report tension-like diagnostics when multiple user segments or data sources disagree at low dimension.
    • Dependencies: Instrumentation to collect segment-level data; Bayesian tooling (PyMC/Stan bridging, or nested sampling when evidence is needed); cultural shift to odds-based decision-making.

Long-Term Applications

The following items require further research, development, scaling, or institutional changes before broad deployment.

  • Standardized, cross-disciplinary framework for dataset tension diagnostics (software; academia; policy)
    • Generalize the unimpeded evidence framework beyond cosmology into a domain-agnostic library (“TensionLab”) with APIs for PyMC/Stan, standardized metrics (R, S, d_G), and reporting templates.
    • Dependencies: Methodological extensions to diverse likelihoods; benchmarking datasets; sustained maintainership and funding.
  • Cloud/HPC evidence-as-a-service (software; industry; academia)
    • Build managed services that run nested sampling at scale for evidence and tension analysis, with reproducibility primitives (versioned priors, chains, and configurations) and dashboards.
    • Dependencies: HPC/cloud integration; cost control; secure handling of proprietary datasets; optimization of nested sampling for large parameter spaces.
  • Survey and experiment design optimized for tension resolution (academia; policy)
    • Use low-dimensional tension diagnostics to target calibration, selection functions, and instrument design that specifically reduce conflicts (e.g., SN photometric calibration improvements that mitigate DESI–SN tensions).
    • Pre-register analysis frameworks that include both Bayesian evidence and tension metrics to prevent model extensions that simply absorb specific conflicts.
    • Dependencies: Collaboration-wide planning; calibration infrastructure; incentives for pre-registration and open analysis.
  • Formalized dual-framework standards for “extraordinary evidence” claims (policy; academia)
    • Establish guidelines requiring concurrent Bayesian evidence thresholds (e.g., minimum odds) and frequentist significance, plus validated asymptotics, before making claims of new physics or major effects.
    • Dependencies: Consensus on thresholds; community enforcement; education and training.
  • Tension-aware machine learning and meta-analysis (software; healthcare; finance)
    • Develop ML algorithms and meta-analytic workflows that incorporate evidence and tension metrics to avoid overfitting to conflicting sources and to prioritize data cleaning or reweighting.
    • Dependencies: Algorithmic research; integration into existing MLOps stacks; high-quality labels and provenance metadata.
  • Education and training programs (academia; policy)
    • Build curricula and workshops that teach evidence-based model comparison, tension diagnostics, and the pitfalls of asymptotic significance in small or compressed datasets, using this paper as a case paper.
    • Dependencies: Course development resources; faculty training; uptake in graduate programs and professional development.
  • Regulatory and risk governance adaptation (policy; finance; healthcare)
    • Incorporate tension diagnostics into regulatory guidance for meta-analyses, risk models, and safety-critical decision systems to ensure extended models aren’t preferred solely because they resolve specific dataset conflicts.
    • Dependencies: Regulator engagement; pilots demonstrating benefit; legal and compliance alignment.

Each long-term application presumes continued tool maturation, broader community adoption, and robust domain-specific validation. In all cases, the feasibility depends on transparent priors, accessible data, reproducible chains, sufficient compute, and governance structures willing to act when tensions—not true signals—drive model preferences.

Ai Generate Text Spark Streamline Icon: https://streamlinehq.com

Glossary

  • asymptotic formulae: Large-sample approximations used to translate test statistics into significances. "it is reasonable to question not the frequentist framework per se, but the applicability of asymptotic formulae in this instance."
  • BAO: Baryon Acoustic Oscillations; a standard ruler from early-Universe sound waves used to constrain cosmology. "We analyse DESI DR2 BAO data~\cite{desi2025,BAOData}"
  • Bayes factor: The ratio of Bayesian evidences for two models, quantifying how the data compare them. "Following Trotta~\cite{2008ConPh..49...71T}, we convert Bayes factors to Gaussian significance via (i) Bayes factor to pp-value"
  • Bayesian dimensionality metric: A measure of the effective dimensionality of the parameter subspace that is constrained by the data. "which the Bayesian dimensionality metric diagnoses as a highly localised, low-dimensional conflict (dG=0.989±0.073d_G = 0.989 \pm 0.073)."
  • Bayesian evidence: The marginal likelihood of the data under a model, integrating the likelihood over the prior. "We compute Bayesian evidence Z=P(DM)\mathcal{Z} = P(D|\mathcal{M})"
  • Bayesian model comparison: Evaluating and ranking models by their Bayesian evidences or Bayes factors. "we present a direct Bayesian model comparison"
  • Bayesian Occam's razor: The automatic penalization of overly complex models via prior-volume integration in the evidence. "(Bayesian Occam's razor)"
  • CAMB: A cosmology code that computes theoretical predictions like CMB anisotropies for parameter inference. "and CAMB~\cite{Lewis:1999bs}"
  • CamSpec likelihood: A Planck CMB likelihood function for temperature and polarization power spectra. "Planck 2018 CMB (CamSpec likelihood~\cite{CamSpec2021})"
  • CMB: Cosmic Microwave Background; relic radiation used as a precise cosmological probe. "for the key combination of the DESI DR2 BAO and the Planck CMB data,"
  • Cobaya: A software framework for Bayesian analysis and sampling in cosmology. "via Cobaya~\cite{Torrado2021Cobaya,cobayaascl}"
  • compressed data vector: A reduced summary of a dataset with fewer components preserving key information. "a 13-component compressed data vector"
  • degrees of freedom: The number of independent parameters in a model that can vary to fit data. "precisely because its additional degrees of freedom are effective at resolving this specific conflict"
  • DESI DR2: The second data release of the Dark Energy Spectroscopic Instrument, providing BAO measurements. "The DESI DR2 data release~\cite{desi2025} reports up to 4.2σ\sigma preference"
  • DES-Y5: The Dark Energy Survey Year 5 Type Ia supernova catalogue. "the DES-Y5 supernova catalogue~\citep{descollaboration2025darkenergysurveycosmology}"
  • evidence ratio: A metric comparing evidences to assess dataset consistency or tension. "This is the only pairwise combination to yield a negative evidence ratio (logR0.17\log R \approx -0.17)"
  • frequentist hypothesis test: A statistical test using data-derived test statistics (e.g., likelihood ratios) to assess models. "based on a frequentist hypothesis test derived from a likelihood ratio based test statistic."
  • frequentist significance: The sigma-level translation of a p-value indicating strength of evidence against a null. "3.1σ\sigma frequentist significance in favoring w0waw_0w_aCDM."
  • Gaussian significance: The equivalent number of standard deviations corresponding to a given p-value. "we convert Bayes factors to Gaussian significance via (i)"
  • geometric analysis: A methodology assessing BAO data consistency using geometric constructs rather than full likelihoods. "finds independent support from the geometric analysis of Efstathiou~\cite{Efstathiou2025BAO}"
  • inverse normal cumulative distribution function: The quantile function mapping probabilities to z-scores in a standard normal. "where Φ1\Phi^{-1} is the inverse normal cumulative distribution function."
  • likelihood ratio: The ratio of likelihoods under competing models, used as a test statistic. "likelihood ratio based test statistic."
  • log-Bayes factor: The natural logarithm of the Bayes factor, commonly reported for model comparison. "log-Bayes factor lnB=0.57±0.26\ln B = -0.57{\scriptstyle\pm0.26}"
  • look-elsewhere threshold: A significance threshold accounting for multiple comparisons across parameter space. "exceeding our 2.88σ2.88\sigma look-elsewhere threshold."
  • Monte Carlo pseudo-experiments: Simulated datasets used to validate statistical approximations and calibrate significances. "by means of extensive Monte Carlo pseudo-experiments"
  • nested sampling: A Monte Carlo algorithm for efficiently computing Bayesian evidence and exploring posteriors. "using nested sampling to compute the Bayesian evidence"
  • Pantheon+: A compilation of Type Ia supernovae used for cosmological parameter inference. "Pairwise combinations with Pantheon+ data strengthen the Bayesian evidence for Λ\LambdaCDM"
  • posterior: The probability distribution over parameters given the data and prior. "Posterior comparisons in w0waw_0w_aCDM"
  • priors: Probability distributions encoding parameter beliefs before observing the data. "we adopt DESI's priors (w0[3,1]w_0 \in [-3, 1], wa[3,2]w_a \in [-3, 2] with w0+wa<0w_0 + w_a < 0)"
  • suspiciousness: A tension metric quantifying direct likelihood-level conflict between datasets. "A strongly negative suspiciousness (logS=3.83±0.03\log S = -3.83 \pm 0.03) confirms a direct likelihood conflict"
  • tension analysis: A systematic evaluation of the consistency between datasets using multiple complementary metrics. "By performing a comprehensive tension analysis, employing five complementary metrics,"
  • Union3: A Type Ia supernova compilation used as an alternative to Pantheon+ and DES-Y5. "Union3 (orange)"
  • unimpeded evidence framework: A suite of tools and metrics for Bayesian model comparison and tension diagnostics. "using the suite of metrics provided by the unimpeded evidence framework~\cite{UnimpededPaper,UnimpededSoftware}"
  • w0waCDM (w0waw_0w_aCDM): A cosmological model with a time-varying dark energy equation of state parameterized by w0w_0 and waw_a. "the w0waw_0w_aCDM model is preferred precisely because its additional parameters act to resolve this specific dataset conflict."
  • LambdaCDM (Λ\LambdaCDM): The standard cosmological model with a cosmological constant and cold dark matter. "Negative lnB\ln B favours Λ\LambdaCDM"
Dice Question Streamline Icon: https://streamlinehq.com

Open Problems

We found no open problems mentioned in this paper.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets

This paper has been mentioned in 1 tweet and received 1 like.

Upgrade to Pro to view all of the tweets about this paper:

Youtube Logo Streamline Icon: https://streamlinehq.com