Papers
Topics
Authors
Recent
Search
2000 character limit reached

Verbalized Probability in AI

Updated 5 March 2026
  • Verbalized probability is a model-agnostic method that explicitly expresses an AI system's confidence through numerical or linguistic cues.
  • It relies on advanced prompting strategies and calibration metrics, such as Expected Calibration Error, to accurately measure uncertainty.
  • Applications include uncertainty quantification, factuality assessment, and probabilistic reasoning, while challenges like overconfidence and ambiguity persist.

Verbalized probability is the practice of eliciting explicit probability estimates or confidence scores from human or artificial agents in natural language, typically as part of task outputs. In modern AI, this term almost always refers to the explicit numerical (or categorical) confidence or probability that a LLM or related system emits alongside, or as part of, an answer—purporting to quantify the likelihood that its response is correct. Unlike low-level internal probabilities (e.g., softmax token logits) or sampling-based uncertainty measures, verbalized probability is fundamentally an output-level, model-agnostic construct, enabling black-box uncertainty quantification, improved interpretability, and practical calibration for safety-critical applications (Yang et al., 2024).

1. Definitions and Conceptual Scope

A verbalized probability CC is a number (e.g., C=0.73C=0.73 or 73%73\%) produced by a model as part of its output, explicitly stating its confidence that the answer YY to prompt XX is correct:

  • Formalism: C=UQ(X,Y)C = \mathrm{UQ}(X, Y), where UQ denotes uncertainty quantification as executed in language.
  • Contrast with other uncertainty quantification (UQ) methods:
    • Internal token probabilities: Formed by aggregating per-token softmax outputs, but typically fail to capture high-level semantic or factual uncertainty, and are unavailable in most black-box APIs.
    • Sampling-based approaches: Estimate uncertainty by generating multiple responses and measuring consensus, but are computationally intensive (sampling overhead is linear in the number of samples) and not independently accessible without repeated generation.
  • Desiderata: Model- and prompt-agnostic (applicable with minimal requirements); low cost (a few extra tokens); can be solicited zero-shot or with few-shot anchoring (Yang et al., 2024, Wang et al., 2024).

Verbalized probability also encompasses categorical or fuzzy/linguistic statements (“likely,” “very probable,” “almost certain”), but in practice, especially for LLMs, the literature overwhelmingly focuses on explicit numerical values.

2. Prompting and Elicitation Strategies

The reliability and calibration of verbalized probabilities depend critically on how the model is prompted. Major dimensions include:

  • Score range: Integer percentage (0–100), floating-point probability (0.0–1.0), or categorical scale (“very low” to “very high” mapped to numbers).
  • Score formulation: Variants such as “confidence score quantifying how confident you are in the correctness of your answer,” “probability that your answer is correct,” or explicit mapping between confidence and probabilistic correctness.
  • Advanced descriptions: Adding context such as “take your uncertainty in the prompt, task difficulty, and knowledge availability into account” significantly improves calibration for large models.
  • Few-shot anchoring: Providing hand-picked examples covering the confidence scale (e.g., 0.1, 0.3, 0.5, 0.7, 0.9) helps anchor the number line and encourages a richer range of outputs.
  • “Combo” prompt: Combining floating-point outputs, advanced description, 5-shot demonstration, and explicit “best guess” phrasing yields the most robust calibration in large LLMs.
  • Baselines: Previous strategies include chain-of-thought, self-consistency, outputs ranked by log probability, and more.

Well-chosen prompting is critical: a poorly designed prompt can lead to ECEs (Expected Calibration Error) of 0.2–0.3, whereas the “combo” design can cut the error nearly in half, particularly in models >70B parameters (Yang et al., 2024).

3. Evaluation: Datasets, Metrics, and Calibration

Benchmarking verbalized probability involves both the range of tasks and models, and the rigor of evaluation metrics:

  • Datasets: Cover closed-book, closed-ended, and objective tasks (arc-c, arc-e, commonsense_qa, logi_qa, mmlu, sciq, social_i_qa, trivia_qa, truthful_qa, and others).
  • Models: Ranging from Gemma 1.1-2B, Llama 3-8B/70B, Qwen-1.5-7B/110B, to GPT-3.5-turbo and GPT-4o (Yang et al., 2024).
  • Calibration metrics:
    • Expected Calibration Error (ECE):

    ECE=m=1MBmNacc(Bm)conf(Bm)\mathrm{ECE} = \sum_{m=1}^M \frac{|B_m|}{N} \cdot |\mathrm{acc}(B_m) - \mathrm{conf}(B_m)|

    Measures the average absolute deviation, over confidence bins, between predicted confidence and empirical accuracy. - Brier score (not emphasized):

    Brier=1Ni=1N(piyi)2,yi{0,1}\mathrm{Brier} = \frac{1}{N} \sum_{i=1}^N (p_i - y_i)^2,\quad y_i\in\{0,1\} - Reliability diagrams: Bar plots of empirical accuracy vs. predicted confidence. - Informativeness: Number of distinct scores (ndistinctn_\text{distinct}) and their variance, to flag trivial/confounded outputs. - Meaningfulness: KL divergence between a model's confidence distribution on a single dataset versus all datasets, probing sensitivity to task difficulty.

  • Key calibration findings (Yang et al., 2024):

    • Small models (7-8B): Calibration improves with simple 0–1 float prompts; high shot/informative prompts can degrade calibration (overfitting).
    • Large models (≥70B or GPT-4): Rich prompts (combo format) yield ECE 0.07\approx 0.07 (gap between confidence and accuracy drops to 7%).

4. Applications, Advantages, and Limitations

Verbalized probability plays several crucial roles across the LLM ecosystem and adjacent areas:

  • Uncertainty quantification for black-box LLMs: Enables trust assessment and risk-sensitive decision-making even when internal model states are inaccessible (Yang et al., 2024).
  • Diversity enhancement: Asking a model to output multiple completions with associated probabilities (Verbalized Sampling, VS) allows debiasing and recovery of base-model generative diversity, especially after RLHF and preference fine-tuning have induced mode collapse (Zhang et al., 1 Oct 2025).
  • Calibration and explanation: Verbalized probabilities serve as externally visible confidence estimates, facilitating post-hoc calibration (e.g., via temperature scaling and Platt scaling, applied to invert-softmaxed values) (Wang et al., 2024).
  • Long-form generation and factuality assessments: Reinforcement learning–based methods can train LLMs to assign fine-grained confidence to each generated fact/sentence, dramatically improving sentence-level calibration without expensive sampling (Zhang et al., 29 May 2025).
  • Probabilistic reasoning and simulation: Embedding entire probabilistic algorithms (e.g., rejection sampling, verbalized graphical modeling) into prompt space enables LLMs to reason about uncertain variables and perform proper stochastic inference (Xiao et al., 11 Jun 2025, Huang et al., 2024).
  • Human factors and interpretability: While numerical confidence is favored in technical contexts, psycholinguistic investigations (e.g., “words of estimative probability”) show that verbal categories are widely used by both laypeople and experts, though with significant variance and asymmetry in interpretation (Sileo et al., 2022, Willems et al., 2019).
  • Limitations: LLM-verbalized probabilities may reflect stylistic conventions or “surface-level” linguistic anchoring rather than genuinely content-grounded beliefs—especially in high-capacity models. Training regimes often teach LLMs to sound confident rather than to be accurate, leading to persistent overconfidence and miscalibration (Xia et al., 15 Jan 2026). Empirical studies also show that internal token-logit confidence typically outperforms verbalized probability as a reliability signal, but requires access to hidden states and external calibration (Ni et al., 2024).

5. Calibration, Post-Processing, and Theoretical Analysis

Several calibration and post-processing techniques have been developed specifically for verbalized probability:

  • Post-hoc temperature and Platt scaling: Rather than applying “re-softmax” (which would flatten the verbalized probability distribution), invert the softmax to recover surrogate logits, then perform standard temperature or logistic scaling. This approach systematically improves calibration metrics, notably lowering ECE on sentiment, intent, and emotion datasets (Wang et al., 2024).
  • Reward-shaping and RL-based calibration: RL with preference or ordinal losses aligns sequence-level confidence with external “oracle” fact-checkers, achieving low sentence-wise calibration error in long-form factual generation (Zhang et al., 29 May 2025). The log-likelihood–based reward penalizes overconfident errors preferentially.
  • Verbalized rejection sampling (VRS): Implements classic probabilistic rejection sampling entirely in language, prompting the LLM to evaluate acceptance stepwise—substantially reducing sampling bias and increasing diversity of output (Xiao et al., 11 Jun 2025).
  • Probabilistic graphical modeling via verbalization: vPGM prompts the LLM to discover and walk through latent-variable Bayesian networks verbally, aggregating per-chain confidence estimates for accurate marginal probability estimation; this approach yields state-of-the-art calibration on compositional reasoning tasks (Huang et al., 2024).

These methods depend on both prompt design and the careful alignment of model outputs to either ground-truth accuracy or internal probability (direct confidence alignment, DCA). DCA can improve or degrade alignment depending on model architecture and the calibration quality of internal token probabilities (Zhang et al., 12 Dec 2025). Calibration improvements are often model- and dataset-specific.

6. Human and Societal Perspectives

Psychometric and cross-linguistic studies provide substantial context for the interpretation of verbal confidence:

  • Categorical verbal probability: Adjectives and adverbs such as “likely,” “improbable,” “almost certain,” or “highly likely” are mapped empirically to numeric intervals, but show substantial inter-individual, cross-linguistic, and contextual variability (Willems et al., 2019, Sileo et al., 2022).
  • Fuzzy-set and possibility theory: Human subjects’ use of verbal uncertainty categories can be modeled with trapezoidal membership functions over [0,1]; once calibrated for an individual, these mappings track actual success rates and enhance internal consistency (Zimmer, 2013).
  • Communication risks: Verbal probability phrases are too ambiguous for high-stakes communication; even among statisticians, interpretations vary by tens of percentage points. The bidirectionality of complementary expressions fails (e.g., “likely” + “unlikely” ≠ 100%), and context dependence is strong (Willems et al., 2019).
  • Explaining probabilistic updates: Empirical work shows that relative update phrases (“much more likely,” “a little less likely”) correspond best to fixed differences in probability (Δp), not ratios or odds, further emphasizing the need for numerically anchored communication in applications requiring precision (1304.1501).

7. Practical Recommendations and Open Problems

Verbalized probability, properly elicited and post-processed, provides a low-overhead, model-agnostic route to trustworthy uncertainty quantification in LLMs. Best practices include:

  • Adopt floating-point responses in [0,1], with explicit instructions tying the score to correctness probability and sources of uncertainty.
  • Use 5-shot calibration spanning the output range to anchor model scale.
  • For class distributions, prompt for Python-dict style output and immediately apply invert-softmax scaling for post-calibration.
  • Prefer deterministic generation temperature to minimize stochastic variations in output probabilities (Wang et al., 2024).
  • Monitor for surface-level mimicry; apply content-grounding diagnostics (e.g., TracVC) to ensure confidence reflects evidential support (Xia et al., 15 Jan 2026).
  • In high-stakes or international contexts, always accompany verbal expressions (“likely,” “very probable”) with explicit numeric anchors to reduce miscommunication (Willems et al., 2019).

Outstanding challenges include robustly aligning verbalized probability with both internal model belief and empirical reality (true correctness), mitigating the influence of generic “confidence” phrases, and systematic fine-tuning for content-grounded calibration across models and domains.

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Verbalized Probability.