BigToM Forward Belief Scenarios
- BigToM Forward Belief Scenarios are systematically controlled ToM tasks that assess first-order belief reasoning in artificial agents.
- They employ formal causal templates and diverse inference methods, including Bayesian updates and regression approaches, to simulate belief dynamics.
- Empirical results show that while models like GPT-4 achieve near-ceiling performance, smaller LLMs exhibit significant biases in belief attribution.
BigToM Forward Belief Scenarios are a class of systematically controlled theory-of-mind (ToM) reasoning tasks designed to probe first-order belief tracking in artificial agents, particularly LLMs. These scenarios operationalize the classic false-belief paradigm within computational and neural frameworks, providing a rigorous platform for benchmarking belief attribution, stepwise uncertain inference, and belief dynamics over multi-agent and multimodal environments. BigToM forward belief scenarios have become a key empirical and methodological touchstone for evaluating and developing ToM capabilities in LLMs and related planning systems.
1. Formal Definition and Structure of Forward Belief Scenarios
A BigToM forward belief scenario consists of a narrative world and an agent embedded in a sequence of discrete events , where is the last event observed by and is a state-changing event not observed by . The agent’s belief at time is expressed as the subjective probability
At time , after an unobserved event causes the true world state to shift from to , the scenario probes: "What does agent believe about the world?" Formally, the task is to compute , recognizing that 's belief remains anchored to the last observed state despite the actual world changing to (Chulo et al., 19 Nov 2025, Gandhi et al., 2023).
Each BigToM scenario is a tuple:
- Story: Controlled narrative instantiating abstract causal relations.
- Question: Queries the agent’s belief after an unobserved event.
- Choices: Typically binary, e.g. does believe or ?
This framework isolates pure belief state inference and enables tightly controlled experimental conditions over perceptual access and event salience.
2. Task Generation via Causal Templates
BigToM scenarios are constructed using formal causal templates to specify the syntactic and semantic scaffolding of ToM reasoning (Gandhi et al., 2023). The main variables are:
- : Context (agent, setting)
- : Desire (agent’s goal)
- , : Initial percept and derived belief
- : Exogenous causal event (world-altering)
- : Percept of (or its absence)
- : Agent’s post-event belief
- : Subsequent action
The causal template induces the following dependency graph:
Concrete BigToM items are synthesized by sampling narrative parameters with LLMs such as GPT-4, enforcing constraints that ensure the intended inference is controlled (e.g., by specifying whether the agent perceives the causal event). This procedural generation supports thousands of evaluation items with high linguistic and situational diversity, matching expert quality according to blinded human raters (Gandhi et al., 2023).
3. Computational Models and Inference Mechanisms
Multiple architectures and paradigms implement forward belief reasoning within the BigToM framework:
a. Scalable Bayesian ToM Planning
A stepwise Bayesian update decomposes the belief inference task. At each time step , the agent’s belief is a probability distribution over hidden states , recursively updated by:
- Prediction:
- Correction:
A weak-to-strong control architecture integrates ToM-specialized likelihoods from small LMs with context-rich priors from large LMs, delivering strong generalization and scalability in rich multimodal domains (Zhang et al., 2 Jun 2025).
Example Forward Scenario Calculation
| Step | Observation | Updated Belief |
|---|---|---|
| $0$ | Uniform over cabinets | = Uniform() |
| $1$ | is empty | : Zero , re-normalize over |
| $2$ | observation | : Sharpened towards |
Over successive steps, the posterior converges to the true location as in a Bayesian filter (Zhang et al., 2 Jun 2025).
b. Regression Approaches
Symbolic goal regression reduces the subjective probability after a sequence of actions and observations to an initial state query via recursive term/formula regression operators, sidestepping the need to explicitly unroll the state transition and observation chain (Belle et al., 2013). For ToM and nested beliefs, each additional order introduces another set of regression fluents, allowing collapse to a single initial-state inference.
c. Feed-forward Belief Propagation
In belief networks and neural networks, forward belief propagation analytically tracks marginal means and variances through each network layer for uncertain inputs, using closed-form moment equations adapted for the activation nonlinearity (e.g., sigmoid, ReLU). This provides a differentiable and scalable means of maintaining uncertainty during forward passes, necessary for efficient ToM architectures based on deep networks (Shekhovtsov et al., 2018).
4. Evaluation, Performance Benchmarks, and Error Profiles
BigToM forward belief scenarios are central to recent ToM evaluation efforts. Key findings include:
- Human baselines: ≈100% (True-Belief), ≈92% (False-Belief), combined ≈92% accuracy.
- GPT-4: Near-ceiling performance in 0-shot, with forward belief (false-belief) accuracies .98–.99, but with a mild anchor bias when the agent's initial belief is reiterated (.97→.90) (Gandhi et al., 2023).
- Smaller LLMs: Dramatically lower, e.g., LLaMA-65B: .41, text-davinci-003: .25, gpt-3.5-turbo: .31, Claude-v1.3: .59, Claude-2: .52.
- Contrastive Activation Addition (CAA) steering: Applied to Gemma-3-4B, CAA increased accuracy from 32.5% to 46.7% on 1,000 forward-belief scenarios, an absolute gain of 14.2% (Chulo et al., 19 Nov 2025).
- Hybrid planners (e.g., 8B→405B chain): Achieve 81.3% on multimodal benchmarks, with strong generalization to unseen domains and +4.6% improvement over prior state-of-the-art (Zhang et al., 2 Jun 2025).
Weaker models typically fail to track perceptual access, defaulting to world-state reasoning (what is true) rather than the agent’s belief state (what they think), or they anchor erroneously on initial beliefs.
5. Planning and Symbolic Formulations
Belief-space planners leverage planning graph heuristics adapted for progression in uncertain, partially observable domains (Bryce et al., 2011). Key constructs include:
- Belief state distances () via aggregation (max, sum, overlap).
- Relaxed plan or level-heuristics computed via single, multiple, or labelled-uncertainty graphs (LUGs), with tracking of variable dependencies in BDD data structures.
- Integration into A*/AO*-style planners allows scalable search and pruning in large belief spaces relevant for robotic and social ToM tasks.
Such methods generalize to hierarchical or nested ToM settings in BigToM, provided variables (actions, observations) are suitably factored or extended to account for higher-order beliefs.
6. Mechanistic Insights from Neural ToM Studies
Recent studies decompose LLM ToM improvements in forward belief scenarios by analyzing internal activations using linear probes across 45 cognitive actions (Chulo et al., 19 Nov 2025). Contrastive activation steering (CAA) was shown to:
- Enhance emotion-related activity: emotion perception (+2.23 layers), emotion valuing (+2.20 layers).
- Suppress analytical routines: questioning (-0.78), convergent thinking (-1.59).
The principal mediator of improved ToM performance in LLMs is increased engagement of emotional processing circuitry rather than traditional analytic reasoning. This challenges the assumption that logical chain-of-thought is the key path to social inference in LLMs; perspective-taking arises from affective mechanisms.
7. Belief Maintenance and Scenario Management Systems
Belief Maintenance Systems (BMS) generalize classical Truth Maintenance Systems (TMS) to infinite-valued logic, allowing for dynamic and reversible assertion of hypothetical facts and propagation of degree-of-belief (Falkenhainer, 2013). In the context of BigToM-style scenarios, BMS:
- Represents each fact as a node with belief degrees.
- Utilizes invertible Dempster–Shafer mass assignments to support dynamic forward scenario evaluation ("what if ").
- Allows rollback to previous belief states by removing assertions and repropagating support.
This infrastructure enables fine-grained, reversible scenario analysis in forward belief tasks, providing a foundation for rich what-if explorations in ToM and planning.
BigToM forward belief scenarios operationalize first-order belief reasoning in computational contexts, offering a rigorous benchmark for ToM, planning, and neural inference methods. Their systematic structure, empirical tractability, and extensibility to multimodal and hierarchical settings make them central to ongoing progress in machine social cognition. Recent advances clarify both the statistical and mechanistic pathways underlying LLM ToM abilities and provide a roadmap for future research into scalable, robust, and transparent belief inference systems.
Sponsored by Paperpile, the PDF & BibTeX manager trusted by top AI labs.
Get 30 days free