Papers
Topics
Authors
Recent
2000 character limit reached

Free-Energy of the Expected Future (FEEF)

Updated 9 December 2025
  • FEEF is a functional in active inference that extends variational free energy to future-oriented planning by balancing expected information gain and goal-directed utility.
  • It decomposes expected free energy into distinct epistemic (information gain) and pragmatic (extrinsic value) components, guiding policy selection.
  • Minimizing FEEF underpins adaptive behaviors in robotics, reinforcement learning, and neuroscience by unifying perception, action, and learning.

The Free-Energy of the Expected Future (FEEF) is a foundational functional in active inference and Bayesian decision-making, unifying epistemic (exploratory) and pragmatic (goal-directed) drives within a single information-theoretic objective. FEEF extends variational free energy from past and present inference into the future—enabling agents to balance information gain and anticipated utility through policy selection. The minimization of FEEF, or equivalently expected free energy (EFE), underlies perception, action, and learning in active inference frameworks, with broad applicability in neuroscience, robotics, reinforcement learning, and computational models of emotion.

1. Mathematical Foundation and Core Definitions

FEEF generalizes the variational free energy (VFE) bound on surprise to the domain of future-oriented planning. Given a generative model over hidden states xt:Tx_{t:T} and observations ot:To_{t:T}, and a candidate policy π\pi, FEEF is:

FEEF(π)=DKL[Q(ot:T,xt:Tπ)p~(ot:T,xt:T)]\mathbf{FEEF}(\pi) = \mathbf{D}_{KL}\left[\,Q(o_{t:T}, x_{t:T}\mid \pi)\,\|\,\tilde p(o_{t:T}, x_{t:T})\,\right]

where:

  • Q(ot:T,xt:Tπ)Q(o_{t:T}, x_{t:T}\mid \pi) is the predictive posterior over future outcomes and states under policy π\pi,
  • p~(ot:T,xt:T)\tilde p(o_{t:T}, x_{t:T}) encodes desired or preferred future trajectories.

A single-step decomposition yields:

FEEFτ(π)=EQ(oτ,xτπ)[lnQ(oτ,xτπ)  lnp~(oτ,xτ)]\mathbf{FEEF}_\tau(\pi) = \mathbb{E}_{Q(o_\tau, x_\tau|\pi)} \left[\ln Q(o_\tau, x_\tau|\pi)\ -\ \ln \tilde p(o_\tau, x_\tau)\right]

This KL-divergence formalization grounds FEEF as a trajectory-level mismatch between predicted and preferred futures (Millidge et al., 2020).

2. Decomposition: Epistemic and Pragmatic Drives

Under Markov and mean-field factorizations, FEEF and EFE exhibit a canonical decomposition:

G(π)=EQ(oτπ)DKL(Q(sτoτ,π)Q(sτπ))Epistemic ValueEQ(oτπ)[logP(oτ)]Extrinsic ValueG(\pi) = \underbrace{\mathbb{E}_{Q(o_\tau|\pi)}\, D_{KL}(Q(s_\tau|o_\tau, \pi)\,\|\,Q(s_\tau|\pi))}_{\text{Epistemic Value}} - \underbrace{\mathbb{E}_{Q(o_\tau|\pi)}[\log P(o_\tau)]}_{\text{Extrinsic Value}}

  • Epistemic value: Encoded as expected information gain over latent states, incentivizing visits to “uncertain” regimes for improved model certainty.
  • Pragmatic value: Captured as the expected log-utility (log probability of preferred outcome), driving the fulfillment of agent goals (Sajid et al., 2021, Pattisapu et al., 2 Jul 2024).

Alternative standard decompositions, such as risk plus ambiguity or entropy plus expected energy, are mathematically equivalent or stand in bound relations depending on how priors over states or observations are handled (Champion et al., 22 Feb 2024).

3. Relationship to Variational Free Energy and Expected Free Energy

FEEF clarifies ambiguities in extending VFE into the domain of action planning:

  • VFE: Targets inference on current (realized) data and bounds surprise.
  • EFE: Introduced as a generalization using priors over future states, decomposing into extrinsic (risk) and intrinsic (information gain) components, but does not directly generalize VFE; it generally forms a lower bound on expected future surprise.
  • FEF: (Free-Energy of the Future) Direct temporal extension of VFE but penalizes information gain and discourages exploration.
  • FEEF: Unifies inference and planning as a single KL-divergence—matching predictive and desired trajectory distributions—and subsumes both extrinsic and epistemic terms (Millidge et al., 2020).

This yields a unique property: With real data, FEEF reduces to VFE; in predictive rollouts, it becomes the planning objective guiding action selection.

4. Limiting Cases and Interpretation

FEEF functionals smoothly interpolate between exploration and exploitation:

  • Pure exploration (optimal Bayesian design): Setting preference terms to uniform (removing goal-directed drives) makes FEEF minimize negative mutual information between hidden states and observations, resulting in pure information-gain maximization.
  • Pure exploitation (classical expected utility): Neglecting epistemic terms reduces FEEF to the negative expected utility under preference distributions (Sajid et al., 2021).

This establishes FEEF as the “master functional” behind Bayesian experiment design, classical expected utility, and their synthesis in active inference.

5. Practical Computation and Policy Selection

Active inference agents employing FEEF implement a three-stage process at each timestep:

  1. Perceptual inference: Minimize VFE for current observations via variational Bayes/message passing.
  2. Policy evaluation: For each candidate policy, calculate FEEF (or EFE) over a finite horizon using predictive state and outcome distributions.
  3. Policy selection: Sample or deterministically choose actions using a softmax over negative FEEF:

P(π)=exp(FEEF(π))πexp(FEEF(π))P(\pi) = \frac{\exp(-\mathbf{FEEF}(\pi))}{\sum_{\pi'} \exp(-\mathbf{FEEF}(\pi'))}

Implementations leverage factor graph representations and efficient message passing to achieve tractability in large, partially observable spaces (Nuijten et al., 4 Aug 2025). Epistemic priors can be operationalized as entropy-minimizing constraints within the variational framework, unifying policy and inference optimization (Nuijten et al., 4 Aug 2025).

6. Empirical Illustration and Model Interpretability

FEEF has been operationalized in interpretable agent models such as Free Energy Projective Simulation (FEPS), where agents construct explicit (e.g., HMM-based) world models, propagate preference signals for long-term goal reasoning, and resolve perceptual ambiguity through “belief superpositions.” Policy selection via FEEF naturally yields behaviors that flexibly adapt to both goal pursuit and uncertainty reduction—demonstrated in tasks such as grid navigation and timed response without requiring external rewards (Pazem et al., 22 Nov 2024). Comparative simulation (e.g., T-maze) establishes that FEEF/IFE minimization outperforms both pure curiosity-driven and pure utility-based agents across context-switching and ambiguous scenarios (Sajid et al., 2021).

7. Theoretical Unification and Applications

Recent formalizations have unified multiple popular decompositions of FEEF (risk/ambiguity; information gain/pragmatic value; risk-over-states/ambiguity; entropy/expected energy) as mathematically equivalent or related via bounds. The choice of root definition constrains the admissible forms of preference priors, with implications for practical expressivity in applications (Champion et al., 22 Feb 2024).

FEEF minimization underlies:

  • Active SLAM and robotic exploration (balancing mapping and targeted locomotion),
  • Curiosity-driven reinforcement learning and model-based RL (intrinsic rewards via information gain),
  • Neurocomputational theories (oculomotor search, dopaminergic signaling),
  • Affective modeling (mapping expected free energy components onto emotional dimensions such as arousal and valence via Shannon entropy and prediction error in circumplex emotion models) (Pattisapu et al., 2 Jul 2024).

The FEEF formalism thus constitutes a principled, unified framework for adaptive behavior in uncertain environments, grounding exploration and exploitation in Bayesian inference and information theory.

Whiteboard

Follow Topic

Get notified by email when new papers are published related to Free-Energy of the Expected Future (FEEF).