Reframing the Expected Free Energy: Four Formulations and a Unification (2402.14460v1)
Abstract: Active inference is a leading theory of perception, learning and decision making, which can be applied to neuroscience, robotics, psychology, and machine learning. Active inference is based on the expected free energy, which is mostly justified by the intuitive plausibility of its formulations, e.g., the risk plus ambiguity and information gain / pragmatic value formulations. This paper seek to formalize the problem of deriving these formulations from a single root expected free energy definition, i.e., the unification problem. Then, we study two settings, each one having its own root expected free energy definition. In the first setting, no justification for the expected free energy has been proposed to date, but all the formulations can be recovered from it. However, in this setting, the agent cannot have arbitrary prior preferences over observations. Indeed, only a limited class of prior preferences over observations is compatible with the likelihood mapping of the generative model. In the second setting, a justification of the root expected free energy definition is known, but this setting only accounts for two formulations, i.e., the risk over states plus ambiguity and entropy plus expected energy formulations.
- Learning perception and planning with deep active inference. In 2020 IEEE International Conference on Acoustics, Speech and Signal Processing, ICASSP 2020, Barcelona, Spain, May 4-8, 2020, pages 3952–3956. IEEE, 2020. doi: 10.1109/ICASSP40776.2020.9054364. URL https://doi.org/10.1109/ICASSP40776.2020.9054364.
- Branching time active inference: Empirical study and complexity class analysis. Neural Networks, 152:450–466, 2022a. ISSN 0893-6080. doi: https://doi.org/10.1016/j.neunet.2022.05.010. URL https://www.sciencedirect.com/science/article/pii/S0893608022001824.
- Branching time active inference: The theory and its generality. Neural Networks, 151:295–316, 2022b. ISSN 0893-6080. doi: https://doi.org/10.1016/j.neunet.2022.03.036. URL https://www.sciencedirect.com/science/article/pii/S0893608022001149.
- Multi-modal and multi-factor branching time active inference, 2022c. URL https://arxiv.org/abs/2206.12503.
- Branching Time Active Inference with Bayesian Filtering. Neural Computation, 34(10):2132–2144, 09 2022d. ISSN 0899-7667. doi: 10.1162/neco˙a˙01529. URL https://doi.org/10.1162/neco_a_01529.
- Deconstructing deep active inference, 2023. URL https://arxiv.org/abs/2303.01618.
- Active inference in openai gym: A paradigm for computational investigations into psychiatric illness. Biological Psychiatry: Cognitive Neuroscience and Neuroimaging, 3(9):809 – 818, 2018. ISSN 2451-9022. doi: https://doi.org/10.1016/j.bpsc.2018.06.010. URL http://www.sciencedirect.com/science/article/pii/S2451902218301617. Computational Methods and Modeling in Psychiatry.
- Dopamine, reward learning, and active inference. Frontiers in Computational Neuroscience, 9:136, 2015. ISSN 1662-5188. doi: 10.3389/fncom.2015.00136. URL https://www.frontiersin.org/article/10.3389/fncom.2015.00136.
- Deep active inference agents using Monte-Carlo methods. In Hugo Larochelle, Marc’Aurelio Ranzato, Raia Hadsell, Maria-Florina Balcan, and Hsuan-Tien Lin, editors, Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, NeurIPS 2020, December 6-12, 2020, virtual, 2020. URL https://proceedings.neurips.cc/paper/2020/hash/865dfbde8a344b44095495f3591f7407-Abstract.html.
- A tutorial on variational Bayesian inference. Artificial Intelligence Review, 38(2):85–95, Aug 2012. ISSN 1573-7462. doi: 10.1007/s10462-011-9236-8. URL https://doi.org/10.1007/s10462-011-9236-8.
- Active inference and learning. Neuroscience & Biobehavioral Reviews, 68:862 – 879, 2016. ISSN 0149-7634. doi: https://doi.org/10.1016/j.neubiorev.2016.06.022.
- Sophisticated Inference. Neural Computation, 33(3):713–763, 03 2021. ISSN 0899-7667. doi: 10.1162/neco˙a˙01351. URL https://doi.org/10.1162/neco_a_01351.
- The graphical brain: Belief propagation and active inference. Network Neuroscience, 1(4):381–414, 12 2017. ISSN 2472-1751. doi: 10.1162/NETN˙a˙00018. URL https://doi.org/10.1162/NETN_a_00018.
- Bayesian surprise attracts human attention. Vision Research, 49(10):1295 – 1306, 2009. ISSN 0042-6989. doi: https://doi.org/10.1016/j.visres.2008.09.007. URL http://www.sciencedirect.com/science/article/pii/S0042698908004380. Visual Attention: Psychophysics, electrophysiology and neuroimaging.
- Edwin T Jaynes. Information theory and statistical mechanics. Physical review, 106(4):620, 1957a.
- Edwin T Jaynes. Information theory and statistical mechanics. ii. Physical review, 108(2):171, 1957b.
- Probabilistic graphical models: principles and techniques. MIT press, 2009.
- Factor graphs and the sum-product algorithm. IEEE Transactions on information theory, 47(2):498–519, 2001.
- Beren Millidge. Combining active inference and hierarchical predictive coding: A tutorial introduction and case study., 2019. URL https://doi.org/10.31234/osf.io/kf6wc.
- Neuronal message passing using mean-field, bethe, and marginal approximations. Scientific reports, 9(1):1889, 2019.
- Active inference: the free energy principle in mind, brain, and behavior. MIT Press, 2022.
- End-to-end pixel-based deep active inference for body perception and action. In Joint IEEE 10th International Conference on Development and Learning and Epigenetic Robotics, ICDL-EpiRob 2020, Valparaiso, Chile, October 26-30, 2020, pages 1–8. IEEE, 2020. doi: 10.1109/ICDL-EpiRob48136.2020.9278105. URL https://doi.org/10.1109/ICDL-EpiRob48136.2020.9278105.
- Computational mechanisms of curiosity and goal-directed exploration. bioRxiv, 2018. doi: 10.1101/411272. URL https://www.biorxiv.org/content/early/2018/09/07/411272.
- Variational message passing. Journal of Machine Learning Research, 6(4), 2005.