Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
167 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
42 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Active Inference as a Model of Agency (2401.12917v1)

Published 23 Jan 2024 in cs.AI

Abstract: Is there a canonical way to think of agency beyond reward maximisation? In this paper, we show that any type of behaviour complying with physically sound assumptions about how macroscopic biological agents interact with the world canonically integrates exploration and exploitation in the sense of minimising risk and ambiguity about states of the world. This description, known as active inference, refines the free energy principle, a popular descriptive framework for action and perception originating in neuroscience. Active inference provides a normative Bayesian framework to simulate and model agency that is widely used in behavioural neuroscience, reinforcement learning (RL) and robotics. The usefulness of active inference for RL is three-fold. \emph{a}) Active inference provides a principled solution to the exploration-exploitation dilemma that usefully simulates biological agency. \emph{b}) It provides an explainable recipe to simulate behaviour, whence behaviour follows as an explainable mixture of exploration and exploitation under a generative world model, and all differences in behaviour are explicit in differences in world model. \emph{c}) This framework is universal in the sense that it is theoretically possible to rewrite any RL algorithm conforming to the descriptive assumptions of active inference as an active inference algorithm. Thus, active inference can be used as a tool to uncover and compare the commitments and assumptions of more specific models of agency.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (84)
  1. A. Barto and R. Sutton. Reinforcement Learning: An Introduction. A Bradford Book, 1992.
  2. Jeremy Bentham. An Introduction to the Principles of Morals and Legislation. Dover Publications Inc., Mineola, N.Y, June 2007.
  3. J. Von Neumann and O. Morgenstern. Theory of Games and Economic Behavior. Princeton University Press, 1944.
  4. J. O. Berger. Statistical Decision Theory and Bayesian Analysis. Springer Series in Statistics. Springer-Verlag, New York, second edition, 1985.
  5. Richard E. Bellman. Dynamic Programming. Princeton University Press, Princeton, NJ, US, 1957.
  6. K. J Åström. Optimal control of Markov processes with incomplete state information. Journal of Mathematical Analysis and Applications, 10(1):174–205, February 1965.
  7. Jürgen Schmidhuber. A possibility for implementing curiosity and boredom in model-building neural controllers. In Proc. of the international conference on simulation of adaptive behavior: From animals to animats, pages 222–227, 1991.
  8. Curiosity-driven exploration by self-supervised prediction. In International conference on machine learning, pages 2778–2787. PMLR, 2017.
  9. Unifying count-based exploration and intrinsic motivation. Advances in neural information processing systems, 29, 2016.
  10. The free energy principle made simpler but not too simple. arXiv:2201.06387 [cond-mat, physics:nlin, physics:physics, q-bio], January 2022.
  11. K. Friston. The free-energy principle: A unified brain theory? Nature Reviews Neuroscience, 11(2):127–138, 2010.
  12. A free energy principle for the brain. J. Physiology-Paris, 100(1-3):70–87, 2006.
  13. The computational neurology of movement under active inference. Brain, March 2021.
  14. Hierarchical Active Inference: A Theory of Motivated Control. Trends in Cognitive Sciences, 22(4):294–306, April 2018.
  15. Computational mechanisms of curiosity and goal-directed exploration. eLife, page 45, 2019.
  16. A Bayesian computational model reveals a failure to adapt interoceptive precision estimates across depression, anxiety, eating, and substance use disorders. PLOS Computational Biology, 16(12):e1008484, December 2020.
  17. Canonical neural networks perform active inference. Communications Biology, 5(1):1–15, January 2022.
  18. Contrastive Active Inference. In Advances in Neural Information Processing Systems, May 2021.
  19. Deep active inference agents using Monte-Carlo methods. arXiv:2006.04176 [cs, q-bio, stat], June 2020.
  20. An empirical evaluation of active inference in multi-armed bandits. Neural Networks, 144:229–246, December 2021.
  21. Exploration and preference satisfaction trade-off in reward-free learning. arXiv preprint arXiv:2106.04316, 2021.
  22. Active Inference in Robotics and Artificial Agents: Survey and Challenges. arXiv:2112.01871 [cs], December 2021.
  23. How Active Inference Could Help Revolutionise Robotics. Entropy, 24(3):361, March 2022.
  24. Robot navigation as hierarchical active inference. Neural Networks, 142:192–204, October 2021.
  25. Cognitive and motor compliance in intentional human-robot interaction. In 2020 IEEE International Conference on Robotics and Automation (ICRA), pages 11291–11297, May 2020.
  26. D. V. Lindley. On a Measure of the Information Provided by an Experiment. The Annals of Mathematical Statistics, 27(4):986–1005, 1956.
  27. PID Control as a Process of Active Inference with Linear Generative Models. Entropy, 21(3):257, March 2019.
  28. Grigorios A. Pavliotis. Stochastic Processes and Applications: Diffusion Processes, the Fokker-Planck and Langevin Equations. Number volume 60 in Texts in Applied Mathematics. Springer, New York, 2014.
  29. Markov blankets, information geometry and stochastic thermodynamics. Philosophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences, 378(2164):20190159, February 2020.
  30. Luc Rey-Bellet. Open Classical Systems. In Stéphane Attal, Alain Joye, and Claude-Alain Pillet, editors, Open Quantum Systems II: The Markovian Approach, Lecture Notes in Mathematics, pages 41–78. Springer, Berlin, Heidelberg, 2006.
  31. The experience of agency: Feelings, judgments, and responsibility. Current Directions in Psychological Science, 18(4):242–246, 2009.
  32. Thomas Minka. Divergence measures and message passing. Technical report, 2005.
  33. Prospect Theory: An Analysis of Decision under Risk. Econometrica, 47(2):263–291, 1979.
  34. Sergey Levine. Reinforcement Learning and Control as Probabilistic Inference: Tutorial and Review. arXiv:1805.00909 [cs, stat], May 2018.
  35. On Stochastic Optimal Control and Reinforcement Learning by Approximate Inference. In Twenty-Third International Joint Conference on Artificial Intelligence, June 2013.
  36. Marc Toussaint. Robot trajectory optimization using approximate inference. In Proceedings of the 26th Annual International Conference on Machine Learning, ICML ’09, pages 1049–1056, Montreal, Quebec, Canada, June 2009. Association for Computing Machinery.
  37. On the Relationship Between Active Inference and Control as Inference. In Tim Verbelen, Pablo Lanillos, Christopher L. Buckley, and Cedric De Boom, editors, Active Inference, Communications in Computer and Information Science, pages 3–11, Cham, 2020. Springer International Publishing.
  38. R. E. Kalman. A New Approach to Linear Filtering and Prediction Problems. Journal of Basic Engineering, 82(1):35–45, March 1960.
  39. Emanuel Todorov. General duality between optimal control and estimation. In 2008 47th IEEE Conference on Decision and Control, pages 4286–4292, December 2008.
  40. Optimal control as a graphical model inference problem. Machine Learning, 87(2):159–182, May 2012.
  41. B. Ziebart. Modeling Purposeful Adaptive Behavior with the Principle of Maximum Causal Entropy. PhD thesis, Carnegie Mellon University, Pittsburgh, 2010.
  42. Abraham Kaplan. The Conduct of Inquiry. Transaction Publishers, 1973.
  43. John S. Bridle. Probabilistic Interpretation of Feedforward Classification Network Outputs, with Relationships to Statistical Pattern Recognition. In Françoise Fogelman Soulié and Jeanny Hérault, editors, Neurocomputing, NATO ASI Series, pages 227–236, Berlin, Heidelberg, 1990. Springer.
  44. R. Duncan Luce. Individual Choice Behavior. Individual Choice Behavior. John Wiley, Oxford, England, 1959.
  45. Jürgen Schmidhuber. Formal Theory of Creativity, Fun, and Intrinsic Motivation (1990–2010). IEEE Transactions on Autonomous Mental Development, 2(3):230–247, September 2010.
  46. David J. C. MacKay. Information-Based Objective Functions for Active Data Selection. Neural Computation, 4(4):590–604, July 1992.
  47. What is Intrinsic Motivation? A Typology of Computational Approaches. Frontiers in Neurorobotics, 1:6, November 2007.
  48. Novelty or Surprise? Frontiers in Psychology, 4, 2013.
  49. Planning to Be Surprised: Optimal Bayesian Exploration in Dynamic Environments. arXiv:1103.5708 [cs, stat], March 2011.
  50. Intrinsic Motivation and Self-Determination in Human Behavior. Perspectives in Social Psychology. Springer US, New York, 1985.
  51. Bayesian surprise attracts human attention. Vision research, 49(10):1295–1306, May 2009.
  52. Generative Models for Active Vision. Frontiers in Neurorobotics, 15, 2021.
  53. H. B. Barlow. Possible Principles Underlying the Transformations of Sensory Messages. The MIT Press, 1961.
  54. R Linsker. Perceptual Neural Organization: Some Approaches Based on Network Models and Information Theory. Annual Review of Neuroscience, 13(1):257–281, 1990.
  55. Temporal encoding of two-dimensional patterns by single units in primate inferior temporal cortex. III. Information theoretic analysis. Journal of Neurophysiology, 57(1):162–178, January 1987.
  56. The Exploration-Exploitation Dilemma: A Multidisciplinary Framework. PLOS ONE, 9(4):e95693, April 2014.
  57. Ready policy one: World building through active learning. In International Conference on Machine Learning, pages 591–601. PMLR, 2020.
  58. Varibad: A very good method for bayes-adaptive deep rl via meta-learning. arXiv preprint arXiv:1910.08348, 2019.
  59. Dream to control: Learning behaviors by latent imagination. arXiv preprint arXiv:1912.01603, 2019.
  60. Action and Perception as Divergence Minimization. arXiv:2009.01791 [cs, math, stat], October 2020.
  61. Exploration in approximate hyper-state space for meta reinforcement learning. In International Conference on Machine Learning, pages 12991–13001. PMLR, 2021.
  62. Synthesizing 3d shapes via modeling multi-view depth maps and silhouettes with deep generative networks. CVPR, 2017.
  63. Deep temporal models and active inference. Neuroscience & Biobehavioral Reviews, 90:486–501, July 2018.
  64. 3dp3: 3d scene perception via probabilistic programming. NeurIPS, 2021.
  65. Toward the neural implementation of structure learning. Current Opinion in Neurobiology, 37:99–105, April 2016.
  66. Building Machines That Learn and Think Like People. arXiv:1604.00289 [cs, stat], April 2016.
  67. Bayesian Models of Conceptual Development: Learning as Building Models of the World. Annual Review of Developmental Psychology, 2(1):533–558, 2020.
  68. Learning a theory of causality. Psychological Review, 118(1):110–119, 2011.
  69. Geometric Methods for Sampling, Optimisation, Inference and Adaptive Agents. In Geometry and Statistics, number 46 in Handbook of Statistics. Academic Press, 2022.
  70. Active inference on discrete state-spaces: A synthesis. J. Math. Psychology, 99:102447, 2020.
  71. Pymdp: A Python library for active inference in discrete state spaces. arXiv:2201.03904 [cs, q-bio], January 2022.
  72. A step-by-step tutorial on active inference and its application to empirical data. Journal of Mathematical Psychology, 107:102632, April 2022.
  73. Active Inference: The Free Energy Principle in Mind, Brain, and Behavior. MIT Press, Cambridge, MA, USA, March 2022.
  74. Active Inference, Belief Propagation, and the Bethe Approximation. Neural Computation, 30(9):2530–2567, September 2018.
  75. Branching Time Active Inference: Empirical study and complexity class analysis. arXiv:2111.11276 [cs], November 2021.
  76. Branching Time Active Inference: The theory and its generality. arXiv:2111.11107 [cs], November 2021.
  77. Active Tree Search in Large POMDPs. arXiv:2103.13860 [cs, math, q-bio], March 2021.
  78. Active Inference: A Process Theory. Neural Computation, 29(1):1–49, January 2017.
  79. Active Inference: Demystified and Compared. Neural Computation, 33(3):674–712, January 2021.
  80. Active Inference, Curiosity and Insight. Neural Computation, 29(10):2633–2683, October 2017.
  81. Thomas Parr. The Computational Neurology of Active Vision. PhD thesis, University College London, London, 2019.
  82. B. Millidge. Deep active inference as variational policy gradients. J. Math. Psychology, 96:102348, 2020.
  83. Path integrals, particular kinds and strange things. 2022.
  84. Active inference, Bayesian optimal design, and expected utility. arXiv:2110.04074 [cs, math, stat], September 2021.
Citations (4)

Summary

  • The paper introduces active inference as a Bayesian approach to model agency by unifying exploration and exploitation under a free energy minimization framework.
  • It demonstrates a shift from traditional reward-maximization paradigms to a principled formulation that mirrors biological decision-making.
  • The framework offers an explainable alternative for agent-based modeling and encourages integrating hierarchical probabilistic models with deep neural architectures.

Overview of Active Inference and Agency

In the paper titled "Active Inference as a Model of Agency," a comprehensive framework for understanding and modeling agency is presented. Agency, within this context, refers to the ability of an entity—be it biological or artificial—to make decisions and take actions that influence the environment. The authors propose active inference as a normative Bayesian approach that naturally integrates both exploratory and exploitative behaviors by minimizing risk and ambiguity concerning external states.

Agency Beyond Reward Maximization

Active inference is highlighted as an advance over traditional reinforcement learning (RL) models, which primarily focus on reward maximization. The authors argue that active inference, by minimizing expected free energy, can address the exploration-exploitation dilemma without resorting to ad hoc methods for inducing exploratory behavior. This paradigm shift is significant as it aligns with the natural behavior observed in biological systems, providing a more realistic and principled methodology for simulating intelligent behavior while offering explainability through explicit generative world models.

Theoretical Foundations in Physics

The foundation of active inference lies in well-established physical principles, as the paper elucidates by discussing the behavior of macroscopic biological systems subject to classical mechanics, known as precise agents. A precise agent responds deterministically to environmental stimuli given the information available up until the current moment. The mathematical formalization of this interaction delineates the nature of decision-making and agency as optimization problems, governed by an expected free energy functional—a criterion that balances objective probabilities of states with subjective preferences or utilities.

Advantages and Applications

Active inference presents several advantages over traditional models of decision-making in living organisms and artificial systems. Firstly, it emphasizes the value of understanding preferences and predictions in a unified framework, leading to robust exploration-exploration solutions. Secondly, the universality of the active inference framework allows RL algorithms to be recast as active inference algorithms, fostering comparisons and advancing theory development in agent-based modeling. Finally, by explicitly encoding predictions and preferences, active inference lays down a blueprint for transparent and justifiable algorithmic decision-making, proving highly applicable to safety-critical domains.

Path Forward

The authors conclude by reflecting on the potential of active inference and pose several open questions important for future research. One area of intrigue is identifying the generative models humans use to represent their environments and the computational mechanisms through which these representations evolve from childhood to adulthood. The integration of hierarchical probabilistic models and deep neural architectures is suggested as a promising path forward, with the goal of understanding and eventually replicating the development of human intelligence in artificial agents.