Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Infinite Ends from Finite Samples: Open-Ended Goal Inference as Top-Down Bayesian Filtering of Bottom-Up Proposals (2407.16770v1)

Published 23 Jul 2024 in cs.AI

Abstract: The space of human goals is tremendously vast; and yet, from just a few moments of watching a scene or reading a story, we seem to spontaneously infer a range of plausible motivations for the people and characters involved. What explains this remarkable capacity for intuiting other agents' goals, despite the infinitude of ends they might pursue? And how does this cohere with our understanding of other people as approximately rational agents? In this paper, we introduce a sequential Monte Carlo model of open-ended goal inference, which combines top-down Bayesian inverse planning with bottom-up sampling based on the statistics of co-occurring subgoals. By proposing goal hypotheses related to the subgoals achieved by an agent, our model rapidly generates plausible goals without exhaustive search, then filters out goals that would be irrational given the actions taken so far. We validate this model in a goal inference task called Block Words, where participants try to guess the word that someone is stacking out of lettered blocks. In comparison to both heuristic bottom-up guessing and exact Bayesian inference over hundreds of goals, our model better predicts the mean, variance, efficiency, and resource rationality of human goal inferences, achieving similar accuracy to the exact model at a fraction of the cognitive cost, while also explaining garden-path effects that arise from misleading bottom-up cues. Our experiments thus highlight the importance of uniting top-down and bottom-up models for explaining the speed, accuracy, and generality of human theory-of-mind.

Open-Ended Goal Inference via Top-Down Bayesian Filtering and Bottom-Up Proposals

In the paper "Infinite Ends from Finite Samples: Open-Ended Goal Inference as Top-Down Bayesian Filtering of Bottom-Up Proposals," the authors Tan Zhi-Xuan et al. propose a computational model to address the problem of inferring human goals in an open-ended setting. Existing models largely operate under the assumption of a fixed and finite set of possible goals, yet human goal inference capabilities seem to significantly transcend these limitations. This paper advances the theoretical and algorithmic understanding of human goal inference in open-ended and complex scenarios.

The proposed model utilizes a Sequential Monte Carlo (SMC) approach to combine top-down Bayesian inverse planning with bottom-up heuristic sampling. This approach enables the model to quickly generate plausible goal hypotheses and subsequently filter them through a rationality framework. This bifurcated approach harmonizes the efficiency and flexibility of human goal inference capabilities.

Methodology

The authors introduce their model in a rigorous computational framework. The core elements include:

  • Goal Prior: gP(g)g \sim P(g)
  • Online Planning: πtP(πtst1,πt1,g)\pi_t \sim P(\pi_t | s_{t-1}, \pi_{t-1}, g)
  • Action Selection: atP(atst1,πt)a_t \sim P(a_t | s_{t-1}, \pi_t)
  • State Transition: stP(stst1,at)s_t \sim P(s_t | s_{t-1}, a_t)

These elements together form a generative model to perform Bayesian inverse planning. Given the computational challenges posed by large goal spaces, typical of open-ended settings, the authors propose a novel integration. They leverage the familiarity of humans with the statistics of their environment to hypothesize goals rapidly via bottom-up proposals.

The bottom-up sampling relies on the hypothesis that humans, based on environmental cues and knowledge of common subgoals, generate and test goal hypotheses. Upon hypothesizing a goal, the model utilizes goal filtering based on top-down rationality constraints, forming a two-stage hypothesis: generate plausibly and then evaluate rigorously.

Experimental Setup

To validate their model, the authors employed the "Block Words" task, where participants observed sequences of actions made by an agent stacking lettered blocks to form a word. Participants then inferred potential goal words at several judgment points within the sequence. The experimental conditions were systematically varied to test different types of inference challenges, such as bottom-up friendly scenarios, irrational alternatives, garden paths, and uncommon words.

Results

The open-ended SMC model demonstrated superior performance in approximating human inferences, achieving higher similarity to human goal distributions, particularly in the intricate scenarios like irrational alternatives. These results substantiate the value of combining bottom-up sampling for hypothesis generation with top-down Bayesian filtering for hypothesis evaluation.

Specifically, the empirical results indicated:

  • Accuracy: The open-ended SMC model showed comparable accuracy to exact Bayesian inference in simpler tasks but also outperformed pure bottom-up heuristics in complex scenarios.
  • Human Similarity: The model displayed higher intersection over union (IoU) with human participants' inferences across all conditions, particularly excelling in scenarios requiring rational evaluation of goals.
  • Algorithmic Properties: The model replicated the variance and sample efficiency observed in human participants, suggesting a plausible cognitive mechanism.
  • Resource Rationality: The open-ended model balanced computational efficiency with high accuracy, supported by sensitivity analysis showing robust performance under different parameter configurations.

Implications and Future Work

The paper has significant theoretical and practical implications. Theoretically, it bridges the gap between fixed-goal and open-ended goal inference models, demonstrating that human-like goal inference can be both computationally tractable and behaviorally accurate. Practically, this model has potential applications in areas requiring human-AI interaction and collaborative systems, enhancing the ability of AI to anticipate and align with human goals dynamically.

Future research could further refine the model's bottom-up proposal mechanism using more advanced LLMs, and explore adaptive algorithms that better mimic human cognitive processes like goal rejuvenation and selective forgetting. Integrating these enhancements could yield more robust and scalable open-ended goal inference systems.

Conclusion

The presented model represents a significant advancement in understanding the computational mechanisms underpinning human goal inference in open-ended scenarios. By effectively combining the strengths of bottom-up sampling and top-down Bayesian filtering, the model not only aligns closely with human reasoning patterns but also offers practical utility in complex real-world tasks. The findings underscore the potential for further developments in AI systems that can better understand and predict human goals in dynamic and open-ended environments.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (52)
  1. (2021). Modeling the mistakes of boundedly rational agents within a Bayesian theory of mind. In Proceedings of the Annual Meeting of the Cognitive Science Society (Vol. 43).
  2. (2009). Action understanding as inverse planning. Cognition, 113(3), 329–349.
  3. (1995). Learning to act using real-time dynamic programming. Artificial Intelligence, 72(1-2), 81–138.
  4. Beale, A.  (2016). 12dicts package. Retrieved from http://wordlist.aspell.net/12dicts-readme/
  5. (2021). Sequential core-set Monte Carlo. In Uncertainty in artificial intelligence (pp. 2165–2175).
  6. (2013). A computational-level explanation of the speed of goal inference. Journal of Mathematical Psychology, 57(3-4), 117–133.
  7. (1997). Fast planning through planning graph analysis. Artificial Intelligence, 90(1-2), 281–300.
  8. (2023). Inferring the future by imagining the past. In Thirty-seventh Conference on Neural Information Processing Systems.
  9. Chopin, N.  (2002). A sequential particle filter method for static models. Biometrika, 89(3), 539–552.
  10. (2019). Gen: A general-purpose probabilistic programming system with programmable inference. In Proceedings of the 40th ACM SIGPLAN Conference on Programming Language Design and Implementation (pp. 221–236).
  11. (2008). The pigeon as particle filter. Advances in Neural Information Processing Systems, 20, 369–376.
  12. (2006). Sequential Monte Carlo samplers. Journal of the Royal Statistical Society Series B: Statistical Methodology, 68(3), 411–436.
  13. (2012). On adaptive resampling strategies for sequential Monte Carlo methods. Bernoulli, 18(1), 252–278.
  14. (2016). Adapting the number of particles in sequential Monte Carlo methods through an online scheme for convergence assessment. IEEE Transactions on Signal Processing, 65(7), 1781–1794.
  15. (2003). Teleological reasoning in infancy: The naıve theory of rational action. Trends in cognitive sciences, 7(7), 287–292.
  16. (1994). Constructing inferences during narrative text comprehension. Psychological review, 101(3), 371.
  17. (2006). Optimal predictions in everyday cognition. Psychological science, 17(9), 767–773.
  18. (2010). Storing the web in memory: Space efficient language models with constant time retrieval. In Proceedings of the 2010 conference on Empirical Methods in Natural Language Processing (pp. 262–272).
  19. (2005). Automatic goal inferences. Journal of Experimental Social Psychology, 41(2), 129–140.
  20. (1944). An experimental study of apparent behavior. The American Journal of Psychology, 57(2), 243–259.
  21. (2007). Improving lrta*(k). In Proceedings of the 20th International Joint Conference on Artifical Intelligence (pp. 2312–2317).
  22. (2015). Children’s understanding of the costs and rewards underlying rational action. Cognition, 140, 14–23.
  23. (2019). The naive utility calculus as a unified, quantitative framework for action understanding. PsyArXiv.
  24. (2016). Coordinate to cooperate or compete: Abstract goals and joint intentions in social interaction. In Proceedings of the Annual Meeting of the Cognitive Science Society.
  25. (2006). Real-Time Adaptive A*. In Proceedings of the Fifth International Joint Conference on Autonomous Agents and Multiagent Systems (pp. 281–288).
  26. (2009). Comparing real-time and incremental heuristic search for real-time situated agents. Autonomous Agents and Multi-Agent Systems, 18, 313–341.
  27. Korf, R. E.  (1990). Real-time heuristic search. Artificial Intelligence, 42(2-3), 189–211.
  28. (2013). Bridging the gap between theory and practice of approximate bayesian inference. Cognitive Systems Research, 24, 2–8.
  29. (2022). Recursive monte carlo and variational inference with auxiliary variables. In Uncertainty in artificial intelligence (pp. 1096–1106).
  30. (2023). Probabilistic programming with stochastic probabilities. Proceedings of the ACM on Programming Languages, 7(PLDI), 1708–1732.
  31. (2023). SMCP3: Sequential Monte Carlo with probabilistic program proposals. In International Conference on Artificial Intelligence and Statistics (pp. 7061–7088).
  32. (2020). Resource-rational analysis: Understanding human cognition as the optimal use of limited computational resources. Behavioral and Brain Sciences, 43, e1.
  33. (2017). Ten-month-old infants infer the value of goals from the costs of actions. Science, 358(6366), 1038–1041.
  34. (1998). PDDL - the Planning Domain Definition Language (Tech. Rep.). Yale Center for Computational Vision and Control.
  35. (2021). Generating options and choosing between them depend on distinct forms of value representation. Psychological science, 32(11), 1731–1746.
  36. (2016). Spontaneous goal inference (sgi). Social and Personality Psychology Compass, 10(1), 64–80.
  37. (2019). How we know what not to think. Trends in cognitive sciences, 23(12), 1026–1040.
  38. (2009). Plan recognition as planning. In Twenty-first international joint conference on artificial intelligence.
  39. (2010). Probabilistic plan recognition using off-the-shelf classical planners. In Proceedings of the AAAI conference on artificial intelligence (Vol. 24).
  40. Schulz, L.  (2012). Finding new facts; thinking new thoughts. Advances in child development and behavior, 43, 269–294.
  41. Speer, R.  (2016). wordfreq: a Python library for looking up the frequencies of words in many languages. Retrieved from https://github.com/rspeer/wordfreq/
  42. (2017). Online learning of symbolic concepts. Journal of Mathematical Psychology, 77, 10–20.
  43. (2009). Help or hinder: Bayesian models of social goal inference. Advances in Neural Information Processing Systems, 22.
  44. (2012). Spontaneous goal inferences are often inferred faster than spontaneous trait inferences. Journal of Experimental Social Psychology, 48(1), 13–18.
  45. (2016). Online goal recognition through mirroring: Humans and agents. In Annual conference on advances in cognitive systems 2016.
  46. (2009). Explaining human multiple object tracking as resource-constrained approximate inference in a dynamic probabilistic model. Advances in Neural Information Processing Systems, 22.
  47. (2022). Symbolic Knowledge Distillation: from general language models to commonsense models. In Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (pp. 4602–4625). Seattle, United States: Association for Computational Linguistics.
  48. (2023). Inferring the goals of communicating agents from actions and instructions. ICML 2023 Workshop on Theory of Mind in Communicating Agents.
  49. Zhi-Xuan, T.  (2020). GenParticleFilters.jl. Retrieved from https://github.com/probcomp/GenParticleFilters.jl
  50. Zhi-Xuan, T.  (2022). PDDL.jl: An extensible interpreter and compiler interface for fast and flexible AI planning. Unpublished master’s thesis, MIT.
  51. (2020). Online Bayesian goal inference for boundedly rational planning agents. Advances in Neural Information Processing Systems, 33.
  52. (2024). Pragmatic instruction following and goal assistance via cooperative language guided inverse plan search. In Proceedings of the 23rd International Conference on Autonomous Agents and Multi-Agent Systems.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Tan Zhi-Xuan (22 papers)
  2. Gloria Kang (1 paper)
  3. Vikash Mansinghka (31 papers)
  4. Joshua B. Tenenbaum (257 papers)
Citations (2)
X Twitter Logo Streamline Icon: https://streamlinehq.com