Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 167 tok/s
Gemini 2.5 Pro 53 tok/s Pro
GPT-5 Medium 31 tok/s Pro
GPT-5 High 31 tok/s Pro
GPT-4o 106 tok/s Pro
Kimi K2 187 tok/s Pro
GPT OSS 120B 443 tok/s Pro
Claude Sonnet 4.5 37 tok/s Pro
2000 character limit reached

Dyna-Mind: Learning to Simulate from Experience for Better AI Agents (2510.09577v1)

Published 10 Oct 2025 in cs.CL, cs.AI, and cs.CV

Abstract: Reasoning models have recently shown remarkable progress in domains such as math and coding. However, their expert-level abilities in math and coding contrast sharply with their performance in long-horizon, interactive tasks such as web navigation and computer/phone-use. Inspired by literature on human cognition, we argue that current AI agents need ''vicarious trial and error'' - the capacity to mentally simulate alternative futures before acting - in order to enhance their understanding and performance in complex interactive environments. We introduce Dyna-Mind, a two-stage training framework that explicitly teaches (V)LM agents to integrate such simulation into their reasoning. In stage 1, we introduce Reasoning with Simulations (ReSim), which trains the agent to generate structured reasoning traces from expanded search trees built from real experience gathered through environment interactions. ReSim thus grounds the agent's reasoning in faithful world dynamics and equips it with the ability to anticipate future states in its reasoning. In stage 2, we propose Dyna-GRPO, an online reinforcement learning method to further strengthen the agent's simulation and decision-making ability by using both outcome rewards and intermediate states as feedback from real rollouts. Experiments on two synthetic benchmarks (Sokoban and ALFWorld) and one realistic benchmark (AndroidWorld) demonstrate that (1) ReSim effectively infuses simulation ability into AI agents, and (2) Dyna-GRPO leverages outcome and interaction-level signals to learn better policies for long-horizon, planning-intensive tasks. Together, these results highlight the central role of simulation in enabling AI agents to reason, plan, and act more effectively in the ever more challenging environments.

Summary

  • The paper introduces a novel two-stage framework combining ReSim and Dyna-GRPO to enhance AI agents' simulation-derived reasoning and decision-making.
  • ReSim constructs structured search trees from actual interactions, enabling more accurate predictions and improved task planning in complex environments.
  • Dyna-GRPO iteratively refines policies using real rollouts and reward feedback, leading to superior performance in long-horizon and planning-intensive tasks.

Dyna-Mind: Learning to Simulate from Experience for Better AI Agents

Introduction

"Dyna-Mind: Learning to Simulate from Experience for Better AI Agents" proposes an innovative framework aimed at enhancing the reasoning and decision-making capabilities of AI agents in complex environments. The framework is motivated by insights from human cognition that emphasize the importance of mental simulations for effective planning and action selection. The paper highlights the limitations of current AI models in long-horizon tasks and introduces a two-stage training framework that explicitly incorporates simulation into the reasoning processes of AI agents.

Dyna-Mind Framework

The core contribution of Dyna-Mind lies in its two-stage training methodology:

  1. Reasoning with Simulations (ReSim): In the first stage, ReSim employs a structured approach to train agents using expanded search trees built from real experiences. This method allows the AI to ground its reasoning in actual world dynamics and anticipate future states. The simulation traces generated through ReSim leverage structured reasoning from these search trees, thus enabling better task planning and execution.
  2. Dyna-GRPO: In the second stage, the paper introduces Dyna-GRPO, an advanced reinforcement learning (RL) technique. This approach strengthens the agent's simulation and decision-making abilities by utilizing both outcome rewards and intermediate states as feedback during real rollouts. Dyna-GRPO iteratively improves policy and world model simulations, optimizing the agent's performance in planning-intensive tasks.

The integration of ReSim and Dyna-GRPO establishes a comprehensive framework for improving simulation-derived reasoning in AI agents, addressing current deficiencies in complex task environments.

Empirical Evaluation

The framework was evaluated using two synthetic benchmarks, Sokoban and ALFWorld, as well as the realistic AndroidWorld benchmark. Key findings include:

  • Enhanced Simulation Ability: ReSim significantly improved the simulation abilities of AI agents, as demonstrated by superior task completion rates in both Sokoban and ALFWorld. By directly learning from real-world interactions, the agent's reasoning took into account more accurate representations of world dynamics.
  • Improvement through Dyna-GRPO: The addition of Dyna-GRPO further bolstered the agent's competencies, particularly in long-horizon and planning-intensive tasks. The iterative policy refinement through real and simulated rollouts proved effective in yielding better long-term task execution strategies. Figure 1

    Figure 1: ReSim integrates simulation into reasoning by using expanded search trees built through real environment interactions.

    Figure 2

    Figure 2: Dyna-GRPO iterates between policy improvement and world model improvement, optimized by GRPO.

Implications and Future Work

The results underline the crucial role of simulation in AI reasoning and planning. By embedding simulation within the framework of AI agents, there is a marked improvement in their ability to navigate and perform complex tasks that require strategic foresight and adaptability.

Advancing toward more interactive and adaptable AI agents, future work could explore enhancing the framework's scalability to broader domains and refining the integration of diverse environmental dynamics. Furthermore, extending this framework to incorporate additional sensory inputs and interaction modalities could further enhance agent performance in multifaceted applications.

Conclusion

Dyna-Mind represents a significant step forward in aligning AI agent capabilities with the intricacies of real-world environments. By leveraging robust simulation strategies within a structured training methodology, the framework offers a comprehensive approach to enhancing AI reasoning and planning capabilities, paving the way for the next generation of autonomous agents.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets

This paper has been mentioned in 6 tweets and received 10 likes.

Upgrade to Pro to view all of the tweets about this paper: