Papers
Topics
Authors
Recent
2000 character limit reached

Retrieval-Augmented Planning (RAP)

Updated 26 December 2025
  • Retrieval-Augmented Planning (RAP) is a framework that integrates external knowledge retrieval with sequential decision-making to enhance planning efficiency and accuracy.
  • It employs dynamic memory, context caching, and multi-step retrieval strategies to tailor tool selection and optimize execution in complex environments.
  • Empirical studies show that RAP improves task completion metrics, reduces latency, and adapts robustly to evolving toolsets across various application domains.

Retrieval-Augmented Planning (RAP) constitutes a paradigm at the intersection of information retrieval and sequential decision-making, specifically designed to enable planning systems—often LLM-based or multi-modal agents—to exploit retrieved, contextually relevant external knowledge or past experience. RAP addresses the foundational challenge of selecting, sequencing, and integrating retrieval operations within a planning procedure, combining the strengths of retrieval-augmented generation (RAG) with structured, often multi-step planning to yield robust, data-grounded action sequences in complex, dynamic environments.

1. Formal Foundations and Core Components

RAP extends traditional planning by introducing explicit retrieval into the loop. Given a user query or task (e.g., in a conversational assistant, robotics control, or multi-hop QA), a RAP agent must select not just which actions or tools to execute, but also which external sources to query—when, in what order, and with what arguments—such that the retrieved evidence is sufficient to drive high-quality downstream reasoning or action while minimizing cost and latency (Joshi et al., 26 Jul 2024).

A canonical RAP formulation is as follows:

  • Let Q\mathcal{Q} denote the space of possible queries or task states.
  • Let T={f1,…,fT}\mathcal{T} = \{f_1,\dots,f_T\} be a set of retrieval tools or APIs, each fif_i parameterized by its arguments.
  • A retrieval plan PP is a sequence [(fi1,ai1),...,(fiK,aiK)][(f_{i_1}, a_{i_1}), ..., (f_{i_K}, a_{i_K})] where arguments aija_{i_j} may depend on prior retrieval outputs.
  • The planner Ï€:(q,c)→P\pi: (q,c) \to P maps query and context to a plan, optimized for both evidence sufficiency and composite cost â„“(P)=∑jcost(fij)+λK\ell(P) = \sum_{j} \text{cost}(f_{i_j}) + \lambda K.

RAP architectures mandate an interplay of several technical components: (1) a planning agent, commonly an LLM or VLM with planning-centric fine-tuning and/or in-context learning; (2) a retrieval mechanism, indexed memory, or database of past demonstrations, trajectories, or facts; (3) an execution engine that realizes the sequential plan by invoking external tools, APIs, or environment actions; and (4) a dynamic memory or context cache to accumulate and summarize relevant prior information (Kagaya et al., 6 Feb 2024, Joshi et al., 26 Jul 2024, Soni et al., 5 Jun 2025).

2. Planning Algorithms and Workflow Variants

A diversity of RAP instantiations exists across domains. Major planning workflows include:

Single-Step RAP (e.g., REAPER): The planner LLM receives tool descriptions, prompt templates, and in-context examples; it outputs a complete, ordered retrieval plan in a single generation pass. The plan is parsed and executed, with intermediate outputs cascaded into successive tool calls (Joshi et al., 26 Jul 2024). This approach yields low-latency, scalable planning and is robust to changing tool APIs.

Multi-Step and Memory-Augmented RAP (e.g., RAP for Multimodal Agents, P-RAG, REAP): The agent alternates between generation and retrieval, dynamically incorporating retrieved past plans or episodic memory at each step. Memory construction can be offline (from expert demonstrations) or online (growing as the agent explores, as in progressive or continual RAP). The planning objective is typically to maximize the probability of task completion or answer correctness, conditioned on retrieved support (Kagaya et al., 6 Feb 2024, Xu et al., 17 Sep 2024, Zhu et al., 13 Nov 2025).

Dynamic Multi-Turn RAP: In dynamic domains (multi-turn dialogues, evolving toolsets), modules such as context caches, LoRA-adaptive retrievers, and context compressors enable RAP agents to maintain relevance, adapt tool selection, and avoid context explosion. The LLM maintains and updates an abstract syntax tree (AST) of tool calls, revising the plan as user intent or available tools change (Soni et al., 5 Jun 2025).

Progressive and Self-Iterative RAP: Progressive RAP pipelines (e.g., P-RAG) grow the retrieval database via self-iteration—each episode's successful trajectory is stored, and subsequent episodes benefit from both task-level and situation-level matching, leading to increasingly specialized and performant plans even in the absence of manual annotation (Xu et al., 17 Sep 2024).

Retrieval-Augmented Diffusion and Multi-Modal Planning: In domains such as motion planning and embodied reasoning, RAP is fused with generative diffusion models or vision-language planners. Here, task-relevant retrieval embeddings (trained jointly with the planner) guide the interpolation of context and action latent spaces during the generative process, dramatically enhancing robustness to rare or safety-critical scenarios (Ding et al., 30 May 2025, Guo et al., 22 Dec 2025).

3. Memory, Retrieval, and Data Structures

Central to RAP is the design of a retrieval corpus or agent memory:

  • Structure: Memories can be flat (logs of past trajectories), graph-based (temporal embodied knowledge graphs, hypergraphs), or database-oriented (API/tool call logs, demonstration banks).
  • Encoding: Textual, visual, or multimodal embeddings are used, often via sentence transformers or CLIP-based encoders (Kagaya et al., 6 Feb 2024, Guo et al., 22 Dec 2025).
  • Scoring and Retrieval: Task, plan, and step embeddings are combined via weighted cosine similarity; multimodal settings use fused representation and adaptive weighting per modality. Entity-weighted overlaps and attention mechanisms bias retrieval toward contextually and semantically relevant sub-trajectories or hyperedges (Zai et al., 14 Oct 2025).
  • Adaptivity: Dynamic memory curation (accumulation, compression, pruning) prevents obsolete or conflicting information from polluting plans. Temporal consistency mechanisms modulate retrieval certainty as underlying facts age (Yoo et al., 10 Sep 2025).

Two-level retrieval schemes (task-matching and situation-matching) surface highly pertinent references for embodied agents (Xu et al., 17 Sep 2024). In hybrid symbolic-neural setups, DAG-based decomposition organizes global reasoning structure and orchestrates retrieval at each hop (Zai et al., 14 Oct 2025, Zhu et al., 13 Nov 2025).

4. Application Domains and Empirical Results

RAP is applied across numerous, often heterogeneous domains:

  • Conversational Assistants: RAP plans retrieval operations across multiple APIs and indexes, achieving 96% tool F1 and 92% argument accuracy in real-world e-commerce dialogue, with a 10× latency reduction versus stepwise agent-based approaches (Joshi et al., 26 Jul 2024).
  • Embodied and Multimodal Agents: Progressive RAP yields >30 percentage point success increases on sequential tasks (e.g., ALFWorld, WebShop, Franka Kitchen, Meta-World) and more than doubles planning success rates in VLM-based embodied robotics when integrated with meta-action abstractions and retrieval-augmented in-context learning (Kagaya et al., 6 Feb 2024, Guo et al., 22 Dec 2025).
  • Multi-Hop Reasoning and QA: Structured decompositional RAP with explicit global state and recursive plan repair significantly outperforms standard RAG on HotpotQA, 2Wiki, MuSiQue, and Bamboogle, with gains exceeding 10 F1 on average (Zhu et al., 13 Nov 2025).
  • Travel and Spatiotemporal Planning: Evolutionary RAP (EvoRAG) combines reference trajectory retrieval with LLM-driven optimization to improve spatial efficiency (distance margin ratio cut by 35%), increase POI rationality, and reduce temporal constraint violations in personalized itineraries (Ni et al., 11 Apr 2025).
  • Driving and Safety-Critical Planning: Retrieval-augmented diffusion models for autonomous driving reduce collision rates by ∼40% on urban driving datasets, with largest gains in long-tail scenarios (Ding et al., 30 May 2025).
  • Continual and Non-Stationary Tasks: Exploratory RAP integrates memory, temporal consistency, and information-theoretic exploration bonuses, improving success rate in continual embodied instruction following by 14–18 points over stepwise LLM planners (Yoo et al., 10 Sep 2025).

5. Comparative Analysis and Design Patterns

RAP subsumes and advances over vanilla RAG (retrieval-augmented generation) by:

  • Treating retrieval not as a passive support mechanism but as a first-class planning variable, with explicit optimization over sources, ordering, and arguments.
  • Integrating memory and retrieval into multi-step, structured, and context-adaptive planning loops, rather than as static one-off augmentations.
  • Supporting progressive, continual learning and memory expansion, accommodating open-ended task spaces and evolving toolsets.
  • Enabling hybrid symbolic-neural reasoning where planning structure (e.g., DAG decomposition, action meta-abstraction) is separated from evidence retrieval, allowing for fine-grained audits and iterative plan refinement.

The table below contrasts representative RAP systems:

System Planning Structure Retrieval Corpus Application Domain
REAPER Single-shot LLM plan Tool/APIs, examples Conversational RAG, multi-API
RAP (2024) Looped LLM + memory Past episodes/blogs Text, multimodal, ALFWorld, robotics
P-RAG Progressive, iterative Grown episodic memory Embodied task learning
REAP Recursive, global state Web corpus, facts Multi-hop QA
PRoH Dynamic DAG/graph Knowledge hypergraph Knowledge-intensive QA
MaP-AVR In-context meta-actions Human-verified plans Embodied VLM planning (robotics)
RealDrive Joint-embedding+diffusion Driving scenario DB Autonomous driving

All information is verbatim or directly derived from the cited works above.

6. Current Limitations and Open Challenges

Despite marked progress, RAP systems face significant technical open questions:

  • Scalability of Memory and Retrieval: As episodic memory or knowledge graphs grow, retrieval speed and memory management become bottlenecks; vector quantization, hierarchical indexing, and continual compression are under-explored (Kagaya et al., 6 Feb 2024, Yoo et al., 10 Sep 2025).
  • Cost-Latency Optimization: Most published planners do not optimize explicit tradeoffs between retrieval latency, cost, and completeness beyond minimization of tool call count; formalizing cost models and integrating them with LLM-based planning is an open direction (Joshi et al., 26 Jul 2024).
  • Adaptivity and Robustness: Static weighting for retrieval or retrieval-only on successful episodes can bias the planner; meta-learned weighting, negative mining, and end-to-end retriever-policy tuning remain to be fully operationalized (Kagaya et al., 6 Feb 2024).
  • Temporal and Non-Stationary Environments: Maintaining temporally consistent, up-to-date environmental memories while avoiding re-exploration or knowledge staleness represents a non-trivial challenge, especially for continual tasks (Yoo et al., 10 Sep 2025).
  • Multi-Modal and Heterogeneous Data: Extending RAP to composite, multi-modal inputs (image, text, actions, sensor data) requires deeper abstraction over retrieval interfaces and plan representation (e.g., action-space, knowledge hypergraphs) (Guo et al., 22 Dec 2025, Zai et al., 14 Oct 2025).
  • Hallucination and Planning Failures: Spurious or hallucinated steps may arise, especially with ambiguous tool definitions or insufficient in-domain examples. Feedback loops for execution-time error correction are under-developed (Joshi et al., 26 Jul 2024).
  • Scalability to Real-World Applications: In highly dynamic domains (smart homes, healthcare, physical robotics), deploying RAP systems necessitates rapid adaptation to evolving APIs, tools, and noisy user intent (Soni et al., 5 Jun 2025).

7. Prospective Directions and Research Frontiers

Multiple research thrusts emerge from current RAP studies:

  • Memory Compression and Continual Learning: End-to-end learned graph summarization, continual memory revision, and federated or multi-agent RAP are active areas for scaling to lifelong learning scenarios (Yoo et al., 10 Sep 2025).
  • Meta-Reasoning and Self-Improvement: Incorporating explicit feedback from execution failures, meta-retrieval for negative cases, and adaptive exploration-exploitation tradeoff tuning.
  • Hybrid Symbolic-Neural Planning: Integration of constraint solvers (e.g., combinatorial optimization for travel), explicit subgoal decomposition, or hierarchical control, combined with neural retrieval/planning for high-level reasoning (Ni et al., 11 Apr 2025, Zai et al., 14 Oct 2025).
  • Cross-Domain Generalization and Data Efficiency: Demonstrated by LLM or VLM planners that adapt to novel APIs or domains via in-context tool definitions and minimal in-domain data (Joshi et al., 26 Jul 2024).
  • Real-World Deployment and Edge Adaptation: RAP components that are lightweight enough for deployment on resource-constrained or real-time edge devices, with parameter-efficient adaptation (e.g., LoRA retrieval, context compression) (Soni et al., 5 Jun 2025).

Retrieval-Augmented Planning thus constitutes a framework of growing centrality in the design of trustworthy, efficient, and extensible sequential decision-making systems, blending retrieval, memory, and stepwise or global planning across diverse domains and modalities. For detailed methodologies, empirical data, and further references, see (Joshi et al., 26 Jul 2024, Kagaya et al., 6 Feb 2024, Xu et al., 17 Sep 2024, Soni et al., 5 Jun 2025, Ni et al., 11 Apr 2025, Zai et al., 14 Oct 2025, Zhu et al., 13 Nov 2025, Guo et al., 22 Dec 2025, Ding et al., 30 May 2025), and (Yoo et al., 10 Sep 2025).

Whiteboard

Topic to Video (Beta)

Follow Topic

Get notified by email when new papers are published related to Retrieval-Augmented Planning (RAP).