Dice Question Streamline Icon: https://streamlinehq.com

Systematic comparison of prompting-based vs. RL-based query augmentation

Determine the relative strengths, weaknesses, and applicability in real-world settings of prompting-based query augmentation methods (e.g., pseudo-document generation) versus reinforcement learning–based query rewriting formulations through a systematic, controlled comparison.

Information Square Streamline Icon: https://streamlinehq.com

Background

The paper highlights two dominant paradigms for improving queries in information retrieval: prompting-based pseudo-document/query expansion and reinforcement learning–based query rewriting. Although each has advantages and limitations, the literature has lacked a rigorous, apples-to-apples comparison to clarify when and why one approach outperforms the other.

Establishing a systematic comparison is important for practitioners who must choose between these approaches for diverse retrieval settings (sparse and dense retrievers) and applications, but prior work left this question unresolved, motivating the present paper.

References

More broadly, the field still lacks a systematic comparison between prompting-based query augmentation methods and RL-based formulations, leaving open questions about their respective strengths, weaknesses, and applicability in real-world settings.

Rethinking On-policy Optimization for Query Augmentation (2510.17139 - Xu et al., 20 Oct 2025) in Section 1, Introduction