"Stochastic RAG: End-to-End Retrieval-Augmented Generation through Expected Utility Maximization" introduces Stochastic RAG, a groundbreaking approach for the optimization of retrieval-augmented generation (RAG) models. Traditional RAG models often depend on simplifying assumptions such as marginalization and document independence, which can limit their performance. This paper aims to address these limitations by presenting a novel formulation that frames the retrieval process as a stochastic sampling without replacement.
The authors use straight-through Gumbel-top-k, a differentiable approximation technique for sampling without replacement, enabling effective end-to-end optimization for retrieval-augmented generation. This method bypasses the need for marginalized distributions, leading to a more direct and integrated optimization process.
To validate the effectiveness of Stochastic RAG, the authors conducted extensive experiments across seven diverse datasets. These datasets cover a broad spectrum of tasks, including:
- Open-domain question answering
- Fact verification
- Slot-filling for relation extraction
- Dialogue systems
Through these experiments, Stochastic RAG demonstrated significant advancements in performance, achieving state-of-the-art results in six out of the seven datasets evaluated. This highlights not only the versatility of the approach across various applications but also its robustness in dealing with different types of language generation and retrieval tasks.
Overall, the paper makes notable contributions by relaxing restrictive assumptions and providing a robust and versatile framework for RAG model optimization. This innovation has the potential to significantly enhance the performance and applicability of RAG models across a wider range of tasks and datasets.