Papers
Topics
Authors
Recent
Search
2000 character limit reached

Towards Reliable Retrieval in RAG Systems for Large Legal Datasets

Published 8 Oct 2025 in cs.CL and cs.IR | (2510.06999v1)

Abstract: Retrieval-Augmented Generation (RAG) is a promising approach to mitigate hallucinations in LLMs for legal applications, but its reliability is critically dependent on the accuracy of the retrieval step. This is particularly challenging in the legal domain, where large databases of structurally similar documents often cause retrieval systems to fail. In this paper, we address this challenge by first identifying and quantifying a critical failure mode we term Document-Level Retrieval Mismatch (DRM), where the retriever selects information from entirely incorrect source documents. To mitigate DRM, we investigate a simple and computationally efficient technique which we refer to as Summary-Augmented Chunking (SAC). This method enhances each text chunk with a document-level synthetic summary, thereby injecting crucial global context that would otherwise be lost during a standard chunking process. Our experiments on a diverse set of legal information retrieval tasks show that SAC greatly reduces DRM and, consequently, also improves text-level retrieval precision and recall. Interestingly, we find that a generic summarization strategy outperforms an approach that incorporates legal expert domain knowledge to target specific legal elements. Our work provides evidence that this practical, scalable, and easily integrable technique enhances the reliability of RAG systems when applied to large-scale legal document datasets.

Summary

  • The paper introduces SAC, a novel method that integrates document-level summaries into text chunks to mitigate retrieval mismatch.
  • It employs LLM-generated summaries and recursive chunking to enhance precision and recall across various legal documents.
  • Results show that generic summarization outperforms expert-guided strategies, offering a scalable improvement in RAG system reliability.

Introduction

The paper provides an in-depth look at improving retrieval reliability within Retrieval-Augmented Generation (RAG) systems, specifically when applied to large legal datasets. The paper identifies Document-Level Retrieval Mismatch (DRM) as a prevalent issue in standard RAG pipelines, where information is mistakenly sourced from incorrect documents. To address this, the authors propose Summary-Augmented Chunking (SAC) as a computationally efficient technique to infuse global context into retrieval stages. The approach integrates a synthetic, document-level summary into each chunk, thus enhancing the retriever's ability to select accurately from large, homogeneous legal databases.

Methodology

The methodology is centered around developing and deploying SAC within a legal dataset context. This involves four key steps: producing a document-level summary using LLMs, recursively character-splitting texts into chunks, prepending the summaries to these chunks, and indexing them for retrieval. Figure 1

Figure 1: Part a) illustrates how the retrieval quality metrics, DRM and text-level precision/recall, are computed in the LegalBench-RAG.

Figure 2

Figure 2

Figure 2: DRM of a standard RAG approach (left) and using SAC (right), applied to the four datasets in the LegalBench-RAG benchmark.

The system was evaluated on the LegalBench-RAG benchmark across a diverse array of legal documents including NDAs, privacy policies, and mergers. The evaluation metrics used were character-level precision/recall and the newly defined DRM, offering quantifiable insight into the retrieval performance.

Results

Using SAC resulted in a marked reduction in DRM across all tested datasets, accompanying improvements in text-level retrieval metrics. The SAC method demonstrated a superior ability to avoid cross-document confusion, a common problem in legal datasets rich with structurally similar documents. Figure 3

Figure 3

Figure 3: Text-level precision (left) and recall (right) of standard RAG and SAC with general or expert-guided summarization strategy.

However, interestingly, the paper found that generic summarization strategies outperformed expert-guided ones. Despite the structured density and specificity of expert-generated summaries, the broader, more semantically aligned summaries proved more effective for retrieval purposes in this context.

Discussion

The broad semantic cues of generic summarization outperform legally precise summaries within embeddings, raising questions about the interaction between chunk and summary information in semantic space. Insights suggest simple, generic context enrichment strategies are not only computationally efficient but also enhance retrieval effectiveness when dealing with legal datasets.

Future work could explore hierarchical context summarization or the adaptation of query optimization techniques to further refine retrieval accuracy. Additionally, examining the utility of post-retrieval reranking stages presents a promising avenue for improving chunk selection before generation, ultimately enhancing downstream generative tasks.

Conclusion

This research highlights a practical method to bolster the reliability of RAG systems in the legal domain, emphasizing the significant role of SAC in addressing Document-Level Retrieval Mismatch. While SAC is not exhaustive in resolving all retrieval challenges, it offers a scalable, easily integrable approach that meaningfully improves the retriever's performance in legally critical AI applications.

By advancing retrieval reliability, this research contributes to foundational work needed to establish AI as a dependable assistant in processing and navigating complex legal documents.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 1 tweet with 4 likes about this paper.