Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Efficient Passage Retrieval with Hashing for Open-domain Question Answering (2106.00882v1)

Published 2 Jun 2021 in cs.CL and cs.IR
Efficient Passage Retrieval with Hashing for Open-domain Question Answering

Abstract: Most state-of-the-art open-domain question answering systems use a neural retrieval model to encode passages into continuous vectors and extract them from a knowledge source. However, such retrieval models often require large memory to run because of the massive size of their passage index. In this paper, we introduce Binary Passage Retriever (BPR), a memory-efficient neural retrieval model that integrates a learning-to-hash technique into the state-of-the-art Dense Passage Retriever (DPR) to represent the passage index using compact binary codes rather than continuous vectors. BPR is trained with a multi-task objective over two tasks: efficient candidate generation based on binary codes and accurate reranking based on continuous vectors. Compared with DPR, BPR substantially reduces the memory cost from 65GB to 2GB without a loss of accuracy on two standard open-domain question answering benchmarks: Natural Questions and TriviaQA. Our code and trained models are available at https://github.com/studio-ousia/bpr.

Efficient Passage Retrieval with Hashing for Open-domain Question Answering

The paper "Efficient Passage Retrieval with Hashing for Open-domain Question Answering" by Yamada et al. addresses a critical challenge in open-domain question answering (QA): the memory-intensive nature of passage retrieval models. By leveraging a novel Binary Passage Retriever (BPR), the authors propose an efficient method that significantly reduces memory requirements while maintaining retrieval accuracy.

Overview of the Approach

Many contemporary QA systems rely on a two-stage retriever-reader framework, notably exemplified by Dense Passage Retrievers (DPR). These models represent passages as continuous vector embeddings, requiring substantial memory to maintain extensive passage indices, such as those derived from large knowledge bases like Wikipedia.

BPR innovatively integrates a learning-to-hash mechanism with DPR to encode passage indices into compact binary codes instead of continuous vectors. This transition from continuous to binary representation enables a drastic reduction in memory usage—shrinking the storage necessity from 65GB to a mere 2GB—without compromising accuracy on established benchmarks like Natural Questions (NQ) and TriviaQA (TQA).

Methodological Insights

BPR employs a multi-task learning approach, simultaneously optimizing two objectives: efficient candidate generation using binary codes and accurate passage reranking using continuous vectors. Binary codes are produced by the application of a sign function to embeddings, with the function approximated by a scaled tanh operation to facilitate backpropagation during training.

The retrieval model operates in two stages:

  1. Candidate Generation: Uses Hamming distance computations on binary codes, ensuring rapid retrieval of candidate passages.
  2. Reranking: Employs inner product calculations on continuous embeddings to order candidate passages accurately.

Experimental Findings

BPR was validated on the NQ and TQA datasets. The results demonstrate its competitive retrieval efficacy—a top-20 recall comparable to full DPR systems—with a dramatically reduced memory footprint. Additionally, the incorporation of an enhanced reader enables BPR to match or exceed the QA accuracy of much larger models while utilizing fewer parameters.

In comparison to other memory-optimized models like those using Hierarchical Navigable Small World (HNSW) graphs or product quantization, BPR offers a balanced trade-off between storage efficiency and query performance.

Implications and Future Directions

The introduction of BPR not only has practical implications—facilitating less resource-intensive deployment of QA systems on large-scale knowledge sources—but also opens avenues for further exploration in retrieval efficiency. The outcomes promote future research in integrating advanced hashing methods and optimizing multi-task learning for retrieval tasks.

By providing a substantially compressed representation without loss of performance, BPR exemplifies an efficient retrieval model that meets contemporary demands for scalable QA systems. This work encourages further exploration into optimizing the alignment between retrieval and reader functionalities, potentially incorporating more sophisticated neural architectures and training regimes.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Ikuya Yamada (22 papers)
  2. Akari Asai (35 papers)
  3. Hannaneh Hajishirzi (176 papers)
Citations (75)
Youtube Logo Streamline Icon: https://streamlinehq.com