Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Learning Dense Representations of Phrases at Scale (2012.12624v3)

Published 23 Dec 2020 in cs.CL
Learning Dense Representations of Phrases at Scale

Abstract: Open-domain question answering can be reformulated as a phrase retrieval problem, without the need for processing documents on-demand during inference (Seo et al., 2019). However, current phrase retrieval models heavily depend on sparse representations and still underperform retriever-reader approaches. In this work, we show for the first time that we can learn dense representations of phrases alone that achieve much stronger performance in open-domain QA. We present an effective method to learn phrase representations from the supervision of reading comprehension tasks, coupled with novel negative sampling methods. We also propose a query-side fine-tuning strategy, which can support transfer learning and reduce the discrepancy between training and inference. On five popular open-domain QA datasets, our model DensePhrases improves over previous phrase retrieval models by 15%-25% absolute accuracy and matches the performance of state-of-the-art retriever-reader models. Our model is easy to parallelize due to pure dense representations and processes more than 10 questions per second on CPUs. Finally, we directly use our pre-indexed dense phrase representations for two slot filling tasks, showing the promise of utilizing DensePhrases as a dense knowledge base for downstream tasks.

Learning Dense Representations of Phrases at Scale

This paper focuses on advancing open-domain question answering (QA) by enhancing phrase retrieval models through the innovation of dense representations. Historically, state-of-the-art QA systems have relied heavily on retriever-reader models which demand on-the-fly document processing. These have shown efficacy but come with high computational costs and complexity. The proposed model, DensePhrases, circumvents these challenges by leveraging purely dense phrase representations and eliminating the need for sparse components traditionally used in similar tasks.

Key Contributions

  1. Dense Representation Learning: The research introduces a method to accurately learn phrase representations using dense vectors, supervised by reading comprehension tasks. By employing data augmentation and knowledge distillation, alongside innovative negative sampling strategies, the model refines its phrase representation capacity. These strategies include in-batch and pre-batch negative sampling which significantly improve discrimination among phrases across a large dataset.
  2. Query-Side Fine-Tuning: This technique addresses the discrepancy between the training and inference phases. It allows robust transfer learning by updating only the question encoder without necessitating re-indexing of phrase representations. This mechanism markedly enhances the model's adaptability across different QA datasets and tasks.
  3. Performance and Efficiency: Evaluated across five standard open-domain QA datasets, the dense representation model demonstrates substantial performance gains of 15%–25% in absolute accuracy over previous models. DensePhrases also achieves accuracy on par with top retriever-reader systems. Moreover, due to its dense nature, the model is inherently more parallelizable, thus improving throughput significantly to over 10 questions per second on CPUs. The reduced storage footprint from 1.5TB to 320GB further illustrates the model's efficiency.
  4. Utilization as a Dense Knowledge Base: Beyond QA, the model's potential extends to other knowledge-intensive tasks, such as slot filling, projecting it as a neural interface for phrase-level knowledge retrieval. In practical terms, this suggests an avenue for DensePhrases to be employed in different NLP contexts without necessitating extensive retraining or restructuring of the knowledge base.

Implications and Future Directions

The introduction of DensePhrases marks a pivotal shift towards fully dense systems which promise scalability and efficiency in open-domain QA landscapes. Its implications for offline indexing and real-time response could transform the accessibility and efficiency of QA systems. Furthermore, the advancement in dense retrieval can aid other domains such as fact extraction and relation classification, enabling seamless, rapid integration of knowledge across diverse NLP tasks.

Looking forward, the ability to leverage simple, dense-only architectures presents fascinating avenues for research progression, including but not limited to more efficient data indexing, refined retrieval pipelines, and broader applications across unexplored datasets and languages. Future efforts could also focus on reducing computational overheads and enhancing the natural language understanding capabilities within more compact models.

In summation, the research provides a comprehensive and practical framework for dense phrase representation and retrieval, emphasizing an intersection of efficiency and accuracy that could set a precedent for future QA systems and related NLP endeavors.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Jinhyuk Lee (27 papers)
  2. Mujeen Sung (20 papers)
  3. Jaewoo Kang (83 papers)
  4. Danqi Chen (84 papers)
Citations (111)
Youtube Logo Streamline Icon: https://streamlinehq.com