Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Decoupled Context Processing for Context Augmented Language Modeling (2210.05758v1)

Published 11 Oct 2022 in cs.CL, cs.AI, and cs.IR

Abstract: LLMs can be augmented with a context retriever to incorporate knowledge from large external databases. By leveraging retrieved context, the neural network does not have to memorize the massive amount of world knowledge within its internal parameters, leading to better parameter efficiency, interpretability and modularity. In this paper we examined a simple yet effective architecture for incorporating external context into LLMs based on decoupled Encoder Decoder architecture. We showed that such a simple architecture achieves competitive results on auto-regressive LLMing and open domain question answering tasks. We also analyzed the behavior of the proposed model which performs grounded context transfer. Finally we discussed the computational implications of such retrieval augmented models.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Zonglin Li (27 papers)
  2. Ruiqi Guo (18 papers)
  3. Sanjiv Kumar (123 papers)
Citations (21)

Summary

We haven't generated a summary for this paper yet.