Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
60 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
8 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

LM-CORE: Language Models with Contextually Relevant External Knowledge (2208.06458v1)

Published 12 Aug 2022 in cs.CL and cs.LG

Abstract: Large transformer-based pre-trained LLMs have achieved impressive performance on a variety of knowledge-intensive tasks and can capture factual knowledge in their parameters. We argue that storing large amounts of knowledge in the model parameters is sub-optimal given the ever-growing amounts of knowledge and resource requirements. We posit that a more efficient alternative is to provide explicit access to contextually relevant structured knowledge to the model and train it to use that knowledge. We present LM-CORE -- a general framework to achieve this -- that allows \textit{decoupling} of the LLM training from the external knowledge source and allows the latter to be updated without affecting the already trained model. Experimental results show that LM-CORE, having access to external knowledge, achieves significant and robust outperformance over state-of-the-art knowledge-enhanced LLMs on knowledge probing tasks; can effectively handle knowledge updates; and performs well on two downstream tasks. We also present a thorough error analysis highlighting the successes and failures of LM-CORE.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Jivat Neet Kaur (7 papers)
  2. Sumit Bhatia (30 papers)
  3. Milan Aggarwal (17 papers)
  4. Rachit Bansal (9 papers)
  5. Balaji Krishnamurthy (68 papers)
Citations (13)