Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Time-Aware Language Models as Temporal Knowledge Bases (2106.15110v2)

Published 29 Jun 2021 in cs.CL

Abstract: Many facts come with an expiration date, from the name of the President to the basketball team Lebron James plays for. But LLMs (LMs) are trained on snapshots of data collected at a specific moment in time, and this can limit their utility, especially in the closed-book setting where the pretraining corpus must contain the facts the model should memorize. We introduce a diagnostic dataset aimed at probing LMs for factual knowledge that changes over time and highlight problems with LMs at either end of the spectrum -- those trained on specific slices of temporal data, as well as those trained on a wide range of temporal data. To mitigate these problems, we propose a simple technique for jointly modeling text with its timestamp. This improves memorization of seen facts from the training time period, as well as calibration on predictions about unseen facts from future time periods. We also show that models trained with temporal context can be efficiently "refreshed" as new data arrives, without the need for retraining from scratch.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Bhuwan Dhingra (66 papers)
  2. Jeremy R. Cole (10 papers)
  3. Julian Martin Eisenschlos (27 papers)
  4. Daniel Gillick (11 papers)
  5. Jacob Eisenstein (73 papers)
  6. William W. Cohen (79 papers)
Citations (237)

Summary

We haven't generated a summary for this paper yet.

Youtube Logo Streamline Icon: https://streamlinehq.com