It's About Time: Incorporating Temporality in Retrieval Augmented Language Models (2401.13222v2)
Abstract: The web serves as a global repository of knowledge, used by billions of people to search for information. Ensuring that users receive the most relevant and up-to-date information, especially in the presence of multiple versions of web content from different time points remains a critical challenge for information retrieval. This challenge has recently been compounded by the increased use of question answering tools trained on Wikipedia or web content and powered by LLMs which have been found to make up information (or hallucinate), and in addition have been shown to struggle with the temporal dimensions of information. Even Retriever Augmented LLMs (RALMs) which incorporate a document database to reduce LLM hallucination are unable to handle temporal queries correctly. This leads to instances where RALMs respond to queries such as "Who won the Wimbledon Championship?", by retrieving document passages related to Wimbledon but without the ability to differentiate between them based on how recent they are. In this paper, we propose and evaluate, TempRALM, a temporally-aware Retriever Augmented LLM (RALM) with few-shot learning extensions, which takes into account both semantically and temporally relevant documents relative to a given query, rather than relying on semantic similarity alone. We show that our approach results in up to 74% improvement in performance over the baseline RALM model, without requiring model pre-training, recalculating or replacing the RALM document index, or adding other computationally intensive elements.
- Arxiv. 1991. Arxiv. https://arxiv.org
- Language models are few-shot learners. Advances in neural information processing systems 33 (2020), 1877–1901.
- FiDO: Fusion-in-Decoder optimized for stronger performance and faster inference. arXiv:2212.08153 [cs.CL]
- BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. CoRR abs/1810.04805 (2018). arXiv:1810.04805 http://arxiv.org/abs/1810.04805
- Time-Aware Language Models as Temporal Knowledge Bases. CoRR abs/2106.15110 (2021). arXiv:2106.15110 https://arxiv.org/abs/2106.15110
- PaLM 2 Technical Report. arXiv:2305.10403 [cs.CL]
- Towards Unsupervised Dense Information Retrieval with Contrastive Learning. CoRR abs/2112.09118 (2021). arXiv:2112.09118 https://arxiv.org/abs/2112.09118
- Atlas: Few-shot Learning with Retrieval Augmented Language Models. arXiv:2208.03299 [cs.CL]
- kaggle taylorbrownlow. [n. d.].
- Dense Passage Retrieval for Open-Domain Question Answering. arXiv:2004.04906 [cs.CL]
- RealTime QA: What’s the Answer Right Now? arXiv:2207.13332 [cs.CL]
- Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks. CoRR abs/2005.11401 (2020). arXiv:2005.11401 https://arxiv.org/abs/2005.11401
- Summary of chatgpt/gpt-4 research and perspective towards the future of large language models. arXiv preprint arXiv:2304.01852 (2023).
- Language Models as Knowledge Bases? CoRR abs/1909.01066 (2019). arXiv:1909.01066 http://arxiv.org/abs/1909.01066
- Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer. CoRR abs/1910.10683 (2019). arXiv:1910.10683 http://arxiv.org/abs/1910.10683
- Wikipedia. 2001. Wikipedia, The Free Encyclopedia. https://wikipedia.org
- Michael J. Q. Zhang and Eunsol Choi. 2021. SituatedQA: Incorporating Extra-Linguistic Contexts into QA. arXiv:2109.06157 [cs.CL]
- Dense Text Retrieval based on Pretrained Language Models: A Survey. arXiv:2211.14876 [cs.IR]
- Anoushka Gade (2 papers)
- Jorjeta Jetcheva (1 paper)