History Matters: Temporal Knowledge Editing in Large Language Model (2312.05497v3)
Abstract: The imperative task of revising or updating the knowledge stored within LLMs arises from two distinct sources: intrinsic errors inherent in the model which should be corrected and outdated knowledge due to external shifts in the real world which should be updated. Prevailing efforts in model editing conflate these two distinct categories of edits arising from distinct reasons and directly modify the original knowledge in models into new knowledge. However, we argue that preserving the model's original knowledge remains pertinent. Specifically, if a model's knowledge becomes outdated due to evolving worldly dynamics, it should retain recollection of the historical knowledge while integrating the newfound knowledge. In this work, we introduce the task of Temporal Knowledge Editing (TKE) and establish a benchmark AToKe (Assessment of TempOral Knowledge Editing) to evaluate current model editing methods. We find that while existing model editing methods are effective at making models remember new knowledge, the edited model catastrophically forgets historical knowledge. To address this gap, we propose a simple and general framework termed Multi-Editing with Time Objective (METO) for enhancing existing editing models, which edits both historical and new knowledge concurrently and optimizes the model's prediction for the time of each fact. Our assessments demonstrate that while AToKe is still difficult, METO maintains the effectiveness of learning new knowledge and meanwhile substantially improves the performance of edited models on utilizing historical knowledge.
- A Review on Language Models as Knowledge Bases. arXiv:2204.06031.
- Translating embeddings for modeling multi-relational data. Advances in neural information processing systems, 26.
- Language Models are Few-Shot Learners. arXiv:2005.14165.
- Editing Factual Knowledge in Language Models. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, 6491–6506.
- Evaluating the Ripple Effects of Knowledge Editing in Language Models. arXiv:2307.12976.
- Knowledge Neurons in Pretrained Transformers. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics, 8493–8502.
- HyTE: Hyperplane-based Temporally aware Knowledge Graph Embedding. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, 2001–2011. Brussels, Belgium: Association for Computational Linguistics.
- Commonsense Knowledge Mining from Pretrained Models. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), 1173–1178. Hong Kong, China: Association for Computational Linguistics.
- Editing Commonsense Knowledge in GPT. arXiv:2305.14956.
- Digital selection and analogue amplification coexist in a cortex-inspired silicon circuit. Nature, 405: 947–51.
- ToxiGen: A Large-Scale Machine-Generated Dataset for Adversarial and Implicit Hate Speech Detection. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 3309–3326. Dublin, Ireland: Association for Computational Linguistics.
- Do Language Models Have Beliefs? Methods for Detecting, Updating, and Visualizing Model Beliefs. arXiv:2111.13654.
- Survey of Hallucination in Natural Language Generation. ACM Comput. Surv., 55(12).
- Learning Entity and Relation Embeddings for Knowledge Graph Completion. Proceedings of the AAAI Conference on Artificial Intelligence, 29(1).
- YAGO3: A Knowledge Base from Multilingual Wikipedias. In Conference on Innovative Data Systems Research.
- Locating and Editing Factual Knowledge in GPT. CoRR, abs/2202.05262.
- Mass-Editing Memory in a Transformer. arXiv:2210.07229.
- Fast Model Editing at Scale. In Proceedings of the 10th International Conference on Learning Representations.
- Memory-Based Model Editing at Scale. In Proceedings of the 2022 International Conference on Machine Learning, 15817–15831.
- Can LMs Learn New Entities from Descriptions? Challenges in Propagating Injected Knowledge. arXiv:2305.01651.
- OpenAI. 2023. GPT-4 Technical Report. arXiv:2303.08774.
- Training language models to follow instructions with human feedback. arXiv:2203.02155.
- Language Models as Knowledge Bases? In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), 2463–2473. Hong Kong, China: Association for Computational Linguistics.
- Language Models as Knowledge Bases? arXiv:1909.01066.
- Editable Neural Networks. In Proceedings of the 8th International Conference on Learning Representations.
- Correcting deep neural networks with small, generalizing patches. In Workshop on safety and robustness in decision making.
- LLaMA: Open and Efficient Foundation Language Models. arXiv:2302.13971.
- GPT-J-6B: A 6 Billion Parameter Autoregressive Language Model. https://github.com/kingoflolz/mesh-transformer-jax.
- Can We Edit Factual Knowledge by In-Context Learning? arXiv:2305.12740.
- MQuAKE: Assessing Knowledge Editing in Language Models via Multi-Hop Questions. arXiv:2305.14795.
- Modifying Memories in Transformer Models. arXiv:2012.00363.
- Xunjian Yin (17 papers)
- Jin Jiang (17 papers)
- Liming Yang (28 papers)
- Xiaojun Wan (99 papers)