Rebuilding ROME : Resolving Model Collapse during Sequential Model Editing
Abstract: Recent work using Rank-One Model Editing (ROME), a popular model editing method, has shown that there are certain facts that the algorithm is unable to edit without breaking the model. Such edits have previously been called disabling edits. These disabling edits cause immediate model collapse and limits the use of ROME for sequential editing. In this paper, we show that disabling edits are an artifact of irregularities in the implementation of ROME. With this paper, we provide a more stable implementation ROME, which we call r-ROME and show that model collapse is no longer observed when making large scale sequential edits with r-ROME, while further improving generalization and locality of model editing compared to the original implementation of ROME. We also provide a detailed mathematical explanation of the reason behind disabling edits.
- The fifth pascal recognizing textual entailment challenge. TAC, 7:8.
- Evaluating the ripple effects of knowledge editing in language models. arXiv preprint arXiv:2307.12976.
- The pascal recognising textual entailment challenge. In Machine learning challenges workshop, pages 177–190. Springer.
- Editing factual knowledge in language models. arXiv preprint arXiv:2104.08164.
- Bill Dolan and Chris Brockett. 2005. Automatically constructing a corpus of sentential paraphrases. In Third International Workshop on Paraphrasing (IWP2005).
- The third pascal recognizing textual entailment challenge. In Proceedings of the ACL-PASCAL workshop on textual entailment and paraphrasing, pages 1–9.
- Model editing can hurt general abilities of large language models. arXiv preprint arXiv:2401.04700.
- Model editing at scale leads to gradual and catastrophic forgetting. arXiv preprint arXiv:2401.07453.
- The second pascal recognising textual entailment challenge. In Proceedings of the Second PASCAL Challenges Workshop on Recognising Textual Entailment, volume 7, pages 785–794.
- Wilke: Wise-layer knowledge editor for lifelong knowledge editing. arXiv preprint arXiv:2402.10987.
- Zero-shot relation extraction via reading comprehension. arXiv preprint arXiv:1706.04115.
- Locating and editing factual associations in gpt. Advances in Neural Information Processing Systems, 35:17359–17372.
- Mass-editing memory in a transformer. arXiv preprint arXiv:2210.07229.
- Fast model editing at scale. arXiv preprint arXiv:2110.11309.
- Memory-based model editing at scale. In International Conference on Machine Learning, pages 15817–15831. PMLR.
- Language models are unsupervised multitask learners. OpenAI blog, 1(8):9.
- Recursive deep models for semantic compositionality over a sentiment treebank. In Proceedings of the 2013 conference on empirical methods in natural language processing, pages 1631–1642.
- Massive editing for large language models via meta learning. arXiv preprint arXiv:2311.04661.
- Glue: A multi-task benchmark and analysis platform for natural language understanding. arXiv preprint arXiv:1804.07461.
- Ben Wang and Aran Komatsuzaki. 2021. GPT-J-6B: A 6 Billion Parameter Autoregressive Language Model. https://github.com/kingoflolz/mesh-transformer-jax.
- Neural network acceptability judgments. Transactions of the Association for Computational Linguistics, 7:625–641.
- The butterfly effect of model editing: Few edits can trigger large language models collapse. arXiv preprint arXiv:2402.09656.
- Editing large language models: Problems, methods, and opportunities. arXiv preprint arXiv:2305.13172.
- Mquake: Assessing knowledge editing in language models via multi-hop questions. arXiv preprint arXiv:2305.14795.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.