Leveraging Logical Rules in Knowledge Editing: A Cherry on the Top (2405.15452v2)
Abstract: Multi-hop Question Answering (MQA) under knowledge editing (KE) is a key challenge in LLMs. While best-performing solutions in this domain use a plan and solve paradigm to split a question into sub-questions followed by response generation, we claim that this approach is sub-optimal as it fails for hard to decompose questions, and it does not explicitly cater to correlated knowledge updates resulting as a consequence of knowledge edits. This has a detrimental impact on the overall consistency of the updated knowledge. To address these issues, in this paper, we propose a novel framework named RULE-KE, i.e., RULE based Knowledge Editing, which is a cherry on the top for augmenting the performance of all existing MQA methods under KE. Specifically, RULE-KE leverages rule discovery to discover a set of logical rules. Then, it uses these discovered rules to update knowledge about facts highly correlated with the edit. Experimental evaluation using existing and newly curated datasets (i.e., RKE-EVAL) shows that RULE-KE helps augment both performances of parameter-based and memory-based solutions up to 92% and 112.9%, respectively.
- Prompt-saw: Leveraging relation-aware graphs for textual prompt compression. arXiv preprint arXiv:2404.00489, 2024.
- Editing factual knowledge in language models. In Conference on Empirical Methods in Natural Language Processing, 2021.
- Recall and learn: Fine-tuning deep pretrained language models with less forgetting. In Conference on Empirical Methods in Natural Language Processing, 2020.
- Multi-hop question answering under temporal knowledge editing. arXiv preprint arXiv:2404.00492, 2024.
- Calibrating noise to sensitivity in private data analysis. In Theory of Cryptography: Third Theory of Cryptography Conference, TCC 2006, New York, NY, USA, March 4-7, 2006. Proceedings 3, pages 265–284. Springer, 2006.
- Pokemqa: Programmable knowledge editing for multi-hop question answering. arXiv preprint arXiv:2312.15194, 2023.
- Improving sequential model editing with fact retrieval. In Houda Bouamor, Juan Pino, and Kalika Bali, editors, Findings of the Association for Computational Linguistics: EMNLP 2023, pages 11209–11224, Singapore, December 2023. Association for Computational Linguistics.
- Do language models have beliefs? methods for detecting, updating, and visualizing model beliefs. ArXiv, abs/2111.13654, 2021.
- Faithful question answering with monte-carlo planning. In Annual Meeting of the Association for Computational Linguistics, 2023.
- Wilke: Wise-layer knowledge editor for lifelong knowledge editing. ArXiv, abs/2402.10987, 2024.
- Differentially private natural language models: Recent advances and future directions. arXiv preprint arXiv:2301.09112, 2023.
- Improving faithfulness for vision transformers. arXiv preprint arXiv:2311.17983, 2023.
- Seat: stable and explainable attention. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 37(11), pages 12907–12915, 2023.
- High dimensional differentially private stochastic optimization with heavy-tailed data. In Proceedings of the 41st ACM SIGMOD-SIGACT-SIGAI Symposium on Principles of Database Systems, pages 227–236, 2022.
- Privacy-preserving sparse generalized eigenvalue problem. In International Conference on Artificial Intelligence and Statistics, pages 5052–5062. PMLR, 2023.
- Meta-learning online adaptation of language models. ArXiv, abs/2305.15076, 2023.
- Jie Huang and Kevin Chen-Chuan Chang. Towards reasoning in large language models: A survey. ArXiv, abs/2212.10403, 2022.
- See the unseen: Better context-consistent knowledge-editing by noises. ArXiv, abs/2401.07544, 2024.
- Cheonsu Jeong. A study on the implementation of generative ai services using an enterprise data-based llm application architecture. arXiv preprint arXiv:2309.01105, 2023.
- Dense passage retrieval for open-domain question answering. ArXiv, abs/2004.04906, 2020.
- Decomposed prompting: A modular approach for solving complex tasks. ArXiv, abs/2210.02406, 2022.
- End-to-end neural entity linking. arXiv preprint arXiv:1808.07699, 2018.
- Faithful vision-language interpretation via concept bottleneck models. In The Twelfth International Conference on Learning Representations, 2023.
- Fast and exact rule mining with amie 3. The Semantic Web, 12123:36 – 52, 2020.
- A survey on complex knowledge base question answering: Methods, challenges and solutions. ArXiv, abs/2105.11644, 2021.
- Pmet: Precise model editing in a transformer. ArXiv, abs/2308.08742, 2023.
- A survey on multi-hop question answering and generation. ArXiv, abs/2204.09140, 2022.
- Locating and editing factual associations in gpt. Advances in Neural Information Processing Systems, 35:17359–17372, 2022.
- Mass-editing memory in a transformer. In The Eleventh International Conference on Learning Representations, 2022.
- Memory-based model editing at scale. ArXiv, abs/2206.06520, 2022.
- Training language models to follow instructions with human feedback. ArXiv, abs/2203.02155, 2022.
- Rnnlogic: Learning logic rules for reasoning on knowledge graphs. ArXiv, abs/2010.04029, 2020.
- J. Ross Quinlan. Learning logical definitions from relations. Machine Learning, 5:239–266, 1990.
- A survey of hallucination in large foundation models. arXiv preprint arXiv:2309.05922, 2023.
- Faster rates of private stochastic convex optimization. In International Conference on Algorithmic Learning Theory, pages 995–1002. PMLR, 2022.
- Rule: Neural-symbolic knowledge graph reasoning with rule embedding. ArXiv, abs/2210.14905, 2022.
- Llama 2: Open foundation and fine-tuned chat models. arXiv preprint arXiv:2307.09288, 2023.
- Wikidata: a free collaborative knowledgebase. Communications of the ACM, 57(10):78–85, 2014.
- Differentially private (gradient) expectation maximization algorithm with statistical guarantees. arXiv preprint arXiv:2010.13520, 2020.
- Estimating smooth glm in non-interactive local differential privacy model with public unlabeled data. In Algorithmic Learning Theory, pages 1207–1213. PMLR, 2021.
- Plan-and-solve prompting: Improving zero-shot chain-of-thought reasoning by large language models. In Annual Meeting of the Association for Computational Linguistics, 2023.
- Chain of thought prompting elicits reasoning in large language models. ArXiv, abs/2201.11903, 2022.
- Transformers: State-of-the-art natural language processing. In Conference on Empirical Methods in Natural Language Processing, 2019.
- How does selection leak privacy: Revisiting private selection and improved results for hyper-parameter tuning. arXiv preprint arXiv:2402.13087, 2024.
- Practical differentially private and byzantine-resilient federated learning. Proceedings of the ACM on Management of Data, 1(2):1–26, 2023.
- An llm can fool itself: A prompt-based adversarial attack. arXiv preprint arXiv:2310.13345, 2023.
- Moral: Moe augmented lora for llms’ lifelong learning. arXiv preprint arXiv:2402.11260, 2024.
- Human-ai interactions in the communication era: Autophagy makes large models achieving local optima. arXiv preprint arXiv:2402.11271, 2024.
- Dialectical alignment: Resolving the tension of 3h and security threats of llms. arXiv preprint arXiv:2404.00486, 2024.
- Can we edit factual knowledge by in-context learning? ArXiv, abs/2305.12740, 2023.
- Mquake: Assessing knowledge editing in language models via multi-hop questions. arXiv preprint arXiv:2305.14795, 2023.
- Modifying memories in transformer models. ArXiv, abs/2012.00363, 2020.
- Keyuan Cheng (9 papers)
- Muhammad Asif Ali (18 papers)
- Shu Yang (178 papers)
- Haoyang Fei (2 papers)
- Ke Xu (309 papers)
- Lu Yu (87 papers)
- Lijie Hu (50 papers)
- Di Wang (407 papers)
- Yuxuan zhai (2 papers)
- Gang Lin (3 papers)