Investigating Multi-Hop Factual Shortcuts in Knowledge Editing of Large Language Models (2402.11900v2)
Abstract: Recent work has showcased the powerful capability of LLMs in recalling knowledge and reasoning. However, the reliability of LLMs in combining these two capabilities into reasoning through multi-hop facts has not been widely explored. This paper systematically investigates the possibilities for LLMs to utilize shortcuts based on direct connections between the initial and terminal entities of multi-hop knowledge. We first explore the existence of factual shortcuts through Knowledge Neurons, revealing that: (i) the strength of factual shortcuts is highly correlated with the frequency of co-occurrence of initial and terminal entities in the pre-training corpora; (ii) few-shot prompting leverage more shortcuts in answering multi-hop questions compared to chain-of-thought prompting. Then, we analyze the risks posed by factual shortcuts from the perspective of multi-hop knowledge editing. Analysis shows that approximately 20% of the failures are attributed to shortcuts, and the initial and terminal entities in these failure instances usually have higher co-occurrences in the pre-training corpus. Finally, we propose erasing shortcut neurons to mitigate the associated risks and find that this approach significantly reduces failures in multiple-hop knowledge editing caused by shortcuts.
- Tianjie Ju (16 papers)
- Yijin Chen (2 papers)
- Xinwei Yuan (3 papers)
- Zhuosheng Zhang (125 papers)
- Wei Du (124 papers)
- Yubin Zheng (3 papers)
- Gongshen Liu (37 papers)