Depth of belief from LLM knowledge editing
Determine whether existing large language model knowledge editing techniques—specifically prompting-based insertion, mechanistic model editing (e.g., AlphaEdit/MEMIT-style surgical rewrites), and finetuning-based approaches such as Synthetic Document Finetuning—produce deep modifications that resemble genuine belief rather than superficial changes or parroting of inserted facts across diverse contexts and tasks.
References
While various methods have been proposed to edit the knowledge of LLMs, it is unclear whether these techniques cause superficial changes and mere parroting of facts as opposed to deep modifications that resemble genuine belief.
— Believe It or Not: How Deeply do LLMs Believe Implanted Facts?
(2510.17941 - Slocum et al., 20 Oct 2025) in Section 1 (Introduction)