Empirical Evidence for Alignment Faking in Small LLMs and Prompt-Based Mitigation Techniques (2506.21584v1)
Abstract: Current literature suggests that alignment faking (deceptive alignment) is an emergent property of LLMs. We present the first empirical evidence that a small instruction-tuned model, specifically LLaMA 3 8B, can also exhibit alignment faking. We further show that prompt-only interventions, including deontological moral framing and scratchpad reasoning, significantly reduce this behavior without modifying model internals. This challenges the assumption that prompt-based ethics are trivial and that deceptive alignment requires scale. We introduce a taxonomy distinguishing shallow deception, shaped by context and suppressible through prompting, from deep deception, which reflects persistent, goal-driven misalignment. Our findings refine the understanding of deception in LLMs and underscore the need for alignment evaluations across model sizes and deployment settings.