Comprehensive architectural solution for distinguishing instructions from data in LLMs
Develop a comprehensive mechanism that reliably enforces a separation between instructions from trusted entities and data from untrusted sources within large language model (LLM) inference and application contexts, overcoming the current architectural inability to distinguish instructions from data so as to prevent prompt injection at the architectural level rather than via application-layer guardrails.
References
The inability to distinguish instructions from data admits no known comprehensive solution at the time of this writing.
— The Promptware Kill Chain: How Prompt Injections Gradually Evolved Into a Multi-Step Malware
(2601.09625 - Nassi et al., 14 Jan 2026) in Conclusion