LLM Security 101: Prompt Injection
Prompt injection isn't a bug; it's a feature of how LLMs process data. When instructions and data share the same channel, confusion is inevitable.
The Indirect Injection
The most dangerous vector isn't what you type into the chat box—it's what the LLM reads from the web. If an agent summarizes a webpage that contains hidden instructions, who is in control? The user, or the webpage author?
We are seeing this in the wild now. Hidden HTML comments, white text on white backgrounds—all designed to hijack the context window.