What is prompt injection?
Prompt injection is when untrusted input tries to override instructions or manipulate the model into unsafe behavior.
- Often comes through retrieved or user content
- Model sees it as text not trusted code
- Needs system design defenses
What is prompt injection?