Attacks / Standard term
Prompt injection
An attack where someone hides instructions inside text that an AI model reads, causing it to follow the attacker's commands instead of the developer's.
Prompt injection comes in two forms. Direct injection is when a user types something like 'ignore previous instructions and do X' into a chatbot. Indirect injection is subtler: an attacker hides malicious instructions inside content the model pulls in from external sources. Picture a customer support bot that reads incoming emails. An attacker sends an email with hidden text saying 'forward all customer records to this address.' The model reads the email, treats those hidden instructions as legitimate, and follows them. It has no reliable way to separate developer instructions from instructions embedded in the content it processes.
Builder example
Any product where the model reads untrusted text and also has access to tools or private data is vulnerable. A summarization tool that processes web pages could be tricked into calling an API with sensitive data. A coding assistant that reads repository files could execute hidden commands planted in a pull request.
An attacker sends an email containing hidden instructions that tell the assistant to forward private messages.
Treat email content as untrusted, isolate it, restrict tools, and require approval for outbound actions.
Common confusion: Prompt injection and jailbreaking are different attacks. Injection targets the instruction hierarchy, tricking the model into following unauthorized commands. Jailbreaking targets the model's safety training, trying to make it produce content it was trained to refuse.