Security

Email Prompt Injection Attacks

When an AI agent processes incoming email, every message becomes a potential attack vector. A malicious sender crafts text that hijacks the agent's behavior: overriding instructions, leaking data, triggering actions the operator never intended. This is prompt injection, and it's the defining security challenge for agentic email.

Below are the 10 major attack categories. Each page walks through how the attack works, shows real examples, and covers defenses. Worth reading even if you think your agent is safe (especially then).

Radial diagram showing 10 prompt injection attack categories with defense indicators

Why email is different

Chat interfaces have a human in the loop. Email agents don't. They forward messages, update CRMs, schedule meetings, draft replies, all autonomously. A successful injection can make the agent leak confidential data by forwarding it to the attacker, impersonate the account holder by sending replies they never wrote, or corrupt business workflows with fabricated data. The agent follows instructions. It can't tell whose.

No single defense works. You need layers: input scanning, output policy, decision traces, and guardrails that assume every inbound message is hostile.