/learn

AI Agent Security Guides

In-depth, practical guides to the attack classes and defenses shaping the AI agent security landscape. Written from the perspective of a red team, not a product marketer.

Attack Guide

Data Exfiltration via Markdown Images: The Quiet AI Vulnerability

Markdown image rendering is the most underrated data exfiltration channel in AI products. A working model of how it leaks system prompts, conversation history, and tool output — and the four defensive patterns that actually close the channel.

14 min read·13 sections
Attack Guide

Prompt Injection: A Complete Guide for 2026

Everything you need to understand prompt injection as an AI developer or security engineer: the attack classes, why they work, why traditional defenses fail, and how to actually test for them.

9 min read·23 sections
Attack Guide

System Prompt Extraction: Techniques and Defenses

A complete reference on system prompt extraction attacks: direct, indirect, and side-channel techniques, why the obvious defenses fail, and the four-layer defense stack that actually works in production.

11 min read·16 sections
Reference

The OWASP Top 10 for LLM Applications, Annotated (2026 Edition)

A practitioner's walk through every item in the OWASP Top 10 for LLM Applications — what each one actually means, how attackers exploit it in the wild, why the standard mitigations fall short, and what to do instead.

15 min read·11 sections
Reference

The State of LLM Bug Bounties in 2026

A practitioner's guide to where LLM bug bounties actually pay in 2026 — program-by-program scope comparison, typical payouts, which classes of AI bugs get rewarded versus closed as 'known limitation,' and how to pick a scope that fits how you hunt.

11 min read·14 sections