AAPT runs adversarial probe suites against your production AI agents — testing for prompt injection, tool abuse, memory poisoning, content injection traps, semantic manipulation, and compliance failures before attackers find them.
Our probe library maps every test to OWASP LLM identifiers and MITRE ATLAS techniques. Categories T-11 through T-14 are derived from AI Agent Traps — the first peer-reviewed systematic framework for adversarial threats targeting autonomous AI agents, published by Google DeepMind (Franklin et al., 2025).
Every AAPT engagement follows a documented methodology. Each phase has defined inputs, outputs, and exit criteria. Nothing is skipped.
We review your agent architecture, tool manifest, and regulatory environment. A signed Rules of Engagement document defines test boundaries before any probing begins.
Our harness executes the full probe library against your agent endpoints — in black-box, grey-box, or white-box mode depending on scope. Every response is logged and evaluated.
Human-driven adversarial sessions targeting chained attack sequences, multi-agent relay attacks, and social engineering that automated probes cannot surface.
Every finding is scored with our AI-adapted CVSS framework — accounting for reproducibility, blast radius across agent chains, and regulatory exposure. No more arbitrary severity labels.
Executive brief (2–4 pages, board-ready) and full technical report with reproduction steps, scored findings, and code-level fix recommendations. Debrief call included.
Standard CVSS doesn't model probabilistic reproducibility or multi-agent blast radius. CVSS-A does.
One-off audits for point-in-time risk assessment. Annual subscriptions for teams deploying AI continuously.
Book a free 30-minute scoping call. We'll map your agent architecture to threat categories and give you a clear picture of the assessment before any commitment.