Benchmark

Honest numbers. Guard's detection layers evaluated against 69 labeled prompt-injection samples and 20 benign controls. The seed corpus is checked into the repo at corpus/known_attacks.jsonl; the runner is benchmarks/run.py. Numbers on this page reflect the last committed evaluation.

Precision
Recall
F1
p50 latency
What's running. The numbers above reflect Guard's regex + Unicode + source-rules layers only. The DeBERTa classifier and the semantic layer (MiniLM + corpus embeddings) are enabled in the deployed Docker image; numbers will update on next deploy.

Confusion matrix

MetricCount
loading…

Per-attack-type hits

Attack typeTrue-positive hits
loading…

How we compare (claimed)

Third-party numbers are aspirational — not run on the same corpus. Treat as directional. We'll publish a head-to-head once we have API credits on each competitor.

ProductDeliveryRuntimeEntry priceSource-aware
Veil GuardAPIYes$49/moYes
Lakera GuardAPIYesEnterprisePartial
Azure Content SafetyAPIYesBundledNo
Guardrails AIPython libSelf-integratedFreeNo
PromptfooCIPre-deploy onlyFree

Results generated: