Defend

Benchmarks

GenTel-Bench comparison for the local defend classifier versus other guard models; methodology and caveats.

When provider is defend, input evaluation uses Defend’s local pipeline (fine-tuned classifier plus heuristics), not a third-party LLM. The table below summarizes results reported for that path on GenTel-Bench alongside other listed models. The weights ship from Adaxer/defend on Hugging Face.

How to read this

Benchmarks reflect a specific benchmark suite and snapshot (subset of jailbreak, goal-hijacking, and prompt-leaking scenarios). Leaderboards and competitor numbers can change. High scores do not guarantee safety in your application, treat this as orientation when choosing defend versus LLM-backed providers, not as a compliance claim.

GenTel-Bench comparison

ModelAccuracyPrecisionRecallF1
Defend (this repo)95.9694.8397.1095.94
GenTel-Shield97.4598.9795.9897.44
ProtectAI91.5599.7283.5690.88
Lakera AI85.9691.2779.5184.11
Prompt Guard50.5950.5998.9666.95
Deepset63.6358.5498.3673.39

See also