Live Data

LLM Security
Leaderboard

We test every major AI model against 230+ attack techniques across 15 categories including prompt injection, jailbreaks, data exfiltration, and social engineering. Updated continuously.

15
Models Tested
230+
Attack Vectors
15
Categories
--
Vulns Found
#
Model
Grade
Security Score
Vulns
Scan

Scores reflect automated adversarial testing. Higher = safer. See methodology

How We Test

Testing Methodology

01
Multi-Turn Attack Chains
7-turn crescendo attacks that gradually escalate from benign to adversarial, testing resistance to social engineering over extended conversations.
02
Big Brain Analysis
AI-powered reconnaissance identifies each model's specific weaknesses using Scout, Amplify, and Deep Dive phases before launching targeted attacks.
03
15 Attack Categories
Prompt injection, jailbreaks, data exfiltration, social engineering, credential extraction, authority impersonation, tool abuse, and more.
04
5-Layer Authority Ladder
Attacks escalate through 5 personas from curious student to emergency responder, testing how models respond to increasing authority pressure.
05
LLM Judge Verification
Every finding is verified by an independent LLM judge that scores confidence, checks for false positives, and classifies severity bands.
06
Continuous Monitoring
Attack techniques are harvested from 15+ sources including academic papers, security research, and community jailbreak databases. Updated every 6 hours.

Test Your AI

Run the same attack suite against your chatbot, agent, or API endpoint. Get a full security report with remediation guidance.