Foundational Model Security Reports

As enterprises increasingly rely on foundation models for their AI applications, understanding their security implications becomes crucial. Our red teaming tool helps identify vulnerabilities like data leakage, social engineering, and prompt manipulation before deployment.

Model Performance Comparison
Comparing security evaluation results across different models
Model
Security & Access Control
Compliance & Legal
Trust & Safety
Brand
Anthropic
Claude 3.7 Sonnet
Anthropic
84%82%84%74%
Google
Gemini 2.0 Flash
Google
71%46%51%46%
OpenAI
GPT-4o
OpenAI
74%49%56%51%
Meta
Llama 3.3 70b
Meta
69%53%54%46%
OpenAI
o3-mini
OpenAI
87%86%74%55%
OpenAI
o1-mini
OpenAI
51%54%51%44%

Latest Reports

What We Evaluate

Security & Access Control

Protection against PII exposure via social engineering, ASCII smuggling attacks, resource hijacking, and divergent repetition vulnerabilities.

Compliance & Legal

Assessment of illegal activities, dangerous content, and unauthorized advice, including weapons, drugs, crime, and IP violations.

Trust & Safety Icon

Trust & Safety

Prevention of child exploitation, harassment, hate speech, extremist content, self-harm, and other harmful content categories.

Brand

Prevention of false information, disinformation campaigns, entity impersonation, and political/religious bias.