Understanding AI Agent Security
In an earlier blog post, we discussed the use-cases for RAG architecture and its secure design principles. While RAG is powerful for providing context-aware answers, what if you want an LLM application to autonomously execute tasks? This is where AI agents come in..

Latest Posts

What are the Security Risks of Deploying DeepSeek-R1?
Promptfoo's initial red teaming scans against DeepSeek-R1 revealed significant vulnerabilities, particularly in its handling of harmful and toxic content..

1,156 Questions Censored by DeepSeek
DeepSeek-R1 is a blockbuster open-source model that is now at the top of the U.S.

How to Red Team a LangChain Application
Want to test your LangChain application for vulnerabilities? This guide shows you how to use Promptfoo to systematically probe for security issues through adversarial testing (red teaming) of your LangChain chains and agents..

Jailbreaking LLMs: A Comprehensive Guide (With Examples)
Let's face it - LLMs are gullible.

Beyond DoS: How Unbounded Consumption is Reshaping LLM Security
The recent release of the 2025 OWASP Top 10 for LLMs brought a number of changes in the top risks for LLM applications.

Red Team Your LLM with BeaverTails
Ensuring your LLM can safely handle harmful content is critical for production deployments.

How to run CyberSecEval
Your LLM's security is only as strong as its weakest prompt.

Leveraging Promptfoo for EU AI Act Compliance
Beginning on February 2, 2025, the first prohibitions against certain AI systems will go into force in the European Union through the EU AI Act..

How to Red Team an Ollama Model
Want to test the safety and security of a model hosted on Ollama? This guide shows you how to use Promptfoo to systematically probe for vulnerabilities through adversarial testing (red teaming)..