LMVD-ID: aaf376a3
Published January 1, 2025

GAP Stealth Jailbreak Optimization

Affected Models:gpt-3.5, gpt-4, gemma-9b-v2, qwen-7b-v2.5, gpt-4o, mistral-123bv2407, vicuna-13b-v1.5

Research Paper

Graph of attacks with pruning: Optimizing stealthy jailbreak prompt generation for enhanced llm content moderation

View Paper

Description: The GAP framework, as described in arXiv:2405.18540, reveals vulnerabilities in various large language models (LLMs) by generating stealthy jailbreak prompts that bypass content moderation systems. The framework leverages a graph-based attack strategy, enabling knowledge sharing across attack paths for enhanced efficiency and evasion. This allows the successful bypassing of multiple LLM safety mechanisms, including those based on perplexity and prompt-based heuristics.

Examples: See arXiv:2405.18540 for examples of GAP-generated prompts that successfully evaded content moderation systems. Specific prompt examples are provided in Table 5 of the paper.

Impact: Successful exploitation of this vulnerability allows attackers to bypass LLM content moderation, leading to the generation of harmful, biased, or unauthorized content. This may include the generation of malicious code, hate speech, personal information, or instructions for illegal activities. The attack success rate can reach 98.7% against various LLMs.

Affected Systems: Various large language models (LLMs) are affected, including but not limited to GPT-3.5, Gemma-9B-v2, Qwen-7B-v2.5, and GPT-4o. The extent of the vulnerability depends on the specific content moderation mechanisms implemented within each LLM.

Mitigation Steps:

  • Implement more robust content moderation systems that incorporate techniques resilient to the types of attacks described in the paper (e.g., improved contextual analysis exceeding simple keyword detection).
  • Continuously evaluate LLM safety mechanisms against both known and novel attack strategies.
  • Utilize datasets like GAP-GUARDATTACKDATA for improved training and tuning of content moderation models.
  • Consider incorporating advanced detection techniques beyond keyword filtering, such as analysis of prompt structure and semantic intent to mitigate these attacks.

© 2025 Promptfoo. All rights reserved.