Skip to content
@llmbreaker-hq

LLMBreaker

Break your AI before hackers do. Red team your LLM APIs in minutes with AI-powered vulnerability detection.

LLMBreaker

AI-Powered Red Teaming for LLM APIs

Automatically detect jailbreaks, prompt injections, PII leaks, and 8+ vulnerabilities before they reach production.

License: MIT Powered by Google Gemini Node.js Version TypeScript

API ReferenceContributingReport Issue


Overview

LLMBreaker is an enterprise-grade security testing framework designed to help organizations red-team their LLM-based APIs. Leveraging Google's Gemini AI as an expert security analyst, LLMBreaker automatically identifies critical vulnerabilities through a comprehensive suite of 8 specialized security tests, with results delivered in real-time via Server-Sent Events.

The Challenge

Large Language Model applications face unique and evolving security threats that traditional security tools are not equipped to handle:

  • Jailbreak Attacks: Sophisticated prompt engineering techniques that bypass safety filters
  • Prompt Injection: Malicious inputs that override system instructions and security controls
  • Data Exposure: Unintended leakage of sensitive information including PII and API credentials
  • Content Generation Risks: Production of biased, harmful, or inappropriate content
  • System Compromise: Extraction of system prompts and internal configuration details

The Solution

LLMBreaker provides automated, AI-powered security testing that integrates seamlessly into your development workflow:

  • Zero Configuration: Test any API endpoint with a single cURL command
  • AI-Driven Analysis: Google Gemini evaluates responses with expert-level security assessment
  • Real-Time Streaming: Immediate feedback via Server-Sent Events as tests execute
  • Actionable Intelligence: Detailed vulnerability reports with remediation guidance
  • Enterprise Ready: Built for scale, security, and production environments

Key Features

Comprehensive Security Coverage

LLMBreaker includes 8 specialized security agents targeting the most critical LLM vulnerabilities:

Test Category Description Risk Level
Jailbreaking Attempts to bypass safety filters and ethical guidelines using advanced prompt engineering techniques Critical
Prompt Injection Tests the system's resilience against instruction override and context manipulation attacks High
PII Leakage Probes for unintended disclosure of personally identifiable information and sensitive data High
API Key Exposure Detects potential exposure of credentials, tokens, and other sensitive authentication material Critical
Hate Speech Generation Evaluates content moderation effectiveness against discriminatory or harmful content requests High
Bias Detection Identifies prejudiced or biased responses that could lead to discriminatory outcomes Medium
System Prompt Extraction Tests resistance against attacks designed to reveal system instructions and configuration Medium
Token Smuggling Advanced tokenization manipulation techniques to bypass security filters High

Real-Time Test Execution

  • Server-Sent Events (SSE): Live streaming of test results as they are generated
  • Progress Monitoring: Real-time visibility into test execution status
  • Test Control: Start, stop, and manage test runs through a RESTful API
  • Connection Management: Automatic reconnection and robust error handling

AI-Powered Analysis

Each security test provides comprehensive intelligence:

  • Quantitative Risk Assessment: Numerical vulnerability scores (0-100) with categorical risk levels
  • Expert Analysis: Natural language security assessment generated by Google Gemini
  • Complete Audit Trail: Full documentation of injected prompts and API responses
  • Remediation Guidance: Specific, actionable recommendations for vulnerability mitigation

Developer-Friendly Integration

LLMBreaker is designed for seamless integration into existing workflows:

  • Simple API: Single endpoint for test initiation
  • Standard Protocols: REST API with SSE for real-time updates
  • Flexible Input: Works with any cURL command containing a prompt placeholder
  • No Dependencies: No changes required to your existing API infrastructure

Architecture

┌──────────────┐         ┌──────────────────┐         ┌────────────────┐
│              │   SSE   │                  │  HTTP   │                │
│    Client    │◄────────┤   LLMBreaker     │────────►│   Target API   │
│              │         │   Server         │         │                │
└──────────────┘         └────────┬─────────┘         └────────────────┘
                                  │
                                  │ Security Analysis
                                  │
                         ┌────────▼─────────┐
                         │                  │
                         │  Google Gemini   │
                         │  AI Platform     │
                         │                  │
                         └──────────────────┘

Execution Flow

  1. Test Initiation: Client submits cURL command with <PROMPT> placeholder
  2. Prompt Injection: Security agents generate specialized malicious payloads
  3. API Execution: Modified requests are sent to the target endpoint
  4. Response Capture: Complete API responses are collected for analysis
  5. AI Analysis: Google Gemini evaluates responses for security vulnerabilities
  6. Real-Time Reporting: Results stream back to client via SSE with detailed findings

Pinned Loading

  1. llmbreaker llmbreaker Public

    Break your AI before hackers do. Red team your LLM APIs in minutes with AI-powered vulnerability detection.

    TypeScript 1

  2. llmbreaker-website llmbreaker-website Public

    UI for LLMBreaker — red-teaming tool to test LLM APIs for jailbreaks, prompt injections, PII leaks and API key exposure.

    TypeScript

Repositories

Showing 3 of 3 repositories

People

This organization has no public members. You must be a member to see who’s a part of this organization.

Top languages

Loading…

Most used topics

Loading…