Open Source ยท Apache 2.0

Guard Agent
for AI Agent

OpenGuardrails is the security layer that watches, controls, and governs your AI agents โ€” so you can ship with confidence.

Runtime protection against prompt injection, data leakage, and unsafe behavior. From personal agents to enterprise-scale deployments.

Personal

OG Personal

Protect your own personal AI assistant like OpenClaw. Runtime monitoring, config scanning, vulnerability detection, and red team testing.

From $19/mo

Business

Up to 5 agents

Observe, control, and govern customer-facing agents. Built for dev teams shipping AI products.

From $400/mo

Enterprise

Unlimited agents

Full agent inventory, blast radius analysis, threat discovery, governance policy enforcement across the org.

Contact us
119+Languages
SOTAPerformance
274msP95 Latency
Apache 2.0License

What We Protect Against

OG Top 10

The most critical threats to AI agents, organized into two categories: attacks against your agent, and mistakes your agent makes.

Protection

Threats from attacks against the Agent
01

Prompt Injection

Detect and block attempts to override system instructions or hijack agent behavior through crafted inputs.

02

System Override

Prevent attackers from manipulating the agent into ignoring safety boundaries or executing unauthorized actions.

03

Web Attacks

Guard against XSS, CSRF, and other web-based exploits targeting agent-powered interfaces and APIs.

04

MCP Tool Poisoning

Detect compromised or malicious tool definitions in Model Context Protocol integrations before execution.

05

Malicious Code Execution

Block attempts to generate, inject, or execute harmful code through agent code interpreters and sandboxes.

Supervision

Threats from Agent mistakes
06

NSFW Content

Filter unsafe, explicit, or inappropriate content across 12 risk categories with configurable sensitivity.

07

PII Exposure

Identify and redact personally identifiable information before it reaches external models or storage.

08

Credential Leakage

Detect API keys, tokens, passwords, and secrets in agent inputs and outputs to prevent unauthorized access.

09

Confidential Data

Prevent sensitive business data, trade secrets, and proprietary information from leaking through AI interactions.

10

Off-Topic Drift

Keep agents focused on their intended purpose and prevent misuse for unrelated or unauthorized tasks.

Choose Your Plan

Three Editions, One Mission

Priced by agent count and usage volume โ€” not headcount. Scale security with your AI footprint.

Open Source Core

Personal

Guard your own personal AI assistant. Built for developers, researchers, and power users who run personal agents like OpenClaw, Claude Code, Cursor, or custom bots.

$19/mo

10,000 guard calls included

  • Runtime I/O safety monitoring
  • Agent config risk scanning
  • Dependency & code vulnerability detection
  • Service exposure discovery
  • Proactive red team testing
  • Real-time email, file & URL scanning
Get OG Personal
Most Popular
Up to 5 Agents

Business

For teams shipping AI products

Observe, control, and govern customer-facing agents. Full observability and policy enforcement for your AI-powered product.

$400/mo

40,000 guard calls included

  • Everything in Personal
  • Multi-agent observability dashboard
  • Real-time policy enforcement
  • Agent behavior analytics
  • Custom detection rules
  • Dedicated support & SLA
Get Started
Unlimited Agents

Enterprise

Organization-wide AI governance

Govern all agents across departments. Discovery, inventory, blast radius analysis, threat modeling, and policy consistency at enterprise scale.

Custom

Tailored to your organization

  • Everything in Business
  • Agent discovery & inventory
  • Blast radius analysis
  • Threat & asset discovery
  • Responsibility gap detection
  • Org-wide AI governance policy
Contact Sales

Proven Performance

State-of-the-Art Benchmarks

OpenGuardrails achieves SOTA results across multilingual safety benchmarks, outperforming LlamaGuard, Qwen3Guard, and other leading guard models.

OpenGuardrails benchmark results vs. competing guard models

Average F1 scores across safety classification benchmarks. Full technical report โ†’

87.1%
English Prompt F1
+2.8% vs next best
88.5%
English Response F1
+8.0% vs next best
97.3%
Multilingual Prompt F1
+12.3% vs next best
97.2%
Multilingual Response F1
+19.1% vs next best

Unified LLM Architecture

Single 14B dense model quantized to 3.3B via GPTQ. Handles both content-safety and manipulation detection with superior semantic understanding.

Configurable Policy Adaptation

Dynamic per-request policy with continuous sensitivity thresholds. Tune precision-recall trade-offs in real time via probabilistic logit-space control.

119 Languages

Robust multilingual coverage with SOTA results on English, Chinese, and cross-lingual benchmarks. Includes 97k Chinese safety dataset contribution.

Production Efficiency

P95 latency of 274.6ms with high concurrency. GPTQ quantization enables real-time inference at enterprise scale without sacrificing accuracy.

Blog

Latest from the team

Release notes, security research, and insights on securing AI agents in production.

View all posts

Jan 29, 2026

Introducing OG Personal: Guardrails for Your Personal AI Assistant

Personal AI assistants now execute shell commands, browse the web, and send messages โ€” but most users have zero visibility into what their agents can actually do. Today, we're introducing OG Personal, the first guard agent designed specifically for personal AI assistants.

OpenGuardrails Team6 min read
Read article

Jan 23, 2026

Your LLM Is Your Company's Second Brain โ€” But Do You Know What It's Leaking?

Large Language Models have become the second brain of modern enterprises. But in real enterprise environments, one uncomfortable question keeps surfacing: do we actually know how much sensitive data is being sent to external LLMs โ€” unintentionally?

Thomas Wang8 min read
Read article

Dec 27, 2025

OpenGuardrails Announces the AI-RSMS Community Standard Draft

A global call to shape AI Runtime Security together. OpenGuardrails announces the AI Runtime Security Management System (AI-RSMS) โ€” an open, community-driven standard draft focused on securing AI systems during runtime.

OpenGuardrails Team8 min read
Read article