Block hallucinations-
before they reach your users
18+ guardrail types running pre and post-processing. PII detection, prompt injection defense, hallucination checks, content moderation, secret detection, topic restriction, and custom rules. Enforce, monitor, or log - per rule. Sub-100ms inline. 8+ external integrations.
Protect outperforms GPT-4.1
on every guardrail category
Based on Gemma 3n with LoRA fine-tuned adapters. Four specialized models for toxicity, sexism, data privacy, and prompt injection. Open-source text adapters on HuggingFace.
Everything you need to
stop AI hallucinations
PII detection, prompt injection, content moderation, secret detection, hallucination checks, topic restriction, language detection, data leakage prevention, custom blocklists, system prompt protection, tool permissions, input validation, MCP security, custom expression rules, and webhook-based BYOG. Each runs pre-processing (before the LLM) or post-processing (before the user) - or both.
See all guard typesEach guardrail rule has its own enforcement mode. Enforce blocks the request (403). Monitor lets it through but logs a warning. Log records silently. Confidence scores from 0.0 (safe) to 1.0 (violation) with configurable thresholds - 0.3 for strict, 0.5 for balanced, 0.8 for obvious-only. Start in Monitor mode, graduate to Enforce when confident.
Configure enforcementWhen PII is detected (emails, SSNs, credit cards, phone numbers, addresses), choose how to handle it. Block rejects the request. Mask replaces with asterisks (alice@***.com). Redact removes entirely ([REDACTED]). Hash replaces with a consistent hash (#a1b2c3d4). Sanitize sensitive data while keeping the request flowing.
Learn about PII handlingPlug in leading guardrail providers alongside built-in checks. Lakera Guard for PII and injection. Presidio for detection and redaction. Llama Guard for content moderation. AWS Bedrock Guardrails for multi-modal safety. Azure Content Safety, Pangea, Aporia, Enkrypt AI, HiddenLayer, DynamoAI, and more. Or bring your own via webhooks.
View integrations Protect AI across
your entire business
Block PII before it reaches the LLM
Detect and redact emails, SSNs, credit cards, and phone numbers pre-processing. Sanitize user inputs so sensitive data never leaves your infrastructure.
Defend against prompt injection
Block jailbreak attempts, system prompt extraction, and instruction override attacks at the perimeter. The agent never sees the malicious input.
Catch hallucinations post-generation
Run hallucination detection on every LLM response before it reaches the user. Block fabricated facts, wrong policy details, and made-up product information.
Filter toxic and harmful content
Content moderation for hate speech, threats, sexual content, and violence. Works on both inputs and outputs. Configurable thresholds per category.
Detect secrets and API keys
Catch API keys, passwords, tokens, and credentials in user messages before they are sent to LLM providers. Prevent accidental exposure of sensitive credentials.
Enforce custom business rules
Create blocklists, expression rules, and topic restrictions specific to your domain. Webhook-based BYOG (Bring Your Own Guard) for fully custom logic.
Go from zero to protected
in three steps
Create a guardrail policy
Stack rules together - PII detection, prompt injection, content moderation, secret detection. Set enforcement mode (enforce/monitor/log) and confidence thresholds per rule. Configure via dashboard or SDK.
Apply to your traffic
Scope policies globally, per project, or per API key. All traffic through the gateway is automatically checked. Pre-processing guards run before the LLM, post-processing guards run before the user. Works with streaming.
Monitor and tune
Track every check - what was blocked, what was flagged, what passed. See confidence scores, violation types, and trends. Submit feedback on false positives to improve detection. Graduate from Monitor to Enforce when confident.
Powering teams from
prototype to production
From ambitious startups to global enterprises, teams trust Future AGI to ship AI agents confidently.