Skip to content

Search is only available in production builds. Try building and previewing the site to test it out locally.

Guardrails

Guardrails are safety policies applied at the gateway level to protect against data leakage, injection attacks, and inappropriate content.

Detects and blocks sensitive data in prompts before they reach the AI.

Detects:

  • API keys and secrets (AWS, GCP, Stripe, etc.)
  • Database connection strings
  • Private keys and certificates
  • Custom patterns (regex-based)

Action: blocks the request and shows which patterns matched.

Identifies personally identifiable information in prompts.

Detects:

  • Email addresses
  • Phone numbers
  • Social Security numbers
  • Credit card numbers
  • Custom PII patterns

Action: configurable — block, warn, or redact.

Detects attempts to override system prompts or inject malicious instructions.

Detects:

  • Common injection patterns (“ignore previous instructions”)
  • Role-playing attacks (“you are now…”)
  • Encoding-based evasion attempts

Action: blocks the request.

Filters requests and responses for inappropriate or off-topic content.

Detects:

  • Content outside the coding domain
  • Harmful or offensive requests
  • Policy-violating responses

Action: configurable — block or flag for review.

Configure guardrails from the dashboard:

  1. Navigate to Guardrails
  2. Enable/disable individual guardrails
  3. Set action for each (block, warn, or log-only)
  4. Add custom patterns for DLP or PII
  5. Save

All guardrail triggers are logged:

  • Timestamp and developer
  • Which guardrail fired
  • The pattern or rule that matched
  • Action taken (blocked, warned, logged)
  • Request summary (truncated for privacy)