Govern your LLMs, Agents, and MCP Servers across the enterprise.

AI Warden is the platform responsible organisations use to secure, govern, and control their AI landscape — so developer and product teams can ship faster with guardrails already in place.

SOC 2 Type II in progress Self-hosted or SaaS Read-only by default

Built for security, risk, and platform teams across regulated markets

  • Financial institutions
  • Insurance organisations
  • Healthcare providers
  • Regulated SaaS companies
  • Public-sector services

The problem

AI is everywhere. Governance is nowhere.

Every team is shipping with AI. Most security teams cannot tell you which models, which agents, or which MCP servers are running today — let alone what data they touch, how much they cost, or whether they would survive a regulator asking for evidence.

Shadow AI & sprawl

Teams adopt LLM APIs, embed copilots, and stand up MCP servers without a central record. By the time security finds out, the data is already flowing out the door.

Read more

Leaked keys, exploding bills

Provider keys land in source repos, terminals, and CI logs. One unbounded loop becomes a six-figure invoice overnight. Procurement is the first to notice.

See the threat model

Regulators are catching up

EU AI Act, NIST AI RMF, ISO 42001, sector regulators. The question stops being "do you use AI?" and becomes "show me your inventory, your controls, and your evidence."

The regulatory landscape

The platform

One control plane. Four surfaces.

AI Warden gives you a single place to set policy, watch every request, enforce in real time, and produce the evidence to prove it. Each surface is independently useful; together they close the loop.

01 — LLM Gateway

Every model call. Cost-bound, key-safe, content-aware.

Route every prompt and completion through one egress. Hold provider keys server-side. Enforce per-team budgets, prompt-injection scanners, PII redaction, and content policies — before the request ever leaves your network.

  • Hard cost ceilings per team, project, model. Alerts at 80%, blocks at 100%.
  • Provider keys never leave the gateway. Clients use a short-lived PAT.
  • Prompt-injection & secret scanners on the request and the response.
Explore the LLM Gateway
before · client.pykey in source
import openai
openai.api_key = "sk-prod-7f2a…"  # committed by mistake
openai.chat.completions.create(model="gpt-4o", ...)
after · client.pypolicy enforced
import openai
openai.base_url = "https://gw.aiwarden.io/v1"
openai.api_key  = os.environ["AIW_PAT"]   # short-lived, scoped
openai.chat.completions.create(model="gpt-4o", ...)
# gateway: enforces policy, redacts PII, charges your team budget

02 — MCP Fleet

Every MCP server. Inventoried, scanned, supervised.

The MCP servers connecting your AI to email, code, finance, CRM are the new attack surface. AI Warden registers them, scans every request and response in real time, and gives you a kill-switch when something looks wrong.

  • Real-time request log with SQLi, secrets, PII, and custom rule scanners.
  • Hosted sandboxes that quarantine risky tools without breaking development.
  • Quotas, IP allowlists, four-eyes approvals per server, per principal.
Explore MCP Fleet
09SQL injection
403blocked at gateway
auditsigned & archived

03 — Governance & Compliance

Policies your auditor can read. Evidence you can export.

Set policies once and apply them to every model, agent, and connector. Map controls to EU AI Act, NIST AI RMF, ISO 42001 and your internal frameworks. Regenerate the evidence pack any time a regulator asks.

  • Versioned policies with diffs, blast-radius preview, and four-eyes approval.
  • Immutable, signed audit log streaming to ClickHouse — query in seconds.
  • One-click evidence packs mapping your controls to the framework of choice.
Explore Governance
Policy diff pol-1287 · v4 → v5 awaiting 2nd approval
# Production LLM egress
  budget_per_team_usd:   8000
+ budget_per_team_usd:   12000
  models_allowed:
+   - "openai:gpt-4o-mini"
- redact_pii:           false
+ redact_pii:            true
  blast_radius:          142 teams · 3,401 agents

04 — Agents & Copilots

Identity for the things that aren't people.

Agents and copilots run continuously, on your behalf, often as system principals. Treat them like employees: identity-rooted, scoped, time-bounded, and revoked when they leave. AI Warden gives every agent its own identity — and a policy you can audit.

  • Per-agent identity rooted in your IDP — Keycloak, Entra, Okta.
  • Per-agent budget & scope: which models, which MCP servers, which data.
  • Behavioural baselines alert when an agent strays from its envelope.
Explore Agent governance
Agent credit-risk-summariser within envelope
Identity
system-principal · sp-92ab @ corp-idp
Models
openai:gpt-4o-mini · anthropic:haiku
MCP scope
finance.read · risk.read · no .write
Budget · 30d
$1,420 / $3,000
Last anomaly
none in 28 days
Owner
risk-platform@corp

Outcomes

What it looks like, week one to quarter one.

Near 100% routed LLM traffic visible in week one

Drop-in OpenAI / Anthropic / Azure-OpenAI compatibility. No client rewrites required.

10-30% month-one spend reduction (common range)

Best estimate from early rollouts. Results vary by workload shape, cache hit rate, and model mix.

0 provider keys on developer machines

Keys live server-side. Clients hold a short-lived, scoped, revocable PAT.

< 1d to regenerate an evidence pack

Map controls once. Export the framework of choice on demand.

Example outcome from a recent banking engagement: nine teams were using LLMs through six routes with unreconciled spend. Within a month, traffic moved to one gateway, keys were centralised server-side, and regulator evidence no longer required a war room.
Illustrative customer case, anonymised

Under the hood

Built for the way real enterprises run.

Self-hosted or SaaS. Keycloak / Entra / Okta on the front, Postgres and ClickHouse on the back, an OpenAPI-described control plane in the middle. No black boxes. No agent on every laptop.

Self-hostable

One VM, your network, your keys, your data. Or a managed instance if you'd rather we operate it. Same code path either way.

Identity-rooted

SSO via OIDC into your IDP. Every actor — human, service principal, agent — has a real identity behind every audit row.

Open standards

OpenAI-compatible egress. OpenAPI-described control plane. OTel, signed audit, S3 / SIEM sinks. No proprietary SDK to adopt.

Read-only by default

Drop AI Warden in front of existing systems and start observing. Enforcement is a deliberate, auditable action — never the default.

Composable

Use the LLM Gateway alone. Or just the MCP Fleet. Or the whole platform. Surfaces share data; you don't have to share scope.

Battle-tested patterns

Built by engineers from the regulated-finance, healthcare, and identity worlds. Patterns chosen because they survive a real audit, not because they look good on a slide.

On watch

This is the Warden.

Every request, every key, every agent — under one steady gaze. The platform takes the name seriously: read-only by default, deliberate when it acts, and never asleep at the post.

Take the next step

See AI Warden against your traffic.

A 45-minute working session with our team. Bring one team, one model provider, and one MCP server. Leave with a working gateway, a real policy, and a draft evidence pack.