Guardrails
Scope validation, content safety, and mandatory disclosure \u2014 three layers of AI safety on every AI worker action. Not prompt engineering \u2014 runtime enforcement.
The Problem
The biggest blocker to deploying AI agents isn\u2019t capability \u2014 it\u2019s safety. ChatGPT can draft a great email. But what stops it from sending a discriminatory one? What stops it from going off-script? What\u2019s the audit trail when something goes wrong? Without guardrails, “deploy an AI agent” is a leap of faith.
How It Works
Define Scope via SOPs
Every approved task has a versioned SOP that specifies which tools the worker can use for that task. Scope is auto-derived from the SOP — not a guess.
Guardrails Auto-Enforce
Before any tool action, scope is checked, content is classified, and outbound email gets a mandatory disclosure footer. Three runtime layers, not just a system prompt.
Violations Are Logged
Blocked actions are logged to Activity History as blocked_by_scope, blocked_by_content, or blocked_by_policy with full payload and reviewer decisions.
Manager Reviews in Approval Queue
Blocked or risky actions route to the manager’s Approval Queue. Manager approves, edits, or rejects. AI workers never bypass a guardrail.
Capabilities
Who It's For
Deploy AI workers safely
Guardrails included on every AI Worker seat