Technical and policy boundaries that keep AI agents safe, legal, and aligned with organisational values.
Guardrails are the price of speed. Pay them up front, or pay them in incidents.
AI guardrails include: abort conditions on every Play, compliance checks embedded in execution workflows (like the WCAG example), data sovereignty rules (what data can agents access, where can it flow), cost limits (maximum spend per agent per loop), output quality gates (automated evaluation before anything reaches users), and ethical boundaries (what agents are not permitted to do regardless of instructions). The Integrity Steward owns the guardrail framework, and Guard agents enforce it in real time.
Guardrails should be versioned and reviewed as the system's autonomy expands.
Key principles
- Abort conditions on every Play.
- Compliance checks embedded in workflows.
- Audit traces on every agent action.
- Risk veto independent of Forum cadence.