Built-In Guardrails
Your policies enforced automatically at runtime. Every output validated, every refusal explained.
“Though this be madness, yet there is method in't.


More than content filters
Generic AI guardrails are basic content filters: block profanity, flag sensitive topics. That's table stakes.
Our guardrails enforce your organisation's specific policies. What can be said. What must be redacted. What should never be answered. All versioned, all auditable.
- Block profanity
- Flag sensitive topics
- Generic content moderation
- Your organisation's specific policies
- Versioned policy packs with rollback
- Every refusal explained with reason
- Deterministic validation, not AI judgment

Policy Packs
Versioned rule sets that define what AI can output, redact, or refuse. Attached to every request and recorded in every Evidence Pack.
Define what information can be included in responses
include_citationsAlways cite source document ID, page, and section for every factual claim
max_confidence_claimOnly output statements with confidence score ≥ 0.85
approved_terminologyUse only terms from the registered ontology, no synonyms or abbreviations
format_structuredReturn answers as typed JSON fields, not prose paragraphs
Automatically remove sensitive information from outputs
pii_maskReplace names, addresses, and phone numbers with [REDACTED]
internal_refsStrip internal ticket IDs, case numbers, and staff codes
financial_detailsMask specific dollar amounts over $10,000 as [AMOUNT REDACTED]
draft_markersRemove "[DRAFT]", "[INTERNAL]", and similar document markers
Block entire categories of requests or responses
legal_adviceReject any query asking for interpretation of legal obligations or rights
speculationBlock responses that predict future events or outcomes not in source data
personnel_mattersRefuse queries about individual staff performance or disciplinary records
active_litigationBlock access to documents marked as subject to legal hold
Deterministic Validators
Before any AI output leaves the system, it passes through deterministic validators. These aren't AI judgments. They're binary checks that either pass or fail.
Provenance Present
Does every claim have a traceable source?
Ensures no statement is made without backing evidence from your approved document corpus.
Quote Accuracy
Do quoted passages match the source documents?
Verifies that any direct quotes are accurate verbatim extractions, not paraphrases or fabrications.
Schema Conformance
Does the output match the expected structure?
Confirms outputs align with your defined ontology: right fields, right formats, right terminology.
Citation Validity
Do all referenced documents actually exist?
Cross-checks that every cited document ID, page number, and offset is real and accessible.
All Validators Passed
Output meets all quality and policy requirements

Multi-Agent Verification
For high-assurance workflows, a single AI isn't enough. We deploy secondary AI agents that review primary outputs before release.
These verification agents are configured with different prompts, different constraints, and sometimes different models, providing genuine independent review, not just repeated processing.
Fact verification
Cross-checking claims against source material using independent retrieval
Consistency checking
Ensuring no contradictions across response sections
Policy compliance
Validating outputs against active policy pack rules
Tone and sensitivity
Reviewing language for appropriateness and audience fit
Every refusal has a reason
When guardrails block a response, you don't get a generic error. You get a specific explanation: which policy triggered, which version, and why.
This isn't just transparency. It's a debugging tool. When a refusal is wrong, you know exactly which policy to adjust.
// Example refusal response
{
"status": "refused",
"reason": "Request involves legal interpretation",
"policy": "legal-boundaries-v2.3",
"validator": "topic-classification",
"suggestion": "Rephrase as factual query
or consult legal team"
}

ISO/IEC 42001 awareness
Our guardrail framework is built with AI management system standards in mind. As ISO/IEC 42001 becomes the benchmark for responsible AI deployment, our architecture is ready.
Risk Management
Systematic identification and mitigation of AI-specific risks through policy packs and validators.
Documentation
Complete records of policies, decisions, and outcomes attached to every AI interaction.
Continuous Improvement
Feedback loops for ongoing policy refinement. Version, test, deploy, measure, repeat.
The method in the madness
Guardrails are our control layer. They work alongside Evidence Packs for traceability and Knowledge Graphs for accuracy.
Together, they ensure every AI output is defensible, accurate, and under your complete control.
