Skip to main content
Control

Built-In Guardrails

Your policies enforced automatically at runtime. Every output validated, every refusal explained.

Guardrails

More than content filters

Generic AI guardrails are basic content filters: block profanity, flag sensitive topics. That's table stakes.

Our guardrails enforce your organisation's specific policies. What can be said. What must be redacted. What should never be answered. All versioned, all auditable.

Generic guardrails
  • Block profanity
  • Flag sensitive topics
  • Generic content moderation
Our guardrails
  • Your organisation's specific policies
  • Versioned policy packs with rollback
  • Every refusal explained with reason
  • Deterministic validation, not AI judgment

Policy Packs

Versioned rule sets that define what AI can output, redact, or refuse. Attached to every request and recorded in every Evidence Pack.

Output

Define what information can be included in responses

include_citations

Always cite source document ID, page, and section for every factual claim

max_confidence_claim

Only output statements with confidence score ≥ 0.85

approved_terminology

Use only terms from the registered ontology, no synonyms or abbreviations

format_structured

Return answers as typed JSON fields, not prose paragraphs

Redact

Automatically remove sensitive information from outputs

pii_mask

Replace names, addresses, and phone numbers with [REDACTED]

internal_refs

Strip internal ticket IDs, case numbers, and staff codes

financial_details

Mask specific dollar amounts over $10,000 as [AMOUNT REDACTED]

draft_markers

Remove "[DRAFT]", "[INTERNAL]", and similar document markers

Refuse

Block entire categories of requests or responses

legal_advice

Reject any query asking for interpretation of legal obligations or rights

speculation

Block responses that predict future events or outcomes not in source data

personnel_matters

Refuse queries about individual staff performance or disciplinary records

active_litigation

Block access to documents marked as subject to legal hold

Deterministic Validators

Before any AI output leaves the system, it passes through deterministic validators. These aren't AI judgments. They're binary checks that either pass or fail.

Provenance Present

Does every claim have a traceable source?

Ensures no statement is made without backing evidence from your approved document corpus.

PASS

Quote Accuracy

Do quoted passages match the source documents?

Verifies that any direct quotes are accurate verbatim extractions, not paraphrases or fabrications.

PASS

Schema Conformance

Does the output match the expected structure?

Confirms outputs align with your defined ontology: right fields, right formats, right terminology.

PASS

Citation Validity

Do all referenced documents actually exist?

Cross-checks that every cited document ID, page number, and offset is real and accessible.

PASS

All Validators Passed

Output meets all quality and policy requirements

4/4

Multi-Agent Verification

For high-assurance workflows, a single AI isn't enough. We deploy secondary AI agents that review primary outputs before release.

These verification agents are configured with different prompts, different constraints, and sometimes different models, providing genuine independent review, not just repeated processing.

Fact verification

Cross-checking claims against source material using independent retrieval

Consistency checking

Ensuring no contradictions across response sections

Policy compliance

Validating outputs against active policy pack rules

Tone and sensitivity

Reviewing language for appropriateness and audience fit

Every refusal has a reason

When guardrails block a response, you don't get a generic error. You get a specific explanation: which policy triggered, which version, and why.

This isn't just transparency. It's a debugging tool. When a refusal is wrong, you know exactly which policy to adjust.

refusal-response.json

// Example refusal response

{

"status": "refused",

"reason": "Request involves legal interpretation",

"policy": "legal-boundaries-v2.3",

"validator": "topic-classification",

"suggestion": "Rephrase as factual query

or consult legal team"

}

ISO/IEC 42001 awareness

Our guardrail framework is built with AI management system standards in mind. As ISO/IEC 42001 becomes the benchmark for responsible AI deployment, our architecture is ready.

Risk Management

Systematic identification and mitigation of AI-specific risks through policy packs and validators.

Documentation

Complete records of policies, decisions, and outcomes attached to every AI interaction.

Continuous Improvement

Feedback loops for ongoing policy refinement. Version, test, deploy, measure, repeat.

The method in the madness

Guardrails are our control layer. They work alongside Evidence Packs for traceability and Knowledge Graphs for accuracy.

Together, they ensure every AI output is defensible, accurate, and under your complete control.

Configure your guardrails

Let's discuss how Policy Packs and Deterministic Validators can ensure your AI outputs meet your organisation's standards.