AI vendor disclosures

AI disclosures

Critical for AI vendors: which models we use, how customer data stays out of training sets, what risks we own, and how we govern agent autonomy.

Customer data does not train foundation models.

This is contractually enforced with every LLM provider in our stack: Anthropic (Commercial Terms §3, zero retention), OpenAI (Enterprise Compliance API zero retention), Google Cloud Vertex AI (GCP Platform Terms — no training on customer data), and Microsoft Azure OpenAI (Products and Services DPA). Claresia itself trains no models on customer content; we are an orchestration layer.

Model inventory

Provider Model Residency
Anthropic Claude Sonnet 4.7 eu-south-1 / eu-central-1
Anthropic Claude Haiku 4.5 eu-south-1 / eu-central-1
OpenAI GPT-5 / GPT-5 mini us / eu region pinned
Google Gemini 2.5 Pro us / eu region pinned
Microsoft Azure OpenAI (GPT-5) eu-italy / eu-west / eu-north region pinned

AI risk taxonomy

The risks an AI orchestration platform is responsible for, and the controls Claresia ships against each. This is the basis for the Claresia Agent Governance Framework (planned formal release Q3 2026).

Prompt injection

Adversarial input that attempts to override system instructions or exfiltrate data.

Mitigation

Skill IR enforces parameter typing + structured tool calls. Output handlers sanitise before persist. Sensitive-data detector scans outputs for credentials, PHI, PCI markers. Outputs flagged through governance_event.

Hallucination

Model produces plausible but factually incorrect output.

Mitigation

Skills are grounded in retrieved context with citation provenance. Hub records every source used; Hub viewer shows the citation chain. Tier-1 skills require human-in-the-loop confirmation for irreversible actions.

Data exfiltration via model

Model returns data the user is not authorised to see.

Mitigation

Permission inheritance applied before retrieval — model only sees data the user has access to per the customer ACL. Connectors enforce ACLs at fetch time, not render time.

Agent autonomy / unsafe actions

Agentic skill takes destructive or irreversible action without human consent.

Mitigation

Alignment-model layer pre-scans every write action against a restricted-topics policy. Tier-1 actions (sending external email, modifying source code, posting publicly) always require explicit user approval. Audit trail in Hub.

Sensitive content leakage

Skill output contains credentials, PHI, PCI, or other sensitive content.

Mitigation

Sensitive-data detector runs on every output, applies redaction templates (one-click admin policies), surfaces matches as governance_event for SecOps review.

Agent governance framework

Pre-action

Alignment scan

Every agentic action is pre-scanned by an alignment model against tenant-configured restricted-topics policies. Blocked actions surface as governance_event for review.

Per-action

Tier-gated execution

Tier-3 (read-only): autonomous. Tier-2 (reversible writes): autonomous with audit. Tier-1 (irreversible / external impact): human approval required, captured in Hub.

Post-action

SOAR-ready audit

All governance events stream to customer SIEM / SOAR (Splunk, Sentinel, Tines) via the Audit Stream API. SOC analysts can triage from existing tools — Claresia does not require a new pane of glass.

Status — the framework operates today on the cc-052 / cc-053 Synthetic Twin runtime and within Mode B/C control planes. Formal published framework specification (Claresia equivalent of Glean AWARE) is targeted Q3 2026, alongside the public Bugcrowd / HackerOne program launch.