ThemissingcontrollayerforproductionAI.
What Vault did for secrets, OGuardAI does for AI data. Detect, tokenize, transform, restore, revoke — every step controlled. Your LLMs work normally. Your data stays safe.
Not another model. Not another redaction script. Not another compliance checkbox. OGuardAI is infrastructure — the runtime layer every LLM stack is missing.
Apache-2.0 · Self-hosted · Provider-neutral
The Data Flow
If your AI sees raw PII, you have already lost control. Watch how OGuardAI prevents that at every step.
Swipe to see all stages
Works Everywhere in Your Stack
RAG Pipelines
Documents ingested with tokens only. Queries and context protected end-to-end.
Agentic Workflows
Each tool call sanitized independently. Per-step policy enforcement.
Chat & Support
German formal replies with gender-aware restoration. Multi-language ready.
Open source. Apache-2.0. Self-hosted.
How It Works
Your App Sends Text
User input containing sensitive data enters the pipeline.
OGuardAI Detects + Tokenizes
PII is identified and replaced with semantic tokens.
LLM Receives Only Tokens
The language model sees safe tokens, never real data.
LLM Responds with Tokens
The model generates output preserving token placeholders.
OGuardAI Restores Originals
Tokens are deterministically replaced with real values.
Detection Technology
Detection Technology
Multi-layered detection combining Rust-native performance with state-of-the-art NLP
Rust Regex Engine
- 30 regex patterns for 16 entity types
- Sub-millisecond detection (p50: 0.8ms)
- Zero external dependencies
- Deterministic: same input = same output
NER / NLP Models
- Person, company, location detection
- GLiNER zero-shot NER model
- spaCy pipeline (optional alternative)
- 30+ language support
- Confidence scoring with 0.65 threshold
Post-Processing
- 3-stage token repair (strict > repair > fuzzy)
- Output guard second-pass detection
- Entity linking (sentence-level + JSON structural)
- Overlap resolution and address merging
Built with
Features
Semantic Tokens
Typed tokens like {{email:e_001}} carry context so LLMs generate correct output.
6 Restore Modes
Full, partial, masked, formatted, abstract, or none — per entity type and channel.
Policy Engine
YAML-driven policies control which entities are masked, passed through, or blocked.
Output Guard
Second-pass scan catches any PII the LLM may have generated in its response.
Token Repair
Three-stage repair (strict, repair, fuzzy) handles malformed tokens from any LLM.
Streaming (SSE)
Transform and rehydrate in real-time over server-sent events for chat UIs.
Entity Revocation
Revoke specific tokens mid-session so restored values are never returned again.
RAG Pipeline
Sanitize document ingestion and query-time context for retrieval-augmented generation.
Multi-Language
Detect and protect PII across languages with gender and formality awareness.
Not another redaction tool
Most tools stop at detection. OGuardAI controls the full lifecycle.
PII detection is solved. What's missing is a runtime layer that protects data flowing to and from LLMs — with reversible tokens, policy enforcement, and identity-level control.
Others detect data. OGuardAI controls how data flows through AI systems — open source, self-hosted, built for production.
View on GitHubEnterprise-Grade Compliance
GDPR by Architecture
Art 17 right to erasure built into the runtime, not bolted on as an afterthought.
HIPAA Ready
PHI protection with audit trails and granular access controls built in.
SOC 2 Aligned
AES encryption with access controls and continuous monitoring support.
Self-Hosted
Your infrastructure, your data. Zero third-party dependency required.
Key Guarantees
- Raw PII never leaves your infrastructure
- AES-256-GCM encrypted sessions
- HMAC-SHA-256 revocation (no PII stored)
- Full audit trail with trace IDs
- Fail-secure: unprotected text is never returned
Drop-In Integration
from openai import OpenAI
client = OpenAI(
base_url="http://localhost:8081/v1"
)
# Standard OpenAI usage. OGuardAI masks PII automatically.|One line change. No code rewrite needed.
Built by Oronts
OGuardAI is developed by Oronts, a technology company specializing in AI infrastructure and data protection systems for enterprises and government organizations.
Ready to protect your AI pipeline?
Get running in seconds with a single command.
$ |