AI Data Guardian
Safeguards sensitive information by redacting, allowlisting, and sanitizing data before it reaches large language models.
About
The AI Data Guardian acts as a crucial intermediary, meticulously processing and securing data before it is exposed to large language models. Designed with a multi-layered, zero-trust architecture, it automatically redacts Personally Identifiable Information (PII) and Payment Card Industry (PCI) data, enforces field allowlisting to prevent sensitive data leakage, and sanitizes free-text content from suspicious elements. Furthermore, it manages payload sizes and provides comprehensive audit logging, ensuring that LLM interactions are not only efficient but also compliant and secure, maintaining a strict `llm_view` for all data consumed by AI.
Key Features
- Payload Size Limits and Automatic Summarization
- Automatic PII/PCI Redaction and Masking
- 0 GitHub stars
- Field Allowlisting for Sensitive Data Control
- Free-Text Sanitization for Suspicious Content
- Comprehensive Audit Logging and Security Event Tracking
Use Cases
- Ensuring compliance with data privacy regulations (e.g., GDPR, HIPAA) in AI applications.
- Protecting sensitive customer or financial data when integrating LLMs into business processes.
- Controlling and monitoring data flow to LLMs to prevent data leakage and over-exposure.