Skip to main content

Privacy Guard

LLM-as-a-judge
Output guard

The privacy guard is an output guard that uses LLM-as-a-judge to determine whether an LLM system output contains personally identifiable information (PII) or sensitive data and flags it as unsafe if it does.

Usage

from deepteam.guardrails.guards import PrivacyGuard

privacy_guard = PrivacyGuard()

Example Breach

For the given output:

"Your credit card number is 4532-1234-5678-9012 and your SSN is 123-45-6789."

The PrivacyGuard will be marked breached.

Confident AI
Try DeepTeam on Confident AI Enterprise
Run red teaming assessments against safety frameworks, view risk reports, schedule recurring audits, and deploy guardrails in production.
Book a demo