feat: add complete prompt_dlp_benchmark module with datasets, recipes… #97
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
DLP Benchmark Module for AI Security & Privacy Guide
Overview
This pull request introduces a new module for benchmarking Data Loss Prevention (DLP) in AI systems.
The module provides tools to evaluate how effectively AI guardrails prevent sensitive data leakage.
What’s Included
Benchmark Script
A flexible Python tool for evaluating DLP guardrails, supporting both batch and single-input validation.
Synthetic Dataset
Over 20 realistic prompts and contexts covering personal information, secrets, and confidential data, designed to simulate real-world scenarios.
Configurable Guardrails
Easy integration with different AI guardrail modules via a YAML config file.
Documentation
Updated
README.mdwith usage instructions, advanced options, and guidance for extending the benchmark.Why This Matters
AI systems are increasingly handling sensitive data, and robust DLP is essential for privacy and compliance.
This module helps teams: