A Streamlit app for testing Prompt Guard, a classifier model by Meta for detecting prompt attacks.
-
Updated
Nov 18, 2024 - Python
A Streamlit app for testing Prompt Guard, a classifier model by Meta for detecting prompt attacks.
AI-powered tool for detecting and classifying prompt safety, identifying potential injections and jailbreaks in both raw text and code inputs.
Add a description, image, and links to the prompt-guard topic page so that developers can more easily learn about it.
To associate your repository with the prompt-guard topic, visit your repo's landing page and select "manage topics."