This repository provides implementation to formalize and benchmark Prompt Injection attacks and defenses
-
Updated
Sep 5, 2024 - Python
This repository provides implementation to formalize and benchmark Prompt Injection attacks and defenses
The Prompt Injection Testing Tool is a Python script designed to assess the security of your AI system's prompt handling against a predefined list of user prompts commonly used for injection attacks. This tool utilizes the OpenAI GPT-3.5 model to generate responses to system-user prompt pairs and outputs the results to a CSV file for analysis.
Image Prompt Injection is a Python script that demonstrates how to embed a secret prompt within an image using steganography techniques. This hidden prompt can be later extracted by an AI system for analysis, enabling covert communication with AI models through images.
ASCII Art Prompt Injection is a novel approach to hacking AI assistants using ASCII art. This project leverages the distracting nature of ASCII art to bypass security measures and inject prompts into large language models, such as GPT-4, leading them to provide unintended or harmful responses.
LLM Security Project with Llama Guard
Manual Prompt Injection / Red Teaming Tool
Client SDK to send LLM interactions to Vibranium Dome
Prompt Engineering Tool for AI Models with cli prompt or api usage
LLM Security Platform Docs
FRACTURED-SORRY-Bench: This repository contains the code and data for the FRACTURED-SORRY-Bench framework, as described in our paper.
Add a description, image, and links to the prompt-injection-tool topic page so that developers can more easily learn about it.
To associate your repository with the prompt-injection-tool topic, visit your repo's landing page and select "manage topics."