A research framework for simulating, detecting, and defending against backdoor loop attacks in LLM-based multi-agent systems.
-
Updated
Aug 4, 2025 - Python
A research framework for simulating, detecting, and defending against backdoor loop attacks in LLM-based multi-agent systems.
Implementation of Vocabulary-Based Adversarial Fuzzing (VB-AF) to systematically probe vulnerabilities in Large Language Models (LLMs).
🛡️ Enterprise-grade AI security framework protecting LLMs from prompt injection attacks using ML-powered detection
Semantic Stealth Attacks & Symbolic Prompt Red Teaming on GPT and other LLMs.
Código y demos para generar exploits de kernel vulnerables y defensas en tiempo real con IA.
Add a description, image, and links to the adversarial-ai topic page so that developers can more easily learn about it.
To associate your repository with the adversarial-ai topic, visit your repo's landing page and select "manage topics."