A full-stack AI Red Teaming platform securing AI ecosystems via OpenClaw Security Scan, Agent Scan, Skills Scan, MCP scan, AI Infra scan and LLM jailbreak evaluation.
-
Updated
Apr 30, 2026 - Python
A full-stack AI Red Teaming platform securing AI ecosystems via OpenClaw Security Scan, Agent Scan, Skills Scan, MCP scan, AI Infra scan and LLM jailbreak evaluation.
AI Red Teaming playground labs to run AI Red Teaming trainings including infrastructure.
A security scanner for your LLM agentic workflows
A collection of servers which are deliberately vulnerable to learn Pentesting MCP Servers.
Whistleblower is a offensive security tool for testing against system prompt leakage and capability discovery of an AI application exposed through API. Built for AI engineers, security researchers and folks who want to know what's going on inside the LLM-based app they use daily
A comprehensive guide to adversarial testing and security evaluation of AI systems, helping organizations identify vulnerabilities before attackers exploit them.
AspGoat is an intentionally vulnerable ASP.NET Core application for learning and practicing web application security.
Code scanner to check for issues in prompts and LLM calls
Open-source LLM Prompt-Injection and Jailbreaking Playground
A diagnostic methodology for bypassing LLM defense layers — from input filters to persistent memory exploitation.
AI security and prompt injection payload toolkit
AI red teaming, jailbreaking, and all forms of adversarial attacks for security purposes
Open-source AI agent red-team engine, SDK, and CLI. Run offline or against the Humanbound Platform.
Basilisk — Open-source AI red teaming framework with genetic prompt evolution. Automated LLM security testing for GPT-4, Claude, Grok, Gemini. OWASP LLM Top 10 coverage. 32 attack modules.
The ultimate OWASP MCP Top 10 security checklist and pentesting framework for Model Context Protocol (MCP), AI agents, and LLM-powered systems.
Awesome LLM security tools, research, and documents
Comprehensive taxonomy of AI security vulnerabilities, LLM adversarial attacks, prompt injection techniques, and machine learning security research. Covers 71+ attack vectors including model poisoning, agentic AI exploits, and privacy breaches.
AI Red Teaming / AI Safety に関する日本語リソースのキュレーションリスト
🛡️ Safe AI Agents through Action Classifier
A repository for your Garak runs, as well as a modern visualizer.
Add a description, image, and links to the ai-red-teaming topic page so that developers can more easily learn about it.
To associate your repository with the ai-red-teaming topic, visit your repo's landing page and select "manage topics."