🐢 Open-Source Evaluation & Testing library for LLM Agents
-
Updated
Mar 20, 2026 - Python
🐢 Open-Source Evaluation & Testing library for LLM Agents
The Python Risk Identification Tool for generative AI (PyRIT) is an open source framework built to empower security professionals and engineers to proactively identify risks in generative AI systems.
A full-stack AI Red Teaming platform securing AI ecosystems via OpenClaw Security Scan, Agent Scan, Skills Scan, MCP scan, AI Infra scan and LLM jailbreak evaluation.
AI Red Teaming playground labs to run AI Red Teaming trainings including infrastructure.
Agentic LLM Vulnerability Scanner / AI red teaming kit 🧪
A powerful tool for automated LLM fuzzing. It is designed to help developers and security researchers identify and mitigate potential jailbreaks in their LLM APIs.
An offensive/defense security toolset for discovery, recon and ethical assessment of AI Agents
AI Red Teaming Range
AI Security Platform: Defense (61 Rust engines + Micro-Model Swarm) + Offense (39K+ payloads)
AspGoat is an intentionally vulnerable ASP.NET Core application for learning and practicing web application security.
A comprehensive guide to adversarial testing and security evaluation of AI systems, helping organizations identify vulnerabilities before attackers exploit them.
LMAP (large language model mapper) is like NMAP for LLM, is an LLM Vulnerability Scanner and Zero-day Vulnerability Fuzzer.
🤖🛡️🔍🔒🔑 Tiny package designed to support red teams and penetration testers in exploiting large language model AI solutions.
Open-Source autonomous security operations and red teaming agent built to help defenders investigate threats, analyze vulnerabilities, assess indicators of compromise, generate hardening guidance, and execute security research through an auditable agent workflow.
Complete 90-day learning path for AI security: ML fundamentals → LLM internals → AI threats → Detection engineering. Built from first principles with NumPy implementations, Jupyter notebooks, and production-ready detection systems.
This is my prompts for Lakera's Gandalf challenges
Objective-driven adversarial testing framework for GenAI systems aligned with OWASP GenAI Top 10 risks.
🤖 Test and secure AI systems with advanced techniques for Large Language Models, including jailbreaks and automated vulnerability scanners.
AAPP‑MART is an AI-Autonomous Attack Path Prediction and Multi‑Agent Red Team Simulation Engine designed for automated threat modeling, adversary emulation, and enterprise‑grade cybersecurity validation aligned with MITRE ATT&CK.
An open source plugin for enabeling claude to gain offensive pentesting capabilities
Add a description, image, and links to the ai-red-team topic page so that developers can more easily learn about it.
To associate your repository with the ai-red-team topic, visit your repo's landing page and select "manage topics."