Agentic LLM Vulnerability Scanner / AI red teaming kit 🧪
-
Updated
Feb 3, 2026 - Python
Agentic LLM Vulnerability Scanner / AI red teaming kit 🧪
DeepTeam is a framework to red team LLMs and LLM systems.
The fastest Trust Layer for AI Agents
Ultra-fast, low latency LLM prompt injection/jailbreak detection ⛓️
The Open Source Firewall for LLMs. A self-hosted gateway to secure and control AI applications with powerful guardrails.
A TypeScript library providing a set of guards for LLM (Large Language Model) applications
LMAP (large language model mapper) is like NMAP for LLM, is an LLM Vulnerability Scanner and Zero-day Vulnerability Fuzzer.
Offical repository for NeurIPS 2025 paper "From Judgment to Interference: Early Stopping LLM Harmful Outputs via Streaming Content Monitoring".
LLM prompt injection detection for Go applications
Runtime-secured AI tooling framework for production-grade LLM applications, protecting against prompt injection, jailbreaks, and adversarial attacks.
Engineered to help red teams and penetration testers exploit large language model AI solutions vulnerabilities.
Veil Armor is an enterprise-grade security framework for Large Language Models (LLMs) that provides multi-layered protection against prompt injections, jailbreaks, PII leakage, and sophisticated attack vectors.
Zero-dependency Claude Code plugin that catches speculation, invented causality, and fake citations before they pollute your context. Install in one command, works offline, no API keys needed.
Exposing Jailbreak Vulnerabilities in LLM Applications with ARTKIT
OpenClaw plugin for Prisma AIRS from Palo Alto Networks
User prompt attack detection system
Runtime guardrails for AI agents that enforce token budgets, loop limits, and tool rate limits locally.
CLI tool for testing production safety controls in LLM/RAG apps - prompt injection, data leakage, hallucinations, cost vulnerabilities
Example of running last_layer with FastAPI on vercel
An OpenAI-compatible reverse proxy you run yourself. It gives you the features of an AI gateway (guardrails, budgets, rate limits, multi-provider routing) but under your control from your client.
Add a description, image, and links to the llm-guardrails topic page so that developers can more easily learn about it.
To associate your repository with the llm-guardrails topic, visit your repo's landing page and select "manage topics."