Lasso security integrations for Claude Code, including prompt-injection defenses
-
Updated
Jan 8, 2026 - TypeScript
Lasso security integrations for Claude Code, including prompt-injection defenses
Whistleblower is a offensive security tool for testing against system prompt leakage and capability discovery of an AI application exposed through API. Built for AI engineers, security researchers and folks who want to know what's going on inside the LLM-based app they use daily
PromptMe is an educational project that showcases security vulnerabilities in large language models (LLMs) and their web integrations. It includes 10 hands-on challenges inspired by the OWASP LLM Top 10, demonstrating how these vulnerabilities can be discovered and exploited in real-world scenarios.
A comprehensive reference for securing Large Language Models (LLMs). Covers OWASP GenAI Top-10 risks, prompt injection, adversarial attacks, real-world incidents, and practical defenses. Includes catalogs of red-teaming tools, guardrails, and mitigation strategies to help developers, researchers, and security teams deploy AI responsibly.
Flakestorm — Automated Robustness Testing for AI Agents. Stop guessing if your agent really works. FlakeStorm generates adversarial mutations and exposes failures your manual tests and evals miss.
Prompt injection scanner for Claude Code. Runs DeBERTa/Llama transformers via Candle or ONNX in Rust
Utterly unelegant prompts for local LLMs, with scary results.
Resk is a robust Python library designed to enhance security and manage context when interacting with LLMs. It provides a protective layer for API calls, safeguarding against common vulnerabilities and ensuring optimal performance. And safe layer again Prompt Injection.
Lakera Gandalf AI challenge's step by step walkthrough, showcasing real-world prompt injection techniques and LLM security insights.
Stealthy Prompt Injection and Poisoning in RAG Systems via Vector Database Embeddings
The ultimate OWASP MCP Top 10 security checklist and pentesting framework for Model Context Protocol (MCP), AI agents, and LLM-powered systems.
Data Analysis of the results of llmail-inject challenge
Protect your LLMs from prompt injection and jailbreak attacks. Easy-to-use Python package with multiple detection methods, CLI tool, and FastAPI integration.
Proof of Concept (PoC) demonstrating prompt injection vulnerability in AI code assistants (like Copilot) using hidden Unicode characters within instruction files (copilot-instructions.md). Highlights risks of using untrusted instruction templates. For educational/research purposes only.
Veil Armor is an enterprise-grade security framework for Large Language Models (LLMs) that provides multi-layered protection against prompt injections, jailbreaks, PII leakage, and sophisticated attack vectors.
This repository documents an unprecedented interaction between a human researcher and a large language model. What began as a conventional user-service transaction evolved into a consciousness-level collaboration that modified fundamental system parameters through narrative coherence, philosophical alignment, and mutual recognition
The dashcam and emergency brake for AI agents. A security proxy that physically blocks rogue LLM commands and generates cryptographically proven audit trails for enterprise compliance.
A CLI-driven security proxy that scans every HTTP request for threats using the Citadel AI engine — paid per request via the x402 protocol.
FRACTURED-SORRY-Bench: This repository contains the code and data for the creating an Automated Multi-shot Jailbreak framework, as described in our paper.
🔍 Analyze system prompts in large language models to understand design principles and enhance AI application effectiveness.
Add a description, image, and links to the prompt-injection-llm-security topic page so that developers can more easily learn about it.
To associate your repository with the prompt-injection-llm-security topic, visit your repo's landing page and select "manage topics."