← Back to Catalog

ZeroLeaks/zeroleaks

↗ GitHub

AI Security Scanner - Test your AI systems for prompt injection and extraction vulnerabilities

539

Stars

86

Forks

9

Watchers

0

Open Issues

TypeScript·Other·Last commit Feb 21, 2026·by @ZeroLeaks·Published April 1, 2026·Analyzed 6d ago
A

Safety Rating A

No hardcoded secrets, malicious code patterns, or dependency vulnerabilities are evident from the repository content. The project is a legitimate, research-backed security testing tool for LLM systems analogous to penetration testing frameworks. The README contains no prompt injection attempts targeting analysts. The attack techniques documented are standard academic and CVE-referenced methods used in defensive security research. The dual-use nature (simulating attacks) is intentional and clearly scoped to authorized testing of one's own systems.

AI-assisted review, not a professional security audit.

AI Analysis

ZeroLeaks is an autonomous AI security scanner built in TypeScript that tests LLM-based systems for prompt injection and system prompt extraction vulnerabilities. It uses a multi-agent architecture (Strategist, Attacker, Evaluator, Mutator, Inspector, Orchestrator) and implements research-backed attack techniques such as Tree of Attacks with Pruning (TAP), Crescendo, Many-Shot, Chain-of-Thought Hijacking, Policy Puppetry, and TombRaider patterns to simulate real-world adversarial attacks against AI systems.

Use Cases

  • Testing LLM applications for system prompt extraction vulnerabilities before deployment
  • Performing red-team assessments on AI chatbots and assistants
  • Integrating automated prompt injection testing into CI/CD pipelines
  • Defense fingerprinting to identify specific guardrail systems in production AI
  • Researching and benchmarking LLM security posture using standardized attack techniques

Tags

#security#llm#ai-agents#multi-agent#testing#cli-tool#library#prompt-management#evaluation

Project Connections