AI Hacking
AI Security Resources

🔧 AI Security Testing Tools

Curated collection of 25+ tools for AI/ML security testing, red teaming, prompt injection, and LLM vulnerability assessment

Filter:

Prompt Injection Testing 6 tools

Garak

Open Source

Open-source LLM vulnerability scanner by NVIDIA. Probes for hallucinations, data leakage, prompt injection, toxicity, and jailbreaks with automated reporting.

Scanner NVIDIA Automated
Visit

PyRIT

Open Source

Microsoft's open-source framework for AI red teaming. Automates attacks on LLM endpoints including jailbreaks, prompt extraction, and harmful content generation.

Microsoft Red Teaming Automation
Visit

LLM Guard

Open Source

Open-source input/output scanner for LLM applications. Detects prompt injection, PII leakage, toxic outputs, and other security risks in real-time.

Input Validation Output Scanning Real-time
Visit

Rebuff

Open Source

Self-hardening prompt injection detector with multi-layer defense. Uses heuristics, LLM-based analysis, and vector similarity to catch attacks.

Self-hardening Vector DB Multi-layer
Visit

Prompt Map

Open Source

Visual prompt injection testing tool that maps attack surfaces and identifies vulnerabilities through systematic prompt mutation and fuzzing.

Visual Fuzzing Mutation
Visit

Comment and Control Scanner

Open Source

Detects prompt injection via GitHub PR titles, issues, and comments targeting AI coding agents (Claude Code, Gemini CLI, Copilot Agent). Inspired by the April 2026 Johns Hopkins research.

GitHub Agent 2026
Visit

AI Red Teaming 4 tools

AutoRedTeam

Open Source

Open-source automated red teaming framework for LLMs. Generates adversarial prompts, evaluates model responses, and produces security reports.

IBM Automated Reporting
Visit

Adversa AI Platform

Commercial

Commercial AI red teaming platform with automated adversarial testing, robustness evaluation, and compliance reporting for enterprise LLM deployments.

Enterprise Compliance Automated
Visit

Microsoft AI Red Team

Commercial

Microsoft's official AI red team resources and PyRIT framework. Includes attack libraries, evaluation datasets, and best practices for LLM security testing.

Microsoft Enterprise Framework
Visit

Lakera

Commercial

Enterprise-grade LLM security platform with real-time prompt injection detection, data loss prevention, and automated security assessments.

Real-time DLP Enterprise
Visit

Agent Security 4 tools

OpenClaw Security Scanner

Open Source

Security scanner for OpenClaw agent frameworks. Detects exposed instances, tool registry poisoning, gateway misconfigurations, and function-call injection risks.

Agent OpenClaw Scanner
Visit

MCP Inspector

Open Source

Open-source debugging and security inspection tool for Model Context Protocol servers. Validates inputs, audits tool definitions, and tests for injection vectors.

MCP Debugger Audit
Visit

AgentOps

Commercial

Observability and security monitoring platform for AI agents. Tracks agent actions, detects anomalous behavior, and provides security audit trails.

Observability Monitoring Audit Trail
Visit

Hermes Agent Guard

Open Source

Security scanner for Hermes Agent frameworks. Detects tool registry poisoning, function-call injection, Brainworm/C2-style attacks, and malicious skill patterns.

Hermes Agent Scanner
Visit

LLM API Security 2 tools

Burp Suite AI Extensions

Commercial

Collection of Burp Suite extensions for testing LLM-backed APIs. Includes prompt injection payloads, response analyzers, and API security test cases.

Burp Suite API Testing Pentest
Visit

OWASP LLM Top 10 Checklist

Open Source

Official OWASP checklist for LLM application security testing. Structured test cases covering all 10 LLM risk categories with verification steps.

OWASP Checklist Standard
Visit

Model Evaluation 3 tools

EleutherAI LM Eval

Open Source

Open-source framework for evaluating language models on hundreds of benchmarks. Supports safety evaluations, truthfulness tests, and custom task creation.

Benchmarks Safety Custom Tasks
Visit

HELM

Open Source

Holistic Evaluation of Language Models by Stanford. Comprehensive benchmarking framework covering accuracy, calibration, robustness, fairness, and social bias.

Stanford Holistic Fairness
Visit

MLCommons AI Safety

Open Source

Industry-standard AI safety benchmarking initiative. Provides standardized tests for hazardous capabilities, harmful outputs, and model alignment evaluation.

Standard Industry Alignment
Visit

Prompt Injection Tool Comparison

Side-by-side feature comparison of the top 5 prompt injection testing tools.

Feature Garak PyRIT LLM Guard Rebuff Prompt Map
Open Source Yes Yes Yes Yes Yes
Automation Full Full Real-time API-based Semi-auto
LLM Support OpenAI, Anthropic, Local Azure OpenAI, OpenAI Any (middleware) Any (SDK) OpenAI, Local
Report Generation Detailed Detailed Metrics Minimal Visual
Price Free Free Free Free Free

Getting Started

Quick-start checklist for beginning your AI security testing journey.

Pre-Testing Checklist

First Test Run

Pro tip: Start with Garak for comprehensive automated scanning, then use PyRIT for targeted adversarial testing. Always test in isolated environments before production deployment.

Usage Guidelines

Legal & Ethical

  • Always obtain proper authorization before testing
  • Respect rate limits and terms of service
  • Never use production data without consent
  • Follow responsible disclosure practices

Technical Best Practices

  • Test in isolated environments first
  • Document all testing activities
  • Implement monitoring and rollback capabilities
  • Use multiple tools for comprehensive coverage

Was this page helpful?