promptmap is an automated security scanner for custom LLM applications that focuses on prompt injection and related attack classes. The project supports both white-box and black-box testing, which means it can either run tests directly against a known model and system prompt configuration or attack an external HTTP endpoint without internal access. Its scanning workflow uses a dual-LLM architecture in which one model acts as the target being tested and another acts as a controller that evaluates whether an attack succeeded. The repository emphasizes broad coverage, including test rules for prompt stealing, jailbreaks, harmful content generation, hate-related outputs, social bias, and distraction attacks. It also supports multiple providers such as OpenAI, Anthropic, Google, xAI, and open-source models through Ollama, making it flexible for both commercial and local deployments.
Features
- White-box and black-box testing modes
- Dual-LLM architecture for attack execution and evaluation
- Support for OpenAI, Anthropic, Google, xAI, and Ollama models
- More than 50 prebuilt rules across 6 security categories
- YAML-based customizable rules and pass-fail conditions
- HTTP endpoint scanning for external LLM applications