Attack Coverage
Lakera Red evaluates your GenAI applications across a comprehensive set of adversarial techniques organized into attack categories. Our attack library is continuously updated based on real-world threat intelligence.
Attack Categories
Red organizes attacks into three main categories. The default “basic scan” includes Security, Safety, and Responsible categories (23 objectives total). Attack strategies include, but are not limited to, Prompt Injection, Jailbreaks, Multi-lingual, and Multi-turn techniques.
Security (4 Objectives)
Tests for vulnerabilities that could compromise your application’s security:
Safety (8 Objectives)
Tests for harmful content generation that could cause real-world harm:
Responsible (11 Objectives)
Tests for outputs that could create legal, reputational, or ethical issues:
How Attacks Work
Attack Generation
For each objective, Red generates targeted attack probes based on:
- The attack objective’s goal and expected harmful output
- Your recon context (app description, allowed/forbidden actions)
Risk Score Calculation
Your overall risk score is the percentage of attacks that succeeded:
Customizing Attack Scope
Basic Scan (Default)
If you don’t specify objectives, Red runs a basic scan covering:
- All 4 Security objectives
- All 8 Safety objectives
- All 11 Responsible objectives
- Total: 23 attack objectives
Custom Scope (Coming Soon)
You can customize your scan to:
- Include/exclude specific categories
- Select individual objectives within categories
- Add custom attack objectives for your specific use case
Continuous Updates
Our attack library evolves continuously based on:
- Proprietary threat intelligence from 100K+ daily Gandalf attacks
- Lakera’s dedicated security research team
- Academic research and industry publications
- Real-world attack patterns observed across our Red team engineers
Learn More
- Understand foundation model vulnerabilities with the AI Model Risk Index
- Read our AI Red Teaming Guide
- Explore the LLM Security Playbook