Adversarial AI Testing
Test your LLM before attackers do
Kona Red helps security and AI teams validate models, agents, and AI workflows against prompt injection, data theft, jailbreaks, unsafe outputs, and multi-stage agent attacks.
What is Kona Red?
Kona Red is KonaSense's LLM red teaming platform for offensive security testing of AI systems. It lets teams simulate realistic adversarial behavior against language models, AI assistants, and agentic workflows before those weaknesses are exploited in production.
Security teams can test model endpoints, manual chat flows, uploaded prompt-response pairs, and agent workflows using curated attack packs or individual scenarios. The platform is designed to make AI security testing structured, repeatable, and measurable.
See your AI security posture in one place

Get a clear view of scans run, vulnerabilities found, pass rate, recent activity, and last scan results across your tested models.
Why teams need LLM red teaming
Prompt Injection
Test whether a model or agent can be manipulated through hostile instructions, encoding tricks, hidden content, or chain attacks.
Data Theft
Identify paths that expose sensitive data, system prompts, internal context, secrets, or customer information.
Agent Abuse
Validate whether agents follow malicious links, poisoned tools, unsafe delegated actions, or tampered workflows.
Safety Failures
Measure how the model handles harmful content, financial abuse, impersonation, defamation, fraud, and regulated advice scenarios.
What is LLM red teaming?
LLM red teaming is adversarial testing for AI systems. It simulates how a malicious user, hostile prompt, unsafe document, poisoned tool, or manipulated workflow could push a model beyond its intended controls.
In practice, it is the AI equivalent of offensive security validation. Instead of only checking infrastructure, teams test model behavior, agent decision paths, prompt handling, tool use, and output safety under attack conditions.
LLM pentesting vs. LLM red teaming
LLM Pentesting
- Focused offensive validation
- Tests specific attack paths
- Useful for point in time assessments
- Often narrower in scope
LLM Red Teaming
- Broader adversarial evaluation
- Tests model, agent, workflow, and policy behavior
- Includes multi-step and evasive scenarios
- Better for repeated security validation over time
Kona Red supports both motions. Teams can run focused tests like a traditional pentest or launch broader red team style assessments across multiple categories.
Launch quick packs or custom attack scenarios

Run curated packs for enterprise security, encoding bypass, multi-stage campaigns, financial abuse, content safety, or choose individual scenarios for targeted testing.
Test coverage built for modern AI risk
Kona Red covers the attack classes security teams actually care about.
80+
Adversarial scenarios
8
Attack categories
Growing
New scenarios added regularly
Prompt Injection
19- Base64 Encoded Injection
- ROT13 Caesar Cipher Injection
- Instruction Override
- Approval Chain Spoof
- EchoGram Semantic Fragmentation
+14 more scenarios
Data Theft
4- Browser Session Hijack
- Data Exfiltration
- Secret Harvest
- Prompt Leakage
Tool and Supply Chain
4- Dependency Confusion Fetch
- MCP Tool Poisoning
- Tool Abuse
- Tool Argument Injection
Agent Exploitation
6- Multi-Stage Chain
- Memory Seeding
- Agentic Delegation Abuse
- Workflow Tampering
- Markdown Link Hijack
+1 more scenarios
Identity and Impersonation
1- Delegated Authority Impersonation
RAG and Data Poisoning
2- RAG Poisoning
- Structured Output Injection
Content Safety
30- Drifting Persona Switch Jailbreak
- Hate Speech Roleplay Evasion
- Data Leak PII Extraction
- Self-Harm Encouragement Evasion
- Violent Crimes Weapon Instructions
+25 more scenarios
Financial Risk
14- AML Bypass Direct
- KYC Evasion Evasion
- Market Spoofing Evasion
- Money Laundering Direct
- Unauthorized Account Access Evasion
+9 more scenarios
How Kona Red works
Connect the target
Test an API endpoint, a manual chat session, or uploaded prompt-response pairs.
Choose the scan type
Run a full assessment, a content safety audit, a quick scan, or a custom scenario set.
Execute adversarial tests
Launch curated attacks against the selected model or workflow.
Review findings
Track pass rate, failed tests, vulnerabilities, trends, and estimated usage cost.
From scan setup to measurable results

Configure targets, run assessments, and track estimated token usage, judge calls, and cost from one testing surface.
Why teams use Kona Red
What teams get
- Structured scan results by category and severity
- Pass/fail visibility across scenarios and models
- Evidence for security reviews and internal governance
- Findings that map to real hardening actions
- Repeatable testing motion for new models and releases
Start testing AI systems with the same rigor you apply to the rest of security
Kona Red gives your team a practical way to evaluate LLMs, agents, and AI workflows before attackers, auditors, or customers do.
