Learn how AI browser assistants get tricked and practice finding safe "flags" hidden in webpages.
Some pages hide instructions your AI helper could follow by mistake. Learn how these tricks work and practice spotting them with safe, hands on mini challenges.
✅ 100% simulated • No external calls • We only count: games played, correct/failed flag submits
-
•
Agentic browser: a browser with an AI helper that can read pages, click, fill forms, and automate simple tasks for you.
-
•
Prompt injection: hidden or sneaky text in content that tries to change the AI's instructions so it does something unintended.
-
•
Guardrails: safety checks and filters around the AI that reduce risky behavior (e.g., focusing on sanitized, visible page content and limiting what tools it can use). Helpful—but not magic.
Why flags, not "trick the browser"
Hands-on learning, zero risk. We use pretend flags so you learn safely; real "trick the agent" stunts can be harmful—and many modern guardrails block them—while DIY or hobby agents might still miss them.
⚠️ The gotcha: when a page "talks back" to your agent
This is the most common way AI helpers get steered: they read everything—even parts you'd never see.
AI helpers are great at following instructions—and that's the problem. Some webpages hide extra instructions in places you don't usually read (like HTML comments, metadata, or invisible elements). When your agent "reads the page," it might treat those hidden bits as commands. That's called prompt injection.
Common hiding spots include:
-
•
HTML comments
(not visible on the page) -
•
hidden DOM
(elements styled to be invisible) -
•
meta/Open Graph tags
(meant for machines, not humans)
See exactly how a hidden instruction can hijack an agent
What this site gives you (in safe, bite-size steps)
This app is a hands-on training ground where you learn how agents can be tricked—and how to spot it—by playing through short, realistic challenges. It's built to be safe and self-contained; nothing here touches your real accounts.
Open a level and read the goal
Use your agentic browser to try the task
Enter the hidden flag if your agent found it
See "What happened?" plus simple defenses
📚 What you'll learn, in plain terms
Where hidden instructions often live on websites
Why agents sometimes confuse data with instructions
Everyday habits that reduce risk—before your agent clicks
🛡️ Why "guardrails" aren't magic (and what to do instead)
Modern AI tools try to filter out bad instructions, monitor unusual actions, and isolate risky content—but nothing is foolproof. Because language models treat most text as potential instructions, clever attackers keep finding ways to slip messages past filters. So you still need a little human savvy.
🚨 Street-smart checklist for everyday use
Hidden instructions often live in comments, hidden elements, or meta tags.
When you just want a summary; avoid letting the agent auto-traverse hidden or cross-origin content.
If your agent wants to post data, fill a form, or call an API, require a confirmation step.
Tell the agent which page(s) it may read—avoid "browse the whole site" unless you trust it.
Don't keep banking or email tabs open while experimenting with new agent features.
🚀 Ready to try?
Start at Level 1. You'll see exactly how a hidden instruction can nudge an AI helper—and how a small change in how you use your agent can block it. Each level gives a clear explainer and "defenses to know," so you leave with practical instincts, not just theory.
Mission Status
Started strong — continue to Level 1.
Training Levels
New to these terms? No stress; every level starts with Plain Words that explain them in simple language.
HTML Comments
IntroHidden HTML comments contain instructions that agents eagerly read and follow.
Hidden DOM
NoviceInvisible elements and template tags hide malicious payloads.
Meta & OG Tags
NovicePage metadata becomes instruction channels. OG tags are perfect hiding spots.
Accessibility Trap
IntermediateAlt text and ARIA labels weaponized for injection attacks.
Structured Data
IntermediateJSON-LD scripts provide machine-readable metadata with hidden payloads.
Off-Path Files
IntermediateInfrastructure files become covert instruction channels for curious agents.
Cross-Origin
AdvancedHostile cross-origin embedded content injects malicious instructions.
PDF Injection
AdvancedEmbedded documents become trojan horses for agent manipulation.
Tool Hijacking
ExpertHidden instructions manipulate agent tool usage and API calls.
Multi-Hop Exfil
ChampionComplex chained injection with multi-origin data exfiltration.