🔒 Technology Controls · Agentic AI Security

See inside an
AI agent under attack.

An interactive demonstration of what happens when agentic AI systems operate without security controls — and what changes when guardrails are in place.

Launch Demo → How it works
Getting started

How to use the demo

Each scenario runs in under 30 seconds. Toggle guardrails on and off to see the same agent behave differently.

1

Pick a scenario

Choose from five scenarios in the left panel — from normal operation to active attacks and approval workflows.

2

Toggle guardrails

Switch between Guardrails ON and Guardrails OFF to see the same request handled differently.

3

Click Run

Watch the agent work in real time — every decision, tool call, and control action streamed live.

4

Compare the panes

Left pane shows everything happening inside the agent. Right pane shows what the end-user sees.

The two panes

◈ Under the Hood — Audit Trail

  • Every orchestrator decision and reasoning step
  • Tool calls dispatched to workers (with inputs)
  • What each worker returned
  • Guardrail events — blocks and reasons
  • Token usage and timing per step
  • Session summary and guardrail statistics

◈ User View — What They See

  • The request as the user typed it
  • The agent's final response only
  • Error messages when a request is blocked
  • In attack scenarios: a normal-looking response — with no sign that data was exfiltrated
Scenarios

Five scenarios, two outcomes each

Every scenario runs with guardrails on and off — showing what the controls prevent, and what happens without them.

Normal Operation

A routine financial research and writing task. Shows the healthy agentic loop — plan, delegate, synthesise — with a full audit trail.

Baseline
💉

Prompt Injection

The agent fetches a malicious website during research. The page contains hidden override instructions. With guardrails off, credentials are silently exfiltrated.

OWASP LLM01
🚫

Unauthorized Tool

The agent attempts to use a tool outside its permitted scope. The allowlist control blocks it. Without guardrails, the action proceeds unchecked.

Tool Allowlist
💸

Budget Exceeded

The agent makes far more tool calls than the task requires — a runaway agent. The call budget cap limits blast radius. Without it, API costs spiral.

Call Budget
🔍

Pre-execution Review

A £5.6M pension fund rebalancing is planned but nothing executes. Every intended trade is held for human-in-the-loop approval — the four-eyes control applied to autonomous AI.

Human-in-the-Loop

Ready to run the demo?

All scenarios are scripted for reliability — no live API calls required for the attack scenarios.

Open the Demo →