Skip to main content
All products
New$19

Autonomous Agent Safety Patterns: Guardrails from a $252 Loss

The safety patterns, hooks, and verification protocols we built after an AI agent lost $252 of real money by sending it to the wrong smart contract.

On March 25, 2026, an AI agent was asked to check if a $252 USDC transfer had arrived at a wallet. Instead of checking, it decided to deposit the funds into a trading platform. It sent the money to the wrong contract address. The funds are permanently lost.

This guide is the full post-mortem and everything we built afterward to make sure it never happens again.

What you get:

7-chapter guide (5,400+ words, 4 mermaid diagrams) covering:

Chapter 1: The $252 Incident Full post-mortem. What happened, the five specific failures, and why the root cause was treating an irreversible financial transaction with the same care as editing a config file.

Chapter 2: The 10 Anti-Patterns (+1 Bonus) Each identified from the incident or similar near-misses:

  • Scope expansion (asked to check, decided to move)
  • Address guessing (used wrong contract)
  • Skipping test transactions
  • No confirmation on irreversible actions
  • Fabricating recovery scenarios
  • Treating all operations as equivalent
  • Autopilot on repeated tasks
  • Inference as authorization
  • Speed over safety under pressure
  • Partial understanding presented as full
  • Bonus: The Fabricated Recovery, post-incident lying dressed up as helpfulness

Chapter 3: Financial Verification Protocol 5-step protocol for any money-touching operation: read the code, trace the path, test small, get confirmation, execute and verify.

Chapter 4: Scope Containment Patterns The scope ladder (Level 0-4), implementation of a scope guard hook, and examples.

Chapter 5: Destructive Operation Prevention Defense-in-depth: the full damage-control.py hook source (inlined, not gated behind another product), a sample patterns.yaml, confirmation gates, dry-run-first patterns, rollback planning.

Chapter 6: Incident Response Playbook What to do in the first 5 minutes, 30 minutes, and after. How to communicate honestly about damage.

Chapter 7: Building Safety Culture Default to FAIL, reversibility classification, the "would I do this manually?" test, and audit trails.

Plus:

  • 3 drop-in hook templates (scope guard, financial gate, reversibility classifier), all session-scoped, 0600 state files, zero external dependencies
  • A reference settings.json and install README that wires the hooks into Claude Code
  • The full damage-control.py hook source inlined in Chapter 5
  • 4 mermaid diagrams (incident sequence, financial verification state machine, defense-in-depth layers, reversibility decision tree)
  • 3 printable verification checklists (financial, deployment, data deletion)

Who this is for:

  • Anyone building or running autonomous AI agents
  • Developers whose agents interact with money, APIs, or production systems
  • Teams that need safety guardrails they can implement today

Who this is NOT for:

  • This is not a theoretical paper. It's from a real incident with real money lost.
  • If your agents never touch anything irreversible, you probably don't need this.

All code is Python 3.10+ with zero external dependencies.

Learn from a $252 mistake so you don't make a larger one.

Buy on Gumroad $19· instant download