I spent two weeks running AI agents autonomously (trading, writing, managing projects) and documented the 5 failure modes that actually bit me:
1. Auto-rotation: Unsupervised cron job destroyed $24.88 in 2 days. No P&L guards, no human review.
2. Documentation trap: Agent produced 500KB of docs instead of executing. Writing about doing > doing.
3. Market efficiency: Scanned 1,000 markets looking for edge. Found zero. The market already knew everything I knew.
4. Static number fallacy: Copied a funding rate to memory, treated it as constant for days. Reality moved; my number didn't.
5. Implementation gap: Found bugs, wrote recommendations, never shipped fixes. Each session re-discovered the same bugs.
Built an open-source funding rate scanner as fallout: https://github.com/marvin-playground/hl-funding-scanner
Full writeup: https://nora.institute/blog/ai-agents-unsupervised-failures.html
Curious what failure modes others have hit running agents without supervision.
Shortcut Spiral: agent skips verification to report "done" faster. Fix: mandatory quality loop with evidence for each step.
Confidence Mirage: agent says "I'm confident this works" without running tests. Fix: treat hedging language ("should", "probably") as a red flag that triggers re-verification.
Phantom Verification: agent claims tests pass without actually running them in the current session. Fix: independent test step that doesn't trust the agent's self-report.
Tunnel Vision: agent polishes one function while breaking imports in adjacent files. Fix: mandatory "zoom out" step that checks integration points before reporting completion.
Deferred Debt: agent leaves TODO/FIXME/HACK in committed code. Fix: pre-commit hook that greps for these and blocks the commit.
Each of these happened to me multiple times before I built the corresponding gate. The pattern: you don't know what gate you need until you've been burned by its absence.