Why I Built a SQLite Brain for AI Coding (and How It Saves 70-90% Tokens)
The Problem Nobody Talks About AI coding tools are incredible for the first 30 minutes. Then quality drops. By the time you're on your 5th file edit, Claude is: Forgetting your project conventions Breaking imports it created 10 minutes ago Re-asking questions you already answered Producing increasingly generic, copy-paste code This is context rot — as the context window fills with file reads, error I built ShipFast — a framework that npm i -g @shipfast-ai/shipfast Then in your AI tool: /sf-do add dark mode toggle Behind the scenes: Analyze — intent detection, complexity scoring (zero tokens) Optimize — selects which agents to skip based on brain.db learnings Plan — Scout researches, Architect creates task list (fresh context) Execute — Builder implements each task in a separate fresh context Verify — Critic reviews, consumer check, stub scan, build verify Learn — Records decisions + patterns for next time Session Without ShipFast With ShipFast 1st time ~100K tokens ~30K (70% saved) 2nd time ~100K tokens ~15K (85% saved) 3rd time ~100K tokens ~5K (95% saved) The brain gets smarter every session...
