Scaling long-running autonomous coding
- •Cursor experiments with hundreds of concurrent autonomous coding agents to build a web browser from scratch.
- •System generates over one million lines of code in a week using a planner-worker-judge architecture.
- •Resulting Rust-based 'FastRender' browser successfully renders live websites, demonstrating massive-scale autonomous programming capabilities.
Wilson Lin at Cursor recently demonstrated the burgeoning power of "agent swarms" by tasking a massive fleet of autonomous AI agents with an incredibly complex software engineering feat: building a functional web browser from scratch. Over the course of a single week, this coordinated system produced over one million lines of code for a project dubbed FastRender, highlighting a shift toward high-volume, long-running AI operations. The experiment utilized a hierarchical workflow where specialized planner agents broke down the massive goal into manageable tasks for worker agents to execute, while judge agents performed final quality checks. This structure mimics the delegation found in large human engineering teams but operates at a vastly accelerated pace, utilizing an underlying LLM to process trillions of tokens in the process. Interestingly, the system even integrated official web specification documents as reference materials to ensure the generated code adhered to industry standards. While the resulting Rust-based browser exhibits some rendering glitches, it successfully loads complex sites like Google and Simon Willison’s personal blog. This milestone suggests that the "cheat code" for AI-driven development lies in providing agents with robust conformance suites—pre-existing tests that act as a ground truth for the AI. As these autonomous fleets scale, the boundary between simple code assistance and full-scale automated software production is rapidly dissolving.