Cursor shipped version 3 on Wednesday, and the release notes read less like an IDE changelog and more like a manifesto. The opening line from the team: "Built for a world where all code is written by agents." That's not marketing fluff — the entire interface was rebuilt around a single bet: you're going to manage AI workers, not type semicolons.

The Agents Window Changes Everything

The headline feature is the Agents Window — a separate surface (not a chat sidebar, not an inline completion popup) where you spin up multiple AI agents working in parallel across different repos. Each agent gets its own tab, its own execution context, and its own branch.

Here's what makes it different from just "chat but bigger": these agents run as background processes. You describe a task — "refactor the auth middleware to use JWT rotation" — and the agent goes off, creates a git worktree, makes changes, runs your tests, takes screenshots of the result, and comes back with a PR-ready diff. While that's happening, you're working on something else entirely. Or launching two more agents.

The cloud/local handoff is surprisingly smooth. Start something locally, push it to a cloud VM when you realize it's going to take twenty minutes, pull the result back when it's done. Cloud agents generate demo videos automatically — no screen recording needed for async review.

There's also the new /best-of-n command that quietly might be the most useful addition: send the same prompt to multiple models (say, Claude Sonnet, GPT-5.4, and Cursor's own Composer 2) and get side-by-side outputs. For architectural decisions where you want to catch blind spots, this is genuinely clever.

Design Mode (Cmd+Shift+D) lets you click on UI elements in your browser and describe changes visually rather than in text. The agent implements the modifications directly. Think Figma-to-code but driven by annotation rather than pixel-perfect mockups.

And then there are Automations — always-on agents triggered by events from Slack, Linear, GitHub, PagerDuty, or plain webhooks. Set up a trigger for "new bug label on GitHub issue," attach an agent instruction like "reproduce, diagnose, and open a fix PR," and walk away. The tool handles scheduling and execution in the cloud. This is where it stops being an editor and starts being infrastructure.

The Philosophy That's Making Developers Uncomfortable

"All code is written by agents" is a bold claim, and the community response has been divided.

One forum user put it bluntly: "I still want to code, not vibe my way through tickets." A Cursor engineer jumped in within minutes to clarify — the classic IDE still exists, the Agents Window is additive, not a replacement, and you can ignore it completely.

But the interface redesign tells a different story. The team clearly believes the future developer workflow looks more like project management than programming. You define specs, review diffs, approve PRs, and occasionally jump in to fix something the agents couldn't handle. Whether that excites or terrifies you probably depends on how much of your identity is wrapped up in writing code by hand.

I think the reality lands somewhere in the middle. For boilerplate, CRUD endpoints, test scaffolding, and migration scripts? Let the agents rip. For core business logic, performance-critical paths, and security boundaries? You probably want human eyes on the actual implementation, not just the review. Cursor 3 gives you both modes. The question is which one you'll default to after a month.

How It Stacks Up Against the Competition

The AI coding tool landscape has gotten genuinely crowded. Here's the honest breakdown as of this week:

Cursor 3 Claude Code Codex
Interface GUI (VS Code fork) Terminal-native Cloud sandbox + CLI
Multi-agent Yes, Agents Window Single session Background tasks
Model flexibility GPT-5.4, Claude, Gemini, Composer 2 Claude models only GPT family only
Token efficiency Average 2-4x better Varies by task
Automations Slack, GitHub, Linear triggers None built-in GitHub integration
Self-hosted option Yes, cloud agents on your infra N/A No
Individual price $20/mo $20/mo Bundled with ChatGPT Pro
Team price $40/user/mo $150/user/mo $200/user/mo

The multi-model flexibility is a genuine advantage. Being locked into a single provider's models feels like a liability when the leaderboards shift monthly. Running /best-of-n across three providers and picking the winner per task is something neither Claude Code nor OpenAI's agent can match right now.

But Claude Code still dominates pure terminal workflows. If your daily loop is SSH into a server, navigate a monorepo, and run a complex refactor across 40 files — the terminal-native experience is hard to beat inside a GUI wrapper. The Cursor team seems to know this, positioning the Agents Window as complementary to the editor rather than replacing it.

For teams, the pricing gap is stark. Cursor at 40/user/month versus 150 for Claude Code or $200 for Codex Enterprise is a significant delta when you're scaling to 50 engineers. The credit pool system charges based on token consumption for premium models, but their "Auto" routing — which picks the most cost-effective model for each request — remains unlimited on Pro.

What to Try First

If you're already a Pro subscriber, the upgrade is free and automatic. Three things worth testing immediately:

Fire up the Agents Window with two parallel tasks on separate branches. The git worktree isolation means they genuinely can't step on each other — something anyone who's dealt with merge conflicts from concurrent AI edits will appreciate.

Try /best-of-n on your next architecture decision. Seeing Claude, GPT-5.4, and Composer 2 reason about the same design problem side by side changes how you evaluate AI suggestions. You stop trusting any single model's confidence and start looking for consensus.

If you touch frontend at all, give Design Mode thirty minutes. Visual annotation beats describing CSS changes in natural language every time, and the agent handles implementation surprisingly well for simple component modifications.

Cursor bet the whole product on agents eating code. Whether they're right probably won't be clear for another year — but the tooling to test that thesis just landed for free on every Pro account.