Cursor 3’s parallel agents let you run up to 8 AI agents simultaneously across isolated git worktrees. Four agents working in parallel — UI, API, database, and tests — can cut wall-clock development time from 8 hours to 2 hours. This tutorial covers all three methods: the Agents Window, /multitask command, and manual worktree setup.

What’s New in Cursor 3: The Agent-First Revolution (April 2026)

Cursor 3 launched on April 2, 2026, with a complete architectural rethink: the classic IDE layout was replaced with an agent-first interface built around parallel AI fleets. The update introduced three major new capabilities — the Agents Window sidebar for managing multiple concurrent agents, the /multitask command for automatic task decomposition, and the in-house Composer 2 model optimized for multi-agent coordination. Unlike Cursor 2.0 where you could technically run parallel agents through manual git worktree commands, Cursor 3 gives every parallelism feature a first-class UI, making it accessible without CLI knowledge. The rebuilt interface treats agents as the primary unit of work: you spawn agents for specific tasks, monitor them in a sidebar, and merge results back via an Apply button. The launch sparked significant community discussion — some developers questioned whether Cursor 3 introduced genuinely new capabilities or rebranded features that power users had already been doing manually. The honest answer: the underlying git worktree technology existed before, but the Cursor 3 interface reduces setup friction from 10+ manual steps to a single click.

What Changed from Cursor 2 to Cursor 3?

Cursor 2 required manually running git worktree add in a terminal, scoping each agent to a directory, and managing merges by hand. Cursor 3 automates all of this: clicking “Worktree” in the agent dropdown creates the isolated branch, runs the agent inside it, and presents a diff for review. The Agents Window replaced the traditional file tree as the primary navigation panel — a deliberate signal that Cursor’s team expects most work to happen through agents, not direct file editing.

How Cursor Parallel Agents Work: Git Worktrees and Isolated Execution

Cursor parallel agents use git worktrees as their foundational technology — each agent gets its own isolated directory, filesystem state, and branch, so agents can’t overwrite each other’s changes. A git worktree is a linked working tree that shares the same .git object store as your main repository but checks out a different branch into a separate folder. When you spawn a Cursor agent with the Worktree option, Cursor automatically runs git worktree add -b agent/task-name ../task-name-worktree in the background, opens that directory in an isolated environment, and launches the agent there. The agent’s file reads and writes are scoped entirely to its worktree — it cannot see uncommitted changes in your main working directory. This isolation prevents the merge conflicts that plagued early multi-agent setups where agents would overwrite shared files. Once an agent finishes, Cursor shows a diff against your main branch; you click Apply to cherry-pick the changes. The Cursor 3 implementation supports up to 8 agents running simultaneously, and the Agents Window sidebar shows real-time status for each: running, paused, waiting for review, or completed.

Why Git Worktrees Prevent Conflicts

Traditional parallel execution on a single branch fails because agents race to write the same files. When two agents both try to update api/routes.ts, one agent’s changes get overwritten. Git worktrees solve this structurally: each agent writes to its own branch, and you control when changes merge. The only conflict zone is the final merge step — which is exactly when you want human review anyway.

Prerequisites: Setting Up Your Environment for Parallel Agent Workflows

Before running parallel agents in Cursor 3, you need four things in place: a git-initialized repository, at minimum a Cursor Pro subscription ($20/month), sufficient RAM for parallel agent context, and a codebase structured for modular work. Parallel agents require a git repo because worktrees are a git primitive — Cursor can’t create isolated agent environments without version control. The Pro plan unlocks the Agents Window and worktree support; the Free tier limits you to single-agent sessions. On the hardware side, each local agent consumes roughly 1–2 GB of RAM for its language server and context window, so running 4 simultaneous local agents requires at least 8 GB free RAM (16 GB recommended). The most important prerequisite is invisible: your codebase should have clear module boundaries. Parallel agents perform best when each agent can own a distinct directory — components/, api/, db/, tests/ — without needing to read each other’s in-progress work. If your architecture has deep coupling between layers, agents will make conflicting assumptions about interfaces that haven’t been defined yet.

Cursor Plan Requirements for Parallel Agents

FeatureFreePro ($20/mo)Pro+ ($60/mo)Ultra ($200/mo)
Agents WindowNoYesYesYes
Local parallel agentsNoUp to 4Up to 8Up to 8
Cloud agentsNoNoYesYes
/multitask commandNoYesYesYes
Background agentsNoLimitedFullFull

Method 1 — The Agents Window: Cursor 3’s Native Parallel UI

The Agents Window is Cursor 3’s primary interface for launching and managing parallel agents — it replaces the file explorer as the default left sidebar panel and shows all active, paused, and completed agents in a single view. To start your first parallel workflow: open Cursor 3, press Cmd+Shift+A (Mac) or Ctrl+Shift+A (Windows/Linux) to open the Agents Window, click the + button to create a new agent, write your task prompt, then click the dropdown arrow next to “Run” and select “Worktree.” Cursor automatically creates an isolated git branch and launches the agent there. You can spawn a second agent immediately — click + again, write a different prompt targeting a different part of the codebase, and run it. Both agents appear in the sidebar with status indicators. The sidebar shows a real-time token counter and a progress description generated by the agent itself. When an agent completes, the status changes to “Review” and a diff view opens. You can pause any agent mid-run if you need to redirect it, or stop it entirely if the approach looks wrong. The Agents Window also supports a multi-model comparison mode where you send the same prompt to agents running different underlying models — useful for evaluating which model handles your specific codebase better.

Agents Window Keyboard Shortcuts

ActionMacWindows/Linux
Open Agents WindowCmd+Shift+ACtrl+Shift+A
New agentCmd+NCtrl+N
Pause active agentCmd+PCtrl+P
Review agent diffCmd+DCtrl+D
Apply agent changesCmd+EnterCtrl+Enter

Method 2 — The /multitask Command: Automatic Task Decomposition

The /multitask command is Cursor 3’s highest-level parallel agent primitive — you describe a large feature and Cursor automatically decomposes it into parallel subtasks, spawns separate agents for each, and coordinates their execution. To use it: open a new agent session, type /multitask, then describe your feature. For example: /multitask Build a user authentication system with JWT tokens, including the API endpoints, database schema, React login form, and test suite. Cursor analyzes your codebase, identifies logical boundaries, and proposes a decomposition — typically something like: Agent 1 (database schema + migrations), Agent 2 (API endpoints), Agent 3 (React components), Agent 4 (test suite). You can edit the proposed decomposition before confirming. Once confirmed, each agent runs in its own worktree. The /multitask decomposition is most effective on features with 4+ clearly separable components. It performs poorly on tightly coupled refactors where each subtask depends on another’s output — in those cases, sequential agents are more reliable. A practical limit: most developers find 2–4 parallel agents manageable; beyond that, the review overhead for each agent’s diff outweighs the time saved.

When to Use /multitask vs Manual Agent Spawning

Use /multitask when you have a large, well-defined feature with independent components. Use manual spawning from the Agents Window when you want fine-grained control over each agent’s prompt, model selection, or starting context. Manual spawning is also better for iterative work where each agent’s result might change the direction of the next.

Method 3 — Manual Worktree Setup for Advanced Parallel Control

Manual worktree setup gives developers the most control over parallel agent execution — it’s the approach that power users preferred before Cursor 3, and it remains useful for complex branching strategies the UI doesn’t support. Start with a clean main branch, then run the following commands to set up three parallel worktrees:

# Create worktrees for parallel agent work
git worktree add -b feature/ui ../myapp-ui
git worktree add -b feature/api ../myapp-api
git worktree add -b feature/tests ../myapp-tests

# Verify worktrees
git worktree list

Now open three separate Cursor windows, each pointed at a different worktree directory. In each window, start an agent session scoped to that directory’s responsibility. The agent in myapp-ui/ should only receive UI-related context; the agent in myapp-api/ should only work on the API layer. This scoping prevents agents from making assumptions about code that doesn’t exist yet in their branch. When agents complete, merge in order of dependency — database schema first, then API, then UI:

# In main branch
git merge feature/database
git merge feature/api
git merge feature/ui
git merge feature/tests

# Remove worktrees
git worktree remove ../myapp-ui
git worktree remove ../myapp-api
git worktree remove ../myapp-tests

Manual worktrees are also the right choice when you’re working on an existing PR branch and want to spawn experimental parallel work without creating a tangled agent history in the Agents Window.

Cloud Agents vs Local Agents: A Practical Decision Framework

Cursor 3 offers two distinct execution environments for parallel agents — local (runs on your machine) and cloud (runs on Cursor’s remote VMs) — and the choice significantly affects both performance and cost. Local agents are fast to start, don’t incur per-minute charges, and keep your code on your own hardware. Cloud agents run in isolated VMs with dedicated compute, handle long-running tasks without tying up your laptop, and support Background Agent mode where they continue working even when Cursor is closed. The critical gotcha: cloud agents bill separately from your subscription using per-minute VM charges, and several early Cursor 3 users reported spending $2,000+ in just two days. The safest default for most developers is local agents for tasks under 30 minutes, cloud agents only for long migrations or overnight runs with hard budget caps set in Cursor’s settings. Cloud agents also require MAX mode, which adds a 20% surcharge on every run on top of the VM cost.

Cloud vs Local Agent Decision Matrix

ScenarioRecommended ModeReason
Feature development (< 2 hours)LocalNo per-minute charges
Large codebase migrationCloudLong-running, needs dedicated VM
Overnight background tasksCloudRuns while laptop is off
Sensitive proprietary codeLocalCode never leaves your machine
Tight monthly budgetLocalPredictable subscription cost
CI/CD integrationCloudBetter API support
Learning/experimentationLocalNo surprise bills

Real-World Tutorial: Full-Stack App with 4 Parallel Agents (UI + API + DB + Tests)

Here is a concrete walkthrough of building a task management API with four parallel Cursor agents — the most common real-world parallel agent pattern. This tutorial builds a full-stack task manager with a React frontend, Express API, PostgreSQL schema, and Jest test suite. Using this approach, an e-commerce platform was built in 18 minutes with three parallel agents versus 45+ minutes using traditional sequential workflows.

Step 1: Define the shared interface contract. Before spawning agents, create a SHARED_SPEC.md in your repo with the data types and API contract that all agents must respect. This prevents agents from making incompatible assumptions:

# Task Schema
interface Task { id: string; title: string; status: 'todo'|'done'; createdAt: Date; }

# API Contract
POST /tasks → 201 { task: Task }
GET /tasks → 200 { tasks: Task[] }
PATCH /tasks/:id → 200 { task: Task }

Step 2: Spawn four agents. Open the Agents Window and create four agents with these prompts:

  • Agent 1 (DB): “Create PostgreSQL schema and migrations for the Task interface in SHARED_SPEC.md. Use Drizzle ORM.”
  • Agent 2 (API): “Build Express routes for the API contract in SHARED_SPEC.md. Assume the DB schema exists as defined.”
  • Agent 3 (UI): “Build React components for a task list. Fetch from the API in SHARED_SPEC.md. Use Tailwind.”
  • Agent 4 (Tests): “Write Jest tests for the API contract in SHARED_SPEC.md. Use supertest for HTTP assertions.”

Step 3: Monitor and coordinate. Watch the Agents Window. The DB agent typically finishes first (schema is straightforward). The API and test agents finish around the same time. The UI agent may take longer. Pause any agent that asks for clarification and answer in its chat window.

Step 4: Review and merge. Apply diffs in dependency order: DB → API → UI → Tests. Run npm test after each merge to catch interface mismatches early.

An 8,000-line Next.js migration from pages/ to app/ router completed in roughly 9 minutes using this parallel agent pattern — a task that would typically take a senior developer a full workday.

Cost Management: Staying on Budget While Running Parallel Agents

Cursor’s parallel agent costs can spiral quickly without guardrails — heavy cloud agent users report monthly overages of 15–30% above their subscription cost, and documented cases exist of entire annual subscriptions depleted in a single day. The primary cost levers are: local vs cloud execution, MAX mode usage, and agent run duration. Here are concrete tactics to manage costs. First, always start with local agents. Local parallel agents run entirely on your hardware and consume only your subscription’s request credits — no per-minute VM charges. Second, set a cloud spending cap before enabling cloud agents: Cursor → Settings → Cloud Agents → Monthly Budget Cap. Third, disable MAX mode unless you’re running tasks that genuinely require the most capable model. MAX mode adds a 20% surcharge on every cloud agent run. Fourth, use the /multitask decomposition preview to estimate agent count before confirming — more agents means more compute. Fifth, for long-running tasks, use Background Agents scheduled for off-peak hours when cloud VM pricing may be lower. Sixth, monitor the token counter in the Agents Window: agents consuming 100k+ tokens on a simple task are probably looping on a problem and should be paused and redirected.

Cursor 3 Cost Comparison: Local vs Cloud

Task TypeLocal Agent CostCloud Agent CostRecommendation
1-hour feature~$0 extra~$3–8Local
8-hour migration~$0 extra~$20–50Local with breaks
Overnight background taskNot possible~$30–80Cloud with cap
4 parallel agents, 2 hours~$0 extra~$15–30Local

Best Practices, Common Pitfalls, and When NOT to Use Parallel Agents

Parallel agents deliver the highest ROI on tasks with clear separation of concerns — but they actively harm productivity when applied to the wrong problems. Most developers find 2–4 parallel agents manageable; beyond that, the overhead of reviewing 8 agent diffs and resolving merge conflicts outweighs the time saved. The sweet spot is 3–4 agents on a well-specced feature. The most common pitfall is spawning agents without a shared interface contract. If Agent A (API) makes different assumptions about the data model than Agent B (UI), you’ll spend more time fixing the mismatch than you would have spent doing the work sequentially. Always write SHARED_SPEC.md first. The second pitfall is over-trusting agent output. Parallel agents produce more code faster — but also more bugs faster. Run tests after each merge, not just at the end. The third pitfall is using cloud agents for short tasks where startup latency dominates: a cloud VM takes 30–60 seconds to initialize, which is significant for a 5-minute task.

When NOT to use parallel agents:

  • Tightly coupled refactors where each change depends on the previous
  • Tasks requiring a single agent to maintain context across the full codebase
  • Debugging sessions where you need to explore the problem space
  • Security-sensitive changes where multiple agents could introduce conflicting assumptions
  • Teams without a clear code review process for agent-generated diffs

When parallel agents excel:

  • Greenfield development with clear module boundaries
  • Large migrations with independent subsystems
  • Adding tests to an existing codebase
  • Building multi-component features from a clear spec
  • Comparing model performance on the same task

FAQ

The questions below address the most common points of confusion for developers adopting Cursor 3 parallel agents in 2026. Cursor parallel agents work by running multiple AI instances simultaneously across isolated git worktrees — each agent gets its own branch, filesystem scope, and context window. The key decisions every developer must navigate are: how many agents to run (2–4 is the practical sweet spot), whether to use local or cloud execution (local for most tasks under 2 hours), and how to structure the shared interface contract that prevents agents from making conflicting architectural assumptions. Pricing is the other critical factor: local parallel agents cost nothing extra on the Pro plan ($20/month), while cloud agents bill per VM minute and can generate unexpected overages without spending caps enabled. The FAQs below cover all of these dimensions in detail.

How many parallel agents can I run in Cursor 3?

Cursor 3 supports up to 8 AI agents running simultaneously across isolated git worktrees. However, most developers find 2–4 parallel agents to be the practical limit — beyond 4 agents, the overhead of reviewing each agent’s diff and resolving merge conflicts at the end typically outweighs the parallelism benefit. Start with 2–3 agents and scale up as you build familiarity with the workflow.

Do I need a paid Cursor subscription to use parallel agents?

Yes. Parallel agents and the Agents Window require at minimum the Cursor Pro plan ($20/month). The Free tier is limited to single-agent sessions. Cloud agents and Background Agents require Pro+ ($60/month) or Ultra ($200/month). Local parallel agents (up to 4 simultaneous) are available on Pro; up to 8 simultaneous agents require Pro+ or Ultra.

What is the difference between the /multitask command and manually spawning agents?

/multitask automatically decomposes a feature description into parallel subtasks and spawns agents for each. Manual spawning via the Agents Window gives you full control over each agent’s prompt, model selection, and starting context. Use /multitask for well-defined features with 4+ clearly separable components. Use manual spawning for iterative or experimental work where you want fine-grained control over what each agent knows.

How do I prevent parallel agents from creating conflicting changes?

The key is defining a shared interface contract before spawning agents. Create a SHARED_SPEC.md file that specifies the data types, API contracts, and file structure that all agents must respect. Each agent’s prompt should reference this spec explicitly. Git worktrees prevent file-level conflicts at write time, but agents can still make incompatible architectural assumptions — the shared spec prevents that class of error.

Are cloud agents worth the extra cost compared to local agents?

For most development tasks, local parallel agents on Pro ($20/month) deliver the best cost-to-value ratio. Cloud agents are worth the extra cost for: long-running background tasks (overnight migrations), tasks where you want computation to continue while your laptop is off, and large codebases that exceed local RAM. The critical warning: set a monthly cloud spending cap before enabling cloud agents. Early adopters reported spending $2,000+ in two days without caps in place.