Cursor releases Composer 2 at $0.50/$2.50 per 1M tokens, undercutting Claude and GPT-4 on pricing
Cursor released Composer 2, a code-specialized model priced at $0.50 per million input tokens and $2.50 per million output tokens—roughly 90% cheaper than Claude Opus 4.6 ($5.00/$25.00) and 60% cheaper than GPT-5.4 ($2.50/$15.00). The model scores 61.3 on Cursor's internal CursorBench, competitive with Claude Opus 4.6 (58.2) but below GPT-5.4 Thinking (63.9).
Cursor Composer 2 — Quick Specs
Cursor Launches Code-Only Model to Break Pricing Dependency
Cursor released Composer 2, its second-generation code model, priced at $0.50 per million input tokens and $2.50 per million output tokens for the standard version. A faster variant costs $1.50/$7.50. Both versions undercut rival API pricing by substantial margins.
Pricing Comparison
| Model | Input | Output |
|---|---|---|
| Composer 2 | $0.50 | $2.50 |
| Composer 2 Fast | $1.50 | $7.50 |
| Claude Opus 4.6 | $5.00 | $25.00 |
| GPT-5.4 (short context) | $2.50 | $15.00 |
| GPT-5.4 (long context) | $5.00 | $22.50 |
Performance Metrics
Composer 2 scores 61.3 on CursorBench, Cursor's internal coding benchmark—a 38% jump from Composer 1.5 (44.2) and competitive with Claude Opus 4.6 (58.2), though below GPT-5.4 Thinking (63.9).
Additional benchmarks show continued improvement across multiple evaluation frameworks:
| Model | CursorBench | Terminal Bench 2.0 | SWE-bench Multilingual |
|---|---|---|---|
| Composer 2 | 61.3 | 61.7 | 73.7 |
| Composer 1.5 | 44.2 | 47.9 | 65.9 |
| Claude Opus 4.6 | 58.2 | 58.0 | 77.8 |
| GPT-5.4 Thinking | 63.9 | 75.1 | N/A |
Cursor co-founder Aman Sanger told Bloomberg the model was trained exclusively on code data, enabling a smaller, cost-effective architecture. "It won't help you do your taxes. It won't be able to write poems," he said.
Training Approach
Quality gains came from stronger continued pretraining followed by reinforcement learning on long-horizon coding tasks—multi-step programming challenges requiring hundreds of individual actions. This approach drove the significant benchmark improvements over Composer 1.5 and Composer 1 (38.0 on CursorBench).
Strategic Necessity for Cursor
Building its own model addresses a structural dilemma: Cursor competes directly with Anthropic and OpenAI while depending on their APIs. As long as Cursor purchases third-party models, it faces pricing constraints its competitors don't—Anthropic and OpenAI can heavily subsidize their own products.
Cursor reportedly estimates a single Claude Code subscription at $200/month generates approximately $5,000 in compute costs for Anthropic. Consumer subscriptions at Cursor currently run at negative margins, with enterprise contracts providing profitability.
With over 1 million daily users and 50,000 enterprise customers, Cursor is discussing funding at a ~$50 billion valuation. As AI coding agents improve, the risk persists that users could bypass the IDE entirely and work directly with model providers—making Composer 2 essential to Cursor's long-term independence.
What This Means
Composer 2 represents a deliberate shift toward self-sufficiency. Cursor's pricing advantage is real but performance remains competitive rather than dominant. The code-only approach is pragmatic: narrower focus enables cheaper training and faster inference. Cursor's bet hinges on whether pricing and adequate performance can retain users against providers with deeper resources and broader models. The benchmark gap with GPT-5.4 Thinking suggests room for improvement, but SWE-bench performance (73.7) demonstrates practical engineering capability.
Related Articles
OpenAI releases GPT-4o mini with 128K context at $0.15/$0.60 per 1M tokens
OpenAI released GPT-4o mini on July 18, 2024, a compact multimodal model with 128,000 token context window priced at $0.15 per million input tokens and $0.60 per million output tokens. The model achieves 82% on MMLU and claims to rank higher than GPT-4 on chat preference leaderboards while costing 60% less than GPT-3.5 Turbo.
OpenAI releases GPT-5.4 mini and nano with 3-4x price increases but major performance gains
OpenAI has released GPT-5.4 mini and GPT-5.4 nano, compact models optimized for coding and subagent tasks. The new models deliver significant performance improvements—GPT-5.4 mini reaches 54.4% on SWE-Bench Pro versus 45.7% for GPT-5 mini—but cost 3-4x more per input token than their predecessors.
OpenAI's GPT-5.4 mini now available in GitHub Copilot
OpenAI has released GPT-5.4 mini, the lightweight variant of its agentic coding model GPT-5.4, in GitHub Copilot. The model represents OpenAI's highest-performing mini offering to date for code generation and completion tasks.
Mistral AI releases Mistral Small 4, claims improved performance on reasoning tasks
Mistral AI has released Mistral Small 4, the latest iteration of its small-scale language model. The company claims improvements in reasoning and coding capabilities, though specific benchmark scores and pricing details have not been publicly disclosed.