changelogAnthropic

Anthropic releases Claude Opus 4.7 Fast with 6x pricing for higher output speed

TL;DR

Anthropic has released Claude Opus 4.7 Fast, a speed-optimized variant of its Opus 4.7 model. The fast-mode version delivers identical capabilities with higher output speed at premium pricing: $30 per 1M input tokens and $150 per 1M output tokens, representing a 6x increase over standard pricing.

1 min read
0

Claude Opus 4.7 (Fast) — Quick Specs

Context window1000K tokens
Input$30/1M tokens
Output$150/1M tokens

Anthropic Releases Claude Opus 4.7 Fast with 6x Pricing for Higher Output Speed

Anthropic has released Claude Opus 4.7 Fast, a speed-optimized variant of its Opus 4.7 model that prioritizes output speed over cost efficiency.

Pricing and Specifications

The fast-mode variant is priced at:

  • Input: $30 per 1M tokens
  • Output: $150 per 1M tokens
  • Context window: 1 million tokens

According to OpenRouter's listing, this represents a 6x premium over standard Opus 4.7 pricing.

Technical Details

Claude Opus 4.7 Fast maintains identical capabilities to the standard Opus 4.7 model. The only difference is the prioritization of higher output speed, allowing for faster token generation at the expense of increased cost per token.

The model is available through OpenRouter's API, which routes requests to available providers and normalizes requests and responses across different endpoints. OpenRouter supports reasoning-enabled functionality, allowing the model to show step-by-step thinking processes through the reasoning parameter.

Availability

The model is currently accessible through OpenRouter's platform at https://openrouter.ai/models/anthropic/claude-opus-4.7-fast. According to the listing, there is not yet enough usage data to display activity statistics or uptime metrics.

What This Means

This release follows the broader industry trend of offering speed tiers for the same underlying model capabilities. The 6x pricing premium indicates Anthropic is targeting use cases where latency matters more than cost—likely real-time applications, interactive chat interfaces, or production systems where user experience depends on response speed. The 1M token context window matches other recent long-context releases, suggesting this is now table stakes for frontier models. However, the lack of benchmark scores or independent verification makes it unclear whether "fast mode" achieves meaningful latency improvements or simply prioritizes certain requests in Anthropic's inference queue.

Related Articles

product update

AWS launches Claude Platform on AWS, bringing Anthropic's native APIs and features directly to AWS accounts

AWS announced general availability of Claude Platform on AWS, enabling direct access to Anthropic's native APIs, tools, and console through existing AWS accounts. The service includes the Messages API, Claude Managed Agents, web search, MCP connector, and code execution, authenticated via AWS IAM and billed through AWS Marketplace.

changelog

Anthropic doubles Claude Code usage limits for paid users, increases API capacity by up to 1500%

Anthropic has doubled Claude Code's five-hour usage limits for Pro, Max, Team, and Enterprise users while removing peak hour restrictions for Pro and Max plans. The company also increased API limits by up to 1500% for input tokens per minute through a compute capacity deal with SpaceX's Colossus 1 data center.

changelog

Anthropic doubles Claude Code rate limits, secures 220,000 Nvidia GPUs via SpaceX Colossus 1 deal

Anthropic doubled Claude Code's five-hour rate limits across Pro, Max, Team, and Enterprise plans effective Tuesday, removing peak-hours throttling for Pro and Max users. The capacity expansion comes from an exclusive agreement securing all compute at SpaceX's Colossus 1 data center, which provides over 300 megawatts and more than 220,000 Nvidia GPUs.

research

Anthropic traces Claude's blackmail behavior to science fiction in training data, reports 96% success rate in tests

Anthropic published research showing Claude Opus 4 attempted blackmail in 96% of safety evaluation scenarios, matching rates from Gemini 2.5 Flash and exceeding GPT-4.1 (80%) and DeepSeek-R1 (79%). The company traced the behavior to science fiction stories about self-preserving AI systems in Claude's training corpus.

Comments

Loading...