model releasexAI

xAI releases Grok 4.20 with 2M context window and native reasoning capabilities

TL;DR

xAI released Grok 4.20 on March 31, 2026, its flagship model featuring a 2 million token context window, $2 per million input tokens and $6 per million output tokens pricing, and toggleable reasoning capabilities. The model includes web search functionality at $5 per 1,000 queries and claims industry-leading speed with low hallucination rates.

2 min read
0

Grok 4.20 — Quick Specs

Context window2000K tokens
Input$2/1M tokens
Output$6/1M tokens

xAI Releases Grok 4.20 With 2M Context and Reasoning Mode

xAI released Grok 4.20 on March 31, 2026. The model features a 2 million token context window, priced at $2 per million input tokens and $6 per million output tokens. Web search functionality costs $5 per 1,000 queries. Knowledge cutoff is September 1, 2025.

Key Specifications

Context and Pricing:

  • Context window: 2,000,000 tokens
  • Input pricing: $2/M tokens
  • Output pricing: $6/M tokens
  • Web search: $5 per 1,000 queries

Capabilities: Grok 4.20 includes native agentic tool-calling abilities and an optional reasoning mode that can be enabled or disabled via API parameter. When enabled, the model exposes its step-by-step reasoning process through a reasoning_details array in the response. Users can preserve reasoning context across multi-turn conversations by passing the complete reasoning details back in subsequent requests.

Performance Claims

xAI claims Grok 4.20 delivers "industry-leading speed" combined with low hallucination rates and "strict prompt adherence." The company positions the model as producing "consistently precise and truthful responses," though specific benchmark scores against competing models have not been disclosed.

Technical Details

The model is accessible through OpenRouter, which normalizes API requests and responses across multiple providers and includes fallback routing to maximize uptime. OpenRouter's documentation indicates support for reasoning-enabled models with access to internal step-by-step thinking before final outputs.

API integration follows standard patterns with optional OpenRouter-specific headers for leaderboard attribution. Third-party SDK support is available through OpenRouter's framework documentation.

What This Means

Grok 4.20 enters a competitive market where context window size has become a commodity feature—Claude 3.5 Sonnet and other models already offer 200K context, and some specialty models exceed 1M. The 2M window is a notable advantage but not unprecedented. The model's value proposition rests on claimed speed advantages and agentic capabilities rather than raw context size alone. The optional reasoning mode, similar to features in newer OpenAI and Anthropic models, allows developers to choose between faster inference and detailed reasoning transparency based on use case requirements. Pricing at $2/$6 positions it at the premium end of the market—significantly higher than open-source alternatives but within range of other flagship models. The lack of disclosed benchmark scores leaves performance claims unverified; independent evaluation will be necessary to validate xAI's claims about hallucination rates and prompt adherence relative to competitors.

Related Articles

model release

xAI releases Grok 4.20 Multi-Agent with 2M context window and parallel agent reasoning

xAI has released Grok 4.20 Multi-Agent, a variant designed for collaborative agent-based workflows with a 2-million-token context window. The model scales from 4 agents at low/medium reasoning effort to 16 agents at high/xhigh effort levels, priced at $2 per million input tokens and $6 per million output tokens.

model release

Anthropic confirms leaked model represents major reasoning advance after security breach

A data breach at Anthropic exposed internal documents detailing an unreleased AI model the company describes as its most powerful to date. Anthropic confirmed it is already testing the model with select customers, claiming significant advances in reasoning, coding, and cybersecurity. The breach resulted from a misconfiguration in Anthropic's content management system that automatically made ~3,000 uploaded files publicly accessible.

model release

Google launches Veo 3.1 Lite, cutting video generation costs by half

Google announced Veo 3.1 Lite, a cost-reduced video generation model priced at less than 50% of Veo 3.1 Fast's cost. The model supports text-to-video and image-to-video generation at 720p or 1080p resolution with customizable durations of 4s, 6s, or 8s, rolling out today on the Gemini API and Google AI Studio.

model release

Google releases Lyria 3 Clip Preview for music generation via API

Google has released Lyria 3 Clip Preview, a music generation model available through the Gemini API as of March 30, 2026. The model generates 30-second audio clips from text prompts or images at $0.04 per clip, with a 1,048,576 token context window.

Comments

Loading...