xAI releases Grok 4.3 reasoning model with 1M token context at $1.25/M input tokens
xAI has released Grok 4.3, a reasoning model with a 1 million token context window and no output token limit. The model accepts text and image inputs, has always-on reasoning that cannot be disabled, and uses tiered pricing starting at $1.25 per million input tokens and $2.50 per million output tokens.
Grok 4.3 — Quick Specs
xAI releases Grok 4.3 reasoning model with 1M token context at $1.25/M input tokens
xAI has released Grok 4.3, a multimodal reasoning model with a 1 million token context window and no output token limit. Released on April 30, 2026, the model is now available through OpenRouter.
Specifications and pricing
Grok 4.3 processes text and image inputs with text output. Input tokens are priced at $1.25 per million tokens, while output tokens cost $2.50 per million tokens. According to xAI, requests exceeding 200,000 total tokens are billed at a higher rate, though the elevated pricing tier has not been disclosed.
The model features always-on reasoning that cannot be disabled or configured by effort level. This distinguishes it from other reasoning models that allow users to adjust computational intensity.
Technical capabilities
xAI positions Grok 4.3 for agentic workflows, instruction-following tasks, and applications requiring high factual accuracy. The absence of an output token limit, combined with the 1 million token context window, enables the model to handle long-document analysis and multi-step agentic tasks without truncation.
The model supports multimodal input, accepting both text and images, but outputs text only.
API access
Grok 4.3 is accessible through OpenRouter's API, which normalizes requests and responses across providers. The platform routes requests to available providers and includes fallback mechanisms for uptime.
OpenRouter's API supports accessing the model's reasoning process through a reasoning_details array in responses. The platform requires preserving complete reasoning details when passing messages back to the model for continued conversations.
What this means
Grok 4.3 enters a competitive reasoning model market where always-on reasoning represents a trade-off: consistent step-by-step thinking for all queries, but no ability to reduce computational cost for simpler tasks. The 1M token context window and unlimited output position it for enterprise document analysis and complex multi-turn interactions. The tiered pricing structure for requests over 200K tokens suggests xAI expects the model to be used for extended contexts, though the lack of disclosed upper-tier pricing creates uncertainty for budget planning at scale.
Related Articles
NVIDIA Releases Nemotron 3 Nano Omni: 31B-Parameter Multimodal Model with 256K Context and Reasoning Mode
NVIDIA has released Nemotron 3 Nano Omni 30B-A3B, a multimodal large language model with 31 billion parameters using a Mamba2-Transformer hybrid Mixture of Experts architecture. The model supports video, audio, image, and text inputs with a 256K token context window and includes a dedicated reasoning mode with chain-of-thought capabilities.
OpenRouter Launches Owl Alpha: Free Foundation Model for Agentic Workflows with 1M Context
OpenRouter has released Owl Alpha, a foundation model specifically designed for agentic workloads with native tool use support and a 1,048,756 token context window. The model is currently free for both input and output tokens and is compatible with Claude Code, OpenClaw, and other productivity tools.
Mistral Releases Medium 3.5: 128B Dense Model With 256k Context and Configurable Reasoning
Mistral AI released Mistral Medium 3.5, a 128B parameter dense model with a 256k context window that unifies instruction-following, reasoning, and coding capabilities. The model features configurable reasoning effort per request and a vision encoder trained from scratch for variable image sizes.
NVIDIA Releases Nemotron 3 Nano Omni: 31B Multimodal Model With 256K Context and Reasoning Mode
NVIDIA released Nemotron 3 Nano Omni, a 31B parameter (30B active, 3B per token) multimodal model supporting video, audio, image, and text inputs. The model features a 256K token context window, reasoning mode with chain-of-thought, and tool calling capabilities.
Comments
Loading...