Alibaba's Qwen Team Releases Qwen3.6 27B With 262K Context Window and Video Processing
Alibaba's Qwen Team has released Qwen3.6 27B, a 27-billion parameter multimodal language model with a 262,144-token context window. The model accepts text, image, and video inputs and includes a built-in thinking mode for extended reasoning, with pricing at $0.195 per million input tokens and $1.56 per million output tokens.
Qwen3.6 27B — Quick Specs
Alibaba's Qwen Team Releases Qwen3.6 27B With 262K Context Window and Video Processing
Alibaba's Qwen Team released Qwen3.6 27B on April 27, 2026, a 27-billion parameter language model with hybrid multimodal capabilities and a 262,144-token context window.
The model accepts text, image, and video inputs and is designed for agentic coding and reasoning tasks. According to the Qwen Team, it shows particular strength in repository-level code comprehension, front-end development workflows, and multi-step problem solving.
Key Specifications
- Parameters: 27 billion (dense architecture)
- Context window: 262,144 tokens
- Input modalities: Text, image, video
- Pricing: $0.195 per million input tokens, $1.56 per million output tokens
- Language support: 201 languages and dialects
- License: Apache 2.0
Technical Features
Qwen3.6 27B includes a built-in thinking mode for extended reasoning tasks. The model preserves thinking context across conversation history, allowing it to continue reasoning from previous exchanges. This feature is accessible through OpenRouter's reasoning parameter API.
The model is available through OpenRouter, which routes requests across multiple providers with automatic fallbacks to maximize uptime.
Positioning and Availability
With its 262K context window, Qwen3.6 27B competes directly with other long-context models in the market. The Apache 2.0 license allows for commercial use without restrictions.
The model is currently available via OpenRouter's API, with standard chat and reasoning endpoints. Model weights are also available for download.
What This Means
Qwen3.6 27B represents Alibaba's continued expansion in multimodal AI, adding video processing to its text and image capabilities. The 262K context window positions it competitively for document analysis and repository-scale coding tasks. At $0.195 per million input tokens, it's priced in the mid-range for models of this size, though benchmark scores were not disclosed at release. The Apache 2.0 license lowers barriers to adoption compared to more restrictive alternatives.
Related Articles
Alibaba Releases Qwen3.6 Max Preview: 1 Trillion Parameter MoE Model With 262K Context Window
Alibaba Cloud has released Qwen3.6 Max Preview, a proprietary frontier model built on sparse mixture-of-experts architecture with approximately 1 trillion total parameters. The model supports a 262,144-token context window and features integrated thinking mode for multi-turn reasoning, priced at $1.30 per million input tokens and $7.80 per million output tokens.
Alibaba Qwen Releases Qwen3.6 Flash with 1M Context Window at $0.25 per 1M Input Tokens
Alibaba's Qwen team has released Qwen3.6 Flash, a multimodal language model supporting text, image, and video input with a 1 million token context window. The model is priced at $0.25 per 1M input tokens and $1.50 per 1M output tokens, with tiered pricing above 256K tokens.
Alibaba Qwen Releases 35B Sparse MoE Model with 262K Context and Multimodal Support
Alibaba Cloud has released Qwen3.6-35B-A3B, an open-weight sparse mixture-of-experts model with 35 billion total parameters but only 3 billion active parameters per token. The model features a 262K native context window (expandable to 1M tokens), multimodal input support, and integrated reasoning mode with preserved thinking traces.
DeepSeek Releases V4-Flash: 284B-Parameter MoE Model With 1M Token Context at 27% Inference Cost
DeepSeek released two Mixture-of-Experts models: V4-Flash with 284B total parameters (13B activated) and V4-Pro with 1.6T parameters (49B activated). Both models support one million token context windows and use a hybrid attention architecture that requires only 27% of the inference FLOPs compared to DeepSeek-V3.2 at 1M token context.
Comments
Loading...