model release

Alibaba's Qwen3.6 Plus reaches 78.8 on SWE-bench with 1M context window

TL;DR

Alibaba released Qwen3.6 Plus on April 2, 2026, featuring a 1 million token context window at $0.50 per million input tokens and $3 per million output tokens. The model combines linear attention with sparse mixture-of-experts routing to achieve a 78.8 score on SWE-bench Verified, with significant improvements in agentic coding, front-end development, and reasoning tasks.

2 min read
0

Qwen 3.6 Plus — Quick Specs

Context window1000K tokens
Input$0.5/1M tokens
Output$3/1M tokens

Alibaba's Qwen3.6 Plus Reaches 78.8 on SWE-bench with Million-Token Context

Alibaba released Qwen3.6 Plus on April 2, 2026, introducing a model that combines efficient linear attention with sparse mixture-of-experts routing to handle complex reasoning and coding tasks at scale.

Key Specifications

Context and Pricing:

  • Context window: 1,000,000 tokens
  • Input pricing: $0.50 per million tokens
  • Output pricing: $3.00 per million tokens

Performance: The model achieves a 78.8 score on SWE-bench Verified, positioning it alongside leading state-of-the-art models. Alibaba claims major improvements over the 3.5 series in agentic coding, front-end development, and overall reasoning capabilities.

Architecture and Capabilities

Qwen3.6 Plus uses a hybrid architecture combining:

  • Efficient linear attention mechanisms for scalability
  • Sparse mixture-of-experts routing for high-performance inference

According to Alibaba, the model excels at complex tasks including 3D scene generation, game development, and repository-level problem solving. The company describes particular improvements in "vibe coding experience," though this term lacks precise technical definition.

The model claims substantial performance gains in both pure-text and multimodal tasks, though specific benchmark comparisons to previous versions remain undisclosed.

Data Collection Notice

Alibaba explicitly states that Qwen3.6 Plus collects prompt and completion data for model improvement purposes. Users should review privacy implications before deploying in sensitive applications.

Deployment Availability

Qwen3.6 Plus is available through OpenRouter, which routes requests across multiple providers to optimize for context window support and uptime. The platform provides normalized API access across providers and supports reasoning-enabled inference with step-by-step thinking visibility.

What This Means

Qwen3.6 Plus positions Alibaba's Qwen line as a competitive option in the 1M-context segment, matching context window sizes offered by Claude 3.5 and other leaders. The 78.8 SWE-bench score places it in the tier of capable coding models, though detailed comparisons to other 1M-context models remain unavailable. Pricing at $0.50/$3.00 per million tokens is competitive with other high-context models. The explicit data collection policy requires careful consideration for enterprises handling proprietary code or sensitive information.

Related Articles

model release

GLM-5.1 achieves 58.4% on SWE-Bench Pro with sustained agentic reasoning over hundreds of iterations

Zhipu AI has released GLM-5.1, a 754-billion parameter model designed for agentic engineering with significantly improved coding capabilities over its predecessor. The model achieves 58.4% on SWE-Bench Pro and demonstrates sustained performance improvement over hundreds of tool calls and iterations, unlike earlier models that plateau quickly.

model release

Z.ai releases GLM-5.1, 754B parameter open-weight model with improved code generation

Z.ai has released GLM-5.1, a 754-billion parameter open-weight model matching the size of its predecessor GLM-5. The model demonstrates improved ability to generate complex, multi-part outputs like HTML pages with SVG graphics and CSS animations, available via Hugging Face and OpenRouter.

model release

Google releases Gemma 4 26B with 256K context and multimodal support, free to use

Google DeepMind has released Gemma 4 26B A4B, a free instruction-tuned Mixture-of-Experts model with 262,144 token context window and multimodal capabilities including text, images, and video input. Despite 25.2B total parameters, only 3.8B activate per token, delivering performance comparable to larger 31B models at reduced compute cost.

model release

Anthropic unveils Claude Mythos model, finds thousands of OS vulnerabilities via Project Glasswing

Anthropic has unveiled Claude Mythos, a new AI model designed for cybersecurity that has already discovered thousands of high-severity vulnerabilities in every major operating system and web browser. The model is being distributed as a preview to over 40 organizations and major technology partners including Apple, Google, Microsoft, and Amazon Web Services through Project Glasswing, a coordinated cybersecurity initiative.

Comments

Loading...