Alibaba's Qwen3.6 Plus reaches 78.8 on SWE-bench with 1M context window
Alibaba released Qwen3.6 Plus on April 2, 2026, featuring a 1 million token context window at $0.50 per million input tokens and $3 per million output tokens. The model combines linear attention with sparse mixture-of-experts routing to achieve a 78.8 score on SWE-bench Verified, with significant improvements in agentic coding, front-end development, and reasoning tasks.
Qwen 3.6 Plus — Quick Specs
Alibaba's Qwen3.6 Plus Reaches 78.8 on SWE-bench with Million-Token Context
Alibaba released Qwen3.6 Plus on April 2, 2026, introducing a model that combines efficient linear attention with sparse mixture-of-experts routing to handle complex reasoning and coding tasks at scale.
Key Specifications
Context and Pricing:
- Context window: 1,000,000 tokens
- Input pricing: $0.50 per million tokens
- Output pricing: $3.00 per million tokens
Performance: The model achieves a 78.8 score on SWE-bench Verified, positioning it alongside leading state-of-the-art models. Alibaba claims major improvements over the 3.5 series in agentic coding, front-end development, and overall reasoning capabilities.
Architecture and Capabilities
Qwen3.6 Plus uses a hybrid architecture combining:
- Efficient linear attention mechanisms for scalability
- Sparse mixture-of-experts routing for high-performance inference
According to Alibaba, the model excels at complex tasks including 3D scene generation, game development, and repository-level problem solving. The company describes particular improvements in "vibe coding experience," though this term lacks precise technical definition.
The model claims substantial performance gains in both pure-text and multimodal tasks, though specific benchmark comparisons to previous versions remain undisclosed.
Data Collection Notice
Alibaba explicitly states that Qwen3.6 Plus collects prompt and completion data for model improvement purposes. Users should review privacy implications before deploying in sensitive applications.
Deployment Availability
Qwen3.6 Plus is available through OpenRouter, which routes requests across multiple providers to optimize for context window support and uptime. The platform provides normalized API access across providers and supports reasoning-enabled inference with step-by-step thinking visibility.
What This Means
Qwen3.6 Plus positions Alibaba's Qwen line as a competitive option in the 1M-context segment, matching context window sizes offered by Claude 3.5 and other leaders. The 78.8 SWE-bench score places it in the tier of capable coding models, though detailed comparisons to other 1M-context models remain unavailable. Pricing at $0.50/$3.00 per million tokens is competitive with other high-context models. The explicit data collection policy requires careful consideration for enterprises handling proprietary code or sensitive information.
Related Articles
GLM-5.1 achieves 58.4% on SWE-Bench Pro with sustained agentic reasoning over hundreds of iterations
Zhipu AI has released GLM-5.1, a 754-billion parameter model designed for agentic engineering with significantly improved coding capabilities over its predecessor. The model achieves 58.4% on SWE-Bench Pro and demonstrates sustained performance improvement over hundreds of tool calls and iterations, unlike earlier models that plateau quickly.
Z.ai releases GLM-5.1, 754B parameter open-weight model with improved code generation
Z.ai has released GLM-5.1, a 754-billion parameter open-weight model matching the size of its predecessor GLM-5. The model demonstrates improved ability to generate complex, multi-part outputs like HTML pages with SVG graphics and CSS animations, available via Hugging Face and OpenRouter.
Google releases Gemma 4 26B with 256K context and multimodal support, free to use
Google DeepMind has released Gemma 4 26B A4B, a free instruction-tuned Mixture-of-Experts model with 262,144 token context window and multimodal capabilities including text, images, and video input. Despite 25.2B total parameters, only 3.8B activate per token, delivering performance comparable to larger 31B models at reduced compute cost.
Anthropic unveils Claude Mythos model, finds thousands of OS vulnerabilities via Project Glasswing
Anthropic has unveiled Claude Mythos, a new AI model designed for cybersecurity that has already discovered thousands of high-severity vulnerabilities in every major operating system and web browser. The model is being distributed as a preview to over 40 organizations and major technology partners including Apple, Google, Microsoft, and Amazon Web Services through Project Glasswing, a coordinated cybersecurity initiative.
Comments
Loading...