Qwen3.5-Flash

Alibaba / Qwen🇨🇳 China
active

Qwen3.5 native vision-language Flash model with hybrid linear attention and sparse MoE architecture. 1M context with ultra-low pricing, a major leap forward over the Qwen 3 series in inference efficiency.

Context window1000K tokens
Input / 1M tokens$0.065
Output / 1M tokens$0.26

Version History

qwen3.5-flash-2026-02-25major

Qwen3.5-Flash debuts with 1M context and ultra-low $0.065/$0.26 pricing. Hybrid architecture delivers a leap in inference efficiency over Qwen 3 series.