Qwen3.6 27B (Reasoning)
描述
Qwen3.6-27B is a dense 27-billion-parameter multimodal model in the Qwen3.6 series, supporting both vision-language thinking and non-thinking modes in a single unified checkpoint. The 64-layer language model uses a hybrid layout of 16 repeats of (3 × Gated DeltaNet → FFN, 1 × Gated Attention → FFN) with hidden dim 5120 and FFN intermediate 17408 — Gated DeltaNet has 48/16 heads for V/QK (head dim 128) and Gated Attention has 24/4 heads for Q/KV (head dim 256). It supports a native 262,144-token context extensible to ~1,010,000 via YaRN and is trained with multi-token prediction. The release delivers flagship-level agentic coding, surpassing the previous-generation open-source flagship Qwen3.5-397B-A17B (397B total / 17B active) on every major coding benchmark including SWE-bench Verified (77.2), SWE-bench Pro (53.5), Terminal-Bench 2.0 (59.3), and SkillsBench (48.2), and reaches 87.8 on GPQA Diamond. Released as open weights under Apache 2.0; accessible via Qwen Studio with the Alibaba Cloud Model Studio API coming soon.
能力雷达图
Science 在缺少专门科学评测时使用推理能力代理估算。
排行榜排名
基准测试分数 (LLM Stats)
Agents
Biology
Chemistry
Code
Embodied
Finance
General
Grounding
Healthcare
Image To Text
Long Context
Math
Multimodal
Reasoning
Spatial Reasoning
AA 评测指数
LLM Stats 分类评分
定价
速度
可用提供商
(LS 内部计价单位)| 提供商 | 输入价格 | 输出价格 |
|---|---|---|
| Novita | 600K | 3.6M |