Qwen3.6 35B A3B (Reasoning)
Description
Qwen3.6-35B-A3B is the first open-weight variant of the Qwen3.6 series, a multimodal Mixture-of-Experts model with 35B total parameters and 3B activated. It pairs a vision encoder with a hybrid 40-layer language model that interleaves Gated DeltaNet linear-attention blocks and Gated Attention blocks (10 × (3 × DeltaNet + 1 × Attention)) over 256 experts (8 routed + 1 shared, expert dim 512). The release prioritizes stability and real-world utility, with substantial gains in agentic coding (frontend workflows, repo-level reasoning) and a new option to preserve reasoning context across turns. Native context length is 262K tokens, extensible to ~1M via YaRN, and the model thinks by default.
Capability Radar
Science uses a reasoning proxy when dedicated science benchmarks are unavailable.
Rankings
| Domain | #Rank | Score | Source |
|---|---|---|---|
| Agents & Tools | 77 | 47.0 | LS |
| Code Ranking | 81 | 65.0 | AA |
| General Ranking | 52 | 78.0 | AA |
| Multimodal Ranking | 34 | 79.0 | LS |
| Reasoning | 44 | 70.0 | LS |
| Science | 84 | 64.0 | AA |
Benchmark Scores (LLM Stats)
Agents
Biology
Chemistry
Code
Embodied
Finance
General
Grounding
Healthcare
Long Context
Math
Multimodal
Reasoning
Spatial Reasoning
Vision
AA Evaluation Indices
LLM Stats Category Scores
Pricing
Speed
Available Providers
(LS internal units)No provider data available