Qwen3.5 35B A3B (Non-reasoning)
AlibabaQwenOpen WeightApache 2.0 · Commercial OK
Description
Qwen3.5-35B-A3B is a multimodal Mixture-of-Experts model with 35 billion total parameters and 3 billion activated parameters. It combines strong reasoning, coding, agentic, and visual understanding performance with production-friendly efficiency and a native 262K context window.
Date de sortie
2026-02-24
Paramètres
35.0B
Longueur du contexte
262K
Modalités
image, text, video
Radar de capacités
27
general
19
coding
82
reasoning
50
scienceest.
60
agents
80
multimodal
Science utilise un proxy de raisonnement lorsque les benchmarks scientifiques dédiés ne sont pas disponibles.
Classements
| Domaine | #Rang | Score | Source |
|---|---|---|---|
| Agents & Tools | 53 | 56.0 | LS |
| Code Ranking | 227 | 37.0 | AA |
| General Ranking | 157 | 57.0 | AA |
| Multimodal Ranking | 62 | 69.0 | LS |
| Reasoning | 52 | 68.0 | LS |
| Science | 144 | 54.0 | AA |
Scores de benchmarks (LLM Stats)
3d
SUNRGBD
0.33 / 100Aut.
Hypersim
0.13 / 100Aut.
Agents
t2-bench
81.2%Aut.
AndroidWorld_SR
71.1%Aut.
BFCL-V4
67.3%Aut.
BrowseComp
61.0%Aut.
FullStackBench en
58.1%Aut.
WideSearch
57.1%Aut.
TIR-Bench
55.5%Aut.
FullStackBench zh
55.0%Aut.
OSWorld-Verified
54.5%Aut.
Terminal-Bench 2.0
40.5%Aut.
VITA-Bench
31.9%Aut.
DeepPlanning
22.8%Aut.
Biology
GPQA
84.2%Aut.
Chemistry
SuperGPQA
63.4%Aut.
Code
SWE-Bench Verified
69.2%Aut.
Communication
Multi-Challenge
60.0%Aut.
Embodied
EmbSpatialBench
0.83 / 100Aut.
Finance
MMLU-Pro
85.3%Aut.
MMLU-ProX
81.0%Aut.
General
MMLU-Redux
93.3%Aut.
IFEval
91.9%Aut.
C-Eval
90.2%Aut.
MAXIFE
86.6%Aut.
Global PIQA
86.6%Aut.
MMMLU
85.2%Aut.
MMStar
81.9%Aut.
MMMU
81.4%Aut.
Include
79.7%Aut.
MMMU-Pro
75.1%Aut.
LiveCodeBench v6
74.6%Aut.
IFBench
70.2%Aut.
LongBench v2
59.0%Aut.
SimpleVQA
0.58 / 100Aut.
NOVA-63
57.1%Aut.
Grounding
RefCOCO-avg
0.89 / 100Aut.
ScreenSpot Pro
68.6%Aut.
RefSpatialBench
0.64 / 100Aut.
Healthcare
VideoMMMU
80.4%Aut.
SlakeVQA
78.7%Aut.
PMC-VQA
62.0%Aut.
MedXpertQA
61.4%Aut.
Image To Text
OCRBench
91.0%Aut.
Language
LingoQA
79.2%Aut.
WMT24++
76.3%Aut.
Long Context
MLVU
85.6%Aut.
LVBench
71.4%Aut.
MMLongBench-Doc
0.59 / 100Aut.
AA-LCR
58.5%Aut.
Math
HMMT25
89.2%Aut.
HMMT 2025
89.0%Aut.
MathVista-Mini
86.2%Aut.
DynaMath
85.0%Aut.
MathVision
83.9%Aut.
CodeForces
0.82 / 3000Aut.
PolyMATH
64.4%Aut.
Humanity's Last Exam
47.4%Aut.
Multimodal
VLMsAreBlind
97.0%Aut.
V*
92.7%Aut.
AI2D
92.6%Aut.
MMBench-V1.1
91.5%Aut.
OmniDocBench 1.5
89.3%Aut.
VideoMME w sub.
86.6%Aut.
VideoMME w/o sub.
82.5%Aut.
CC-OCR
80.7%Aut.
CharXiv-R
77.5%Aut.
MVBench
74.8%Aut.
MMVU
72.3%Aut.
BabyVision
38.4%Aut.
ZEROBench-Sub
0.34 / 100Aut.
Nuscene
14.6%Aut.
ZEROBench
0.08 / 100Aut.
Reasoning
CountBench
0.98 / 100Aut.
BrowseComp-zh
69.5%Aut.
Hallusion Bench
67.9%Aut.
ERQA
64.8%Aut.
Seal-0
41.4%Aut.
OJBench
36.0%Aut.
Spatial Reasoning
RealWorldQA
84.1%Aut.
Vision
ODinW
42.6%Aut.
Indices d'évaluation AA
Intelligence Index30.7
Coding Index16.8
Tau20.9
Gpqa0.8
Lcr0.6
Ifbench0.4
Scicode0.3
Hle0.1
Terminalbench Hard0.1
Scores par catégorie LLM Stats
Structured Output80
Text-to-image80
Video80
Biology80
Embodied80
Finance80
General80
Image To Text80
Instruction Following80
Language80
Legal80
Math80
Physics80
Spatial Reasoning70
Vision70
Chemistry70
Frontend Development70
Grounding70
Healthcare70
Long Context70
Multimodal70
Reasoning70
Tool Calling60
Code60
Communication60
Economics60
Search60
Agents50
3d20
Spatial10
Tarification
Prix d'entrée$0.25 / 1M tokens
Prix de sortie$2 / 1M tokens
Prix mixte (3:1)$0.688 / 1M tokens
Vitesse
Tokens/sec138.7 tokens/s
Délai du premier token1.15s
Temps de réponse1.15s
Fournisseurs disponibles
(Unités internes LS)| Fournisseur | Prix d'entrée | Prix de sortie |
|---|---|---|
| Novita | 250K | 2.0M |