Qwen3.5 35B A3B (Non-reasoning)
AlibabaQwenOpen WeightApache 2.0 · Commercial OK
Descripción
Qwen3.5-35B-A3B is a multimodal Mixture-of-Experts model with 35 billion total parameters and 3 billion activated parameters. It combines strong reasoning, coding, agentic, and visual understanding performance with production-friendly efficiency and a native 262K context window.
Fecha de lanzamiento
2026-02-24
Parámetros
35.0B
Longitud del contexto
262K
Modalidades
image, text, video
Radar de capacidades
27
general
19
coding
82
reasoning
50
scienceest.
60
agents
80
multimodal
Science usa un proxy de razonamiento cuando los benchmarks científicos dedicados no están disponibles.
Rankings
| Dominio | #Posición | Puntuación | Fuente |
|---|---|---|---|
| Agents & Tools | 53 | 56.0 | LS |
| Code Ranking | 227 | 37.0 | AA |
| General Ranking | 157 | 57.0 | AA |
| Multimodal Ranking | 62 | 69.0 | LS |
| Reasoning | 52 | 68.0 | LS |
| Science | 144 | 54.0 | AA |
Puntuaciones de benchmarks (LLM Stats)
3d
SUNRGBD
0.33 / 100Aut.
Hypersim
0.13 / 100Aut.
Agents
t2-bench
81.2%Aut.
AndroidWorld_SR
71.1%Aut.
BFCL-V4
67.3%Aut.
BrowseComp
61.0%Aut.
FullStackBench en
58.1%Aut.
WideSearch
57.1%Aut.
TIR-Bench
55.5%Aut.
FullStackBench zh
55.0%Aut.
OSWorld-Verified
54.5%Aut.
Terminal-Bench 2.0
40.5%Aut.
VITA-Bench
31.9%Aut.
DeepPlanning
22.8%Aut.
Biology
GPQA
84.2%Aut.
Chemistry
SuperGPQA
63.4%Aut.
Code
SWE-Bench Verified
69.2%Aut.
Communication
Multi-Challenge
60.0%Aut.
Embodied
EmbSpatialBench
0.83 / 100Aut.
Finance
MMLU-Pro
85.3%Aut.
MMLU-ProX
81.0%Aut.
General
MMLU-Redux
93.3%Aut.
IFEval
91.9%Aut.
C-Eval
90.2%Aut.
MAXIFE
86.6%Aut.
Global PIQA
86.6%Aut.
MMMLU
85.2%Aut.
MMStar
81.9%Aut.
MMMU
81.4%Aut.
Include
79.7%Aut.
MMMU-Pro
75.1%Aut.
LiveCodeBench v6
74.6%Aut.
IFBench
70.2%Aut.
LongBench v2
59.0%Aut.
SimpleVQA
0.58 / 100Aut.
NOVA-63
57.1%Aut.
Grounding
RefCOCO-avg
0.89 / 100Aut.
ScreenSpot Pro
68.6%Aut.
RefSpatialBench
0.64 / 100Aut.
Healthcare
VideoMMMU
80.4%Aut.
SlakeVQA
78.7%Aut.
PMC-VQA
62.0%Aut.
MedXpertQA
61.4%Aut.
Image To Text
OCRBench
91.0%Aut.
Language
LingoQA
79.2%Aut.
WMT24++
76.3%Aut.
Long Context
MLVU
85.6%Aut.
LVBench
71.4%Aut.
MMLongBench-Doc
0.59 / 100Aut.
AA-LCR
58.5%Aut.
Math
HMMT25
89.2%Aut.
HMMT 2025
89.0%Aut.
MathVista-Mini
86.2%Aut.
DynaMath
85.0%Aut.
MathVision
83.9%Aut.
CodeForces
0.82 / 3000Aut.
PolyMATH
64.4%Aut.
Humanity's Last Exam
47.4%Aut.
Multimodal
VLMsAreBlind
97.0%Aut.
V*
92.7%Aut.
AI2D
92.6%Aut.
MMBench-V1.1
91.5%Aut.
OmniDocBench 1.5
89.3%Aut.
VideoMME w sub.
86.6%Aut.
VideoMME w/o sub.
82.5%Aut.
CC-OCR
80.7%Aut.
CharXiv-R
77.5%Aut.
MVBench
74.8%Aut.
MMVU
72.3%Aut.
BabyVision
38.4%Aut.
ZEROBench-Sub
0.34 / 100Aut.
Nuscene
14.6%Aut.
ZEROBench
0.08 / 100Aut.
Reasoning
CountBench
0.98 / 100Aut.
BrowseComp-zh
69.5%Aut.
Hallusion Bench
67.9%Aut.
ERQA
64.8%Aut.
Seal-0
41.4%Aut.
OJBench
36.0%Aut.
Spatial Reasoning
RealWorldQA
84.1%Aut.
Vision
ODinW
42.6%Aut.
Índices de evaluación AA
Intelligence Index30.7
Coding Index16.8
Tau20.9
Gpqa0.8
Lcr0.6
Ifbench0.4
Scicode0.3
Hle0.1
Terminalbench Hard0.1
Puntuaciones por categoría LLM Stats
Structured Output80
Text-to-image80
Video80
Biology80
Embodied80
Finance80
General80
Image To Text80
Instruction Following80
Language80
Legal80
Math80
Physics80
Spatial Reasoning70
Vision70
Chemistry70
Frontend Development70
Grounding70
Healthcare70
Long Context70
Multimodal70
Reasoning70
Tool Calling60
Code60
Communication60
Economics60
Search60
Agents50
3d20
Spatial10
Precios
Precio de entrada$0.25 / 1M tokens
Precio de salida$2 / 1M tokens
Precio mixto (3:1)$0.688 / 1M tokens
Velocidad
Tokens/seg138.7 tokens/s
Retraso del primer token1.15s
Tiempo hasta la respuesta1.15s
Proveedores disponibles
(Unidades internas LS)| Proveedor | Precio de entrada | Precio de salida |
|---|---|---|
| Novita | 250K | 2.0M |