Skip to main content

GLM-4.5V (Non-reasoning)

Z AIGLMOpen WeightMIT · Commercial OK

Description

GLM-4.5V is a multimodal (vision-language) model based on GLM-4.5-Air (106B total, 12B active) that extends hybrid reasoning to images and video. It achieves state-of-the-art results across 40+ VLM benchmarks (image reasoning, video understanding, GUI tasks, chart/document parsing, grounding) while supporting a Thinking Mode switch for deep reasoning. Released under MIT with FP8/BF16 variants and tooling in Transformers, vLLM, and SGLang.

Release Date
2025-08-11
Parameters
108.0B
Context Length
66K
Modalities
image, text, video

Capability Radar

30
general
20
coding
22
reasoning
33
scienceest.
0
agents
90
multimodal

Science uses a reasoning proxy when dedicated science benchmarks are unavailable.

Rankings

Domain#RankScoreSource
Code Ranking368
17.0
AA
General Ranking363
31.0
AA
Math Reasoning321
15.0
AA
Science353
31.0
AA

Benchmark Scores (LLM Stats)

No benchmark data available

AA Evaluation Indices

Math Index
15.3
Intelligence Index
12.7
Coding Index
10.8
Mmlu Pro
0.8
Gpqa
0.6
Livecodebench
0.4
Ifbench
0.3
Tau2
0.2
Scicode
0.2
Aime 25
0.2
Terminalbench Hard
0.1
Hle
0.0
Lcr
0.0

LLM Stats Category Scores

No category score data available

Pricing

Input Price$0.6 / 1M tokens
Output Price$1.8 / 1M tokens
Blended Price (3:1)$0.9 / 1M tokens

Speed

Tokens/sec48.4 tokens/s
Time to First Token37.83s
Time to Answer37.83s

Available Providers

(LS internal units)

No provider data available

External Sources