Skip to main content

GLM-5 (Reasoning)

Z AIGLMOpen WeightMIT · Commercial OK

Description

GLM-5 is Zhipu AI's flagship foundation model designed for complex system engineering and long-range Agent tasks, shifting focus from coding to engineering. It features 744B total parameters (40B activated) in a Mixture of Experts architecture, trained on 28.5T tokens. GLM-5 integrates DeepSeek Sparse Attention for higher token efficiency while preserving long-context quality. It supports 200K context length and 128K max output tokens, with capabilities including thinking modes, real-time streaming, function calling, context caching, and structured output. GLM-5 approaches Claude Opus 4.5 in code-logic density and systems-engineering capability.

Release Date
2026-02-11
Parameters
744.0B
Context Length
203K
Modalities
text

Capability Radar

45
general
45
coding
82
reasoning
59
scienceest.
70
agents
0
multimodal

Science uses a reasoning proxy when dedicated science benchmarks are unavailable.

Rankings

Domain#RankScoreSource
Agents & Tools14
71.0
LS
Code Ranking34
76.0
AA
General Ranking19
87.0
AA
Science34
75.0
AA

Benchmark Scores (LLM Stats)

Agents

t2-bench89.7%SR
BrowseComp75.9%SR
MCP Atlas67.8%SR
Terminal-Bench 2.056.2%SR

Code

SWE-Bench Verified77.8%SR

AA Evaluation Indices

Intelligence Index
49.8
Coding Index
44.2
Tau2
1.0
Gpqa
0.8
Ifbench
0.7
Lcr
0.6
Scicode
0.5
Terminalbench Hard
0.4
Hle
0.3

LLM Stats Category Scores

Frontend Development
80
Search
80
Tool Calling
70
Agents
70
Code
70
Reasoning
70

Pricing

Input Price$1 / 1M tokens
Output Price$3.2 / 1M tokens
Blended Price (3:1)$1.55 / 1M tokens

Speed

Tokens/sec67.1 tokens/s
Time to First Token0.68s
Time to Answer46.97s

Available Providers

(LS internal units)
ProviderInput PriceOutput Price
ZAI1.0M3.2M

External Sources