Skip to main content

DeepSeek V3 (Dec '24)

DeepSeekDeepSeekOpen WeightMIT + Model License (Commercial use allowed)

Description

A powerful Mixture-of-Experts (MoE) language model with 671B total parameters (37B activated per token). Features Multi-head Latent Attention (MLA), auxiliary-loss-free load balancing, and multi-token prediction training. Pre-trained on 14.8T tokens with strong performance in reasoning, math, and code tasks.

Release Date
2024-12-26
Parameters
671.0B
Context Length
164K
Modalities
text

Capability Radar

32
general
25
coding
38
reasoning
38
scienceest.
0
agents
0
multimodal

Science uses a reasoning proxy when dedicated science benchmarks are unavailable.

Rankings

Domain#RankScoreSource
Code Ranking273
29.0
AA
General Ranking306
36.0
AA
Math Reasoning226
39.0
AA
Reasoning36
76.0
LS
Science272
40.0
AA

Benchmark Scores (LLM Stats)

Biology

GPQA59.1%SR

Code

Aider-Polyglot Edit79.7%SR
Aider-Polyglot49.6%SR
SWE-Bench Verified42.0%SR
LiveCodeBench37.6%SR

Factuality

SimpleQA24.9%SR

Finance

MMLU88.5%SR
MMLU-Pro75.9%SR

General

MMLU-Redux89.1%SR
C-Eval86.5%SR
IFEval86.1%SR
CSimpleQA64.8%SR
LongBench v248.7%SR

Language

CLUEWSC90.9%SR

Math

DROP91.6%SR
MATH-50090.2%SR
CNMO 202443.2%SR
AIME 202439.2%SR

Reasoning

HumanEval-Mul82.6%SR
FRAMES73.3%SR

AA Evaluation Indices

Math Index
26.0
Intelligence Index
16.5
Coding Index
16.4
Math 500
0.9
Mmlu Pro
0.8
Gpqa
0.6
Livecodebench
0.4
Scicode
0.4
Ifbench
0.3
Lcr
0.3
Aime 25
0.3
Aime
0.3
Tau2
0.2
Terminalbench Hard
0.1
Hle
0.0

LLM Stats Category Scores

Instruction Following
90
Finance
80
Healthcare
80
Language
80
Legal
80
Structured Output
70
General
70
Math
70
Reasoning
70
Biology
60
Chemistry
60
Physics
60
Code
50
Long Context
50
Frontend Development
40
Factuality
20

Pricing

Input Price$0.4 / 1M tokens
Output Price$0.89 / 1M tokens
Blended Price (3:1)$0.523 / 1M tokens

Speed

Tokens/sec0.0 tokens/s
Time to First Token0.00s
Time to Answer0.00s

Available Providers

(LS internal units)

No provider data available

External Sources