DeepSeek V4 Flash (Reasoning, Max Effort)
Description
DeepSeek-V4-Flash-Max is the maximum reasoning effort mode of DeepSeek-V4-Flash, a 284B-parameter MoE model with 13B activated parameters and a 1M-token context window. Sharing the V4 series' hybrid attention architecture (Compressed Sparse Attention combined with Heavily Compressed Attention), Manifold-Constrained Hyper-Connections, and Muon optimizer, V4-Flash-Max delivers reasoning performance comparable to V4-Pro when given a larger thinking budget while operating at a fraction of the parameter scale. It is pre-trained on more than 32T tokens and post-trained with a two-stage paradigm of domain-specific expert cultivation followed by on-policy distillation.
Radar de capacités
Science utilise un proxy de raisonnement lorsque les benchmarks scientifiques dédiés ne sont pas disponibles.
Classements
| Domaine | #Rang | Score | Source |
|---|---|---|---|
| Agents & Tools | 49 | 57.0 | LS |
| Code Ranking | 64 | 68.0 | AA |
| General Ranking | 24 | 85.0 | AA |
| Science | 25 | 80.0 | AA |
Scores de benchmarks (LLM Stats)
Agents
Biology
Code
Factuality
Finance
General
Math
Indices d'évaluation AA
Scores par catégorie LLM Stats
Tarification
Vitesse
Fournisseurs disponibles
(Unités internes LS)| Fournisseur | Prix d'entrée | Prix de sortie |
|---|---|---|
| DeepSeek | 140K | 280K |