DeepSeek V4 Flash (Reasoning, Max Effort)
Description
DeepSeek-V4-Flash-Max is the maximum reasoning effort mode of DeepSeek-V4-Flash, a 284B-parameter MoE model with 13B activated parameters and a 1M-token context window. Sharing the V4 series' hybrid attention architecture (Compressed Sparse Attention combined with Heavily Compressed Attention), Manifold-Constrained Hyper-Connections, and Muon optimizer, V4-Flash-Max delivers reasoning performance comparable to V4-Pro when given a larger thinking budget while operating at a fraction of the parameter scale. It is pre-trained on more than 32T tokens and post-trained with a two-stage paradigm of domain-specific expert cultivation followed by on-policy distillation.
Capability Radar
Science uses a reasoning proxy when dedicated science benchmarks are unavailable.
Rankings
| Domain | #Rank | Score | Source |
|---|---|---|---|
| Agents & Tools | 49 | 57.0 | LS |
| Code Ranking | 64 | 68.0 | AA |
| General Ranking | 24 | 85.0 | AA |
| Science | 25 | 80.0 | AA |
Benchmark Scores (LLM Stats)
Agents
Biology
Code
Factuality
Finance
General
Math
AA Evaluation Indices
LLM Stats Category Scores
Pricing
Speed
Available Providers
(LS internal units)| Provider | Input Price | Output Price |
|---|---|---|
| DeepSeek | 140K | 280K |