Passer au contenu principal

Mistral Large (Feb '24)

MistralMistralOpen WeightApache 2.0 · Commercial OK

Description

Mistral Large 3 (675B Instruct 2512) is a state-of-the-art general-purpose Multimodal granular Mixture-of-Experts model with 41B active parameters and 675B total parameters trained from scratch with 3000 H200s. This model is the instruct post-trained version in FP8, fine-tuned for instruction tasks, making it ideal for chat, agentic and instruction based use cases. A no-loss FP8 version to reduce resource requirements. Can be deployed on a node of B200s or H200s. Designed for reliability and long-context comprehension - It is engineered for production-grade assistants, retrieval-augmented systems, scientific workloads, and complex enterprise workflows.

Date de sortie
2024-02-26
Paramètres
675.0B
Longueur du contexte
128K
Modalités
image, text

Radar de capacités

21
general
18
coding
23
reasoning
24
scienceest.
0
agents
75
multimodal

Science utilise un proxy de raisonnement lorsque les benchmarks scientifiques dédiés ne sont pas disponibles.

Classements

Domaine#RangScoreSource
Code Ranking355
19.0
AA
General Ranking410
24.0
AA
Math Reasoning288
25.0
AA
Science404
23.0
AA

Scores de benchmarks (LLM Stats)

Biology

GPQA43.9%Aut.

Code

LiveCodeBench34.4%Aut.

Factuality

SimpleQA23.8%Aut.

General

MMMLU85.5%Aut.

Math

AMC_2022_2352.0%Aut.

Indices d'évaluation AA

Intelligence Index
9.9
Math 500
0.5
Mmlu Pro
0.5
Gpqa
0.4
Scicode
0.2
Livecodebench
0.2
Hle
0.0
Aime
0.0

Scores par catégorie LLM Stats

Language
90
Math
70
General
50
Reasoning
50
Biology
40
Chemistry
40
Physics
40
Code
30
Factuality
20

Tarification

Prix d'entrée$4 / 1M tokens
Prix de sortie$12 / 1M tokens
Prix mixte (3:1)$6 / 1M tokens

Vitesse

Tokens/sec0.0 tokens/s
Délai du premier token0.00s
Temps de réponse0.00s

Fournisseurs disponibles

(Unités internes LS)
FournisseurPrix d'entréePrix de sortie
Mistral AI500K1.5M

Sources externes