Back
Mistral: Mistral Nemo
Mistral
Input: text
Output: text
Released: Jul 19, 2024•Updated: Mar 28, 2025
A 12B parameter model with a 128k token context length built by Mistral in collaboration with NVIDIA.
The model is multilingual, supporting English, French, German, Spanish, Italian, Portuguese, Chinese, Japanese, Korean, Arabic, and Hindi.
It supports function calling and is released under the Apache 2.0 license.
131,072 Token Context
Process and analyze large documents and conversations.
Advanced Coding
Improved capabilities in front-end development and full-stack updates.
Agentic Workflows
Autonomously navigate multi-step processes with improved reliability.
Available On
Provider | Model ID | Context | Max Output | Input Cost | Output Cost | Throughput | Latency |
---|---|---|---|---|---|---|---|
Kluster | klusterAi | 131K | 131K | $0.01/M | $0.03/M | 94.0 t/s | 918 ms |
DeepInfra | deepInfra | 131K | 16K | $0.01/M | $0.03/M | 43.6 t/s | 308 ms |
Enfer | enfer | 131K | 66K | $0.02/M | $0.07/M | 36.6 t/s | 1655 ms |
NextBit | nextBit | 128K | - | $0.03/M | $0.07/M | 38.3 t/s | 1712 ms |
InferenceNet | inferenceNet | 16K | 16K | $0.04/M | $0.10/M | 63.3 t/s | 1008 ms |
Parasail | parasail | 131K | 131K | $0.04/M | $0.11/M | 79.2 t/s | 722 ms |
Nebius | nebiusAiStudio | 128K | - | $0.04/M | $0.12/M | 43.1 t/s | 468 ms |
Novita | novitaAi | 60K | 32K | $0.04/M | $0.17/M | 55.8 t/s | 1037 ms |
Atoma | atoma | 128K | 80K | $0.10/M | $0.10/M | 73.8 t/s | 669 ms |
InoCloud | inoCloud | 131K | 131K | $0.14/M | $0.14/M | 100.9 t/s | 1243 ms |
Mistral | mistral | 131K | - | $0.15/M | $0.15/M | 119.9 t/s | 245 ms |
Azure | azure | 128K | - | $0.30/M | $0.30/M | 100.3 t/s | 1144 ms |
Standard Pricing
Input Tokens
$0.00000001
per 1K tokens
Output Tokens
$0.000000029
per 1K tokens