Back

Mistral: Mistral Nemo

Mistral
Input: text
Output: text
Released: Jul 19, 2024Updated: Mar 28, 2025

A 12B parameter model with a 128k token context length built by Mistral in collaboration with NVIDIA.

The model is multilingual, supporting English, French, German, Spanish, Italian, Portuguese, Chinese, Japanese, Korean, Arabic, and Hindi.

It supports function calling and is released under the Apache 2.0 license.

131,072 Token Context

Process and analyze large documents and conversations.

Advanced Coding

Improved capabilities in front-end development and full-stack updates.

Agentic Workflows

Autonomously navigate multi-step processes with improved reliability.

Available On

ProviderModel IDContextMax OutputInput CostOutput CostThroughputLatency
kluster.aiklusterAi131K131K$0.02/M$0.07/M107.3 t/s738 ms
Enferenfer131K66K$0.03/M$0.07/M58.2 t/s6404 ms
NextBitnextBit128K-$0.03/M$0.07/M42.8 t/s1556 ms
DeepInfradeepInfra131K16K$0.04/M$0.08/M54.9 t/s251 ms
inference.netinferenceNet16K16K$0.04/M$0.10/M59.6 t/s992 ms
Nebius AI StudionebiusAiStudio128K-$0.04/M$0.12/M33.1 t/s603 ms
NovitaAInovitaAi131K-$0.04/M$0.17/M42.1 t/s1173 ms
Atomaatoma128K80K$0.10/M$0.10/M60.8 t/s646 ms
Parasailparasail131K131K$0.11/M$0.11/M90.9 t/s747 ms
Mistralmistral131K-$0.15/M$0.15/M143.8 t/s219 ms
Azureazure128K-$0.30/M$0.30/M98.6 t/s1146 ms
Standard Pricing
Input Tokens
$0.000000025

per 1K tokens

Output Tokens
$0.00000007

per 1K tokens

Do Work. With AI.