Back

Mistral: Ministral 8B

Mistral
Input: text
Output: text
Released: Oct 17, 2024Updated: Mar 28, 2025

Ministral 8B is an 8B parameter model featuring a unique interleaved sliding-window attention pattern for faster, memory-efficient inference. Designed for edge use cases, it supports up to 128k context length and excels in knowledge and reasoning tasks. It outperforms peers in the sub-10B category, making it perfect for low-latency, privacy-first applications.

128,000 Token Context

Process and analyze large documents and conversations.

Advanced Coding

Improved capabilities in front-end development and full-stack updates.

Agentic Workflows

Autonomously navigate multi-step processes with improved reliability.

Available On

ProviderModel IDContextMax OutputInput CostOutput CostThroughputLatency
Mistralmistral128K-$0.10/M$0.10/M124.5 t/s225 ms
Standard Pricing
Input Tokens
$0.0000001

per 1K tokens

Output Tokens
$0.0000001

per 1K tokens

Do Work. With AI.