Mistral Small 3.2

Mistral AI · Mistral Small

Mistral's open 24B model balancing strong instruction quality with low API cost for production assistants.

Type
language
Context
128K tokens
Max Output
128K tokens
Status
current
Input
$0.1/1M tok
Output
$0.3/1M tok
API Access
Yes
License
Apache 2.0
western europe open-weights coding instruction cost-efficient
Released June 2025 · Updated March 1, 2026

Overview

Freshness note: Model capabilities, limits, and pricing can change quickly. This profile is a point-in-time snapshot last verified on March 1, 2026.

Mistral Small 3.2 is Mistral AI’s open-weight 24B-class model tuned for production-grade instruction following and coding support at low operating cost. It is one of the more practical Western open-weight defaults when teams need a balance of quality, controllability, and throughput.

In the broader lineup, it sits below Mistral Large tiers on absolute peak capability, but often wins on cost-latency tradeoffs and deployment flexibility.

Capabilities

Mistral Small 3.2 is strongest for everyday assistant workloads: coding help, structured content generation, extraction/classification, and tool-augmented workflow orchestration. It is especially useful when response consistency and cost predictability are more important than frontier benchmark peaks.

For multilingual products, it handles major European languages well while remaining lightweight enough for self-hosted or high-volume API routing.

Technical Details

Official Mistral docs list a 128K context window for Small 3.2. The model is distributed as open weights under Apache 2.0, with optimized inference guidance for common serving stacks.

Compared with many larger frontier models, Small 3.2 usually requires more retrieval/prompt structure for difficult reasoning tasks, but that tradeoff is often acceptable in production systems that prioritize stable cost envelopes.

Pricing & Access

Published Mistral API pricing for Small 3.2 is:

  • Input: $0.10 per 1M tokens
  • Output: $0.30 per 1M tokens

Access paths include Mistral’s API platform and official open-weight releases for self-managed deployments.

Best Use Cases

Use Mistral Small 3.2 for high-throughput assistant backends, code review helpers, enterprise summarization/extraction pipelines, and agent systems where model unit cost is a hard constraint.

It is also a strong Western complement to Chinese open-weight models in multi-region routing designs.

Comparisons

  • Llama 4 Scout (Meta): Scout is a strong open option with Meta ecosystem gravity; Small 3.2 is often simpler to cost and operate for API-first teams.
  • Qwen3.5 (Alibaba): Qwen3.5 is stronger for Chinese-first bilingual workloads; Small 3.2 is a solid Western default for EU/global deployments.
  • DeepSeek-R1 (DeepSeek): DeepSeek can outperform on some reasoning tasks per dollar; Small 3.2 is attractive when Apache 2.0 licensing and Mistral ecosystem alignment are priorities.