You’re staring at Mistral’s model list: Small, Medium, Large. Names that promise nothing about what they actually cost or what they’re good at. You pick Large because “more parameters = better,” run a batch job, and the bill is 3x what you budgeted.
The real question isn’t “what’s the biggest Mistral model?” It’s “which model solves my problem without burning cash?”
Mistral AI’s lineup spans 3 billion to 675 billion parameters (as of December 2025) – tiny edge models running on your phone to frontier reasoning models competing with GPT-4. Tutorials organize by model size. That’s useless. Here’s the map: task → model → cost.
Pick Your Model in Three Seconds
Simple chat, summaries, Q&A: Mistral Small 3.1 (24B). $0.02 input, cheap, fast.
Coding, agents, tool use: Devstral 2 (123B) or Devstral Small 2 (24B). Free via API as of December 2025.
Long documents, multilingual, vision: Mistral Medium 3 (released May 2025). $0.40 input, 128K context.
Complex reasoning, math, chain-of-thought: Magistral Medium or Mistral Large 3 (256K context, 41B active / 675B total params). $2.00 input.
Audio transcription: Voxtral Mini ($0.003/min) or Voxtral Realtime ($0.006/min).
Running offline or on-device: Ministral 3 (3B/8B/14B). Apache 2.0, fits on a laptop.
That’s the decision tree. Everything else is tuning.
What’s Different Here
Mistral AI: French startup founded April 2023 by ex-Google DeepMind and Meta researchers. Valuation hit $14 billion by September 2025.
The pitch? Open-weight models (weights published, Apache 2.0 license) you can run anywhere – your cloud, your data center, a Raspberry Pi if you pick the right size. No vendor lock-in. Europe-first privacy (GDPR-compliant by default). Cheaper than OpenAI for comparable tasks.
You’re not buying a polished product. You’re getting an engine. Integration, deployment, fine-tuning – that’s on you. Community feedback consistently mentions steep learning curve and slow support response times.
Have a dev team and want control? Mistral is one of the few credible alternatives to the OpenAI/Anthropic duopoly.
Model Breakdown (By What They Do)
Production Apps: Mistral Medium 3
Released May 7, 2025. $0.40 per million input tokens, $2.00 per million output. Context: 128K tokens.
This is the best balance for most business use cases. Handles coding, document analysis, summarization, multilingual chat, visual understanding (multimodal). Third-party analysis claims “90% of GPT-4-level reasoning for 20% of the cost” – math only holds if your workload is prompt-heavy and output-light. Generating long responses? Output pricing starts to matter.
128K context looks good until you hit the wall: prompt_tokens + max_tokens must stay under 128,000. Prompt is 100K tokens? You can only generate 28K tokens of output before the API rejects the request. No partial completion – just fails.
from mistralai import Mistral
import os
client = Mistral(api_key=os.environ["MISTRAL_API_KEY"])
response = client.chat.complete(
model="mistral-medium-latest",
messages=[{"role": "user", "content": "Summarize this 50-page document..."}],
max_tokens=4096
)
print(response.choices[0].message.content)
Good default choice. Scales well.
Reasoning Tasks: Magistral and Large 3
Mistral Large 3 (December 2, 2025): 41 billion active parameters, 675 billion total in mixture-of-experts architecture. Context: 256K tokens. Pricing: $2.00 input / $6.00 output per million tokens.
Magistral models (June 2025, detailed in arXiv:2506.10910) add chain-of-thought reasoning. Magistral Medium: proprietary. Magistral Small: open-source under Apache 2.0.
Pick these for step-by-step logic – math proofs, code debugging, complex legal analysis. Magistral Medium hits 90% accuracy on AIME-24 (high-school math competition) with majority voting. Frontier-level performance.
Slow and expensive. Don’t use for simple Q&A – you’re paying for reasoning tokens you don’t need.
Coding: Devstral 2
| Model | Parameters | Context | Use Case | Price (Dec 2025) |
|---|---|---|---|---|
| Devstral 2 | 123B | 256K | Code agents, multi-file tasks | Free via API |
| Devstral Small 2 | 24B | 256K | Local coding, IDE integration | Apache 2.0 (self-host) |
| Codestral 2 | Undisclosed | 256K | Fill-in-the-middle completion | Check pricing page |
Devstral 2 scores 72.2% on SWE-bench Verified – a benchmark of real GitHub issues. Built for agentic workflows: exploring codebases, editing multiple files, running tests. Like Cursor or Cline, but you control the model.
Devstral Small 2 runs locally. 24 billion parameters. Fast enough for real-time autocomplete – supports image inputs (multimodal), so it can read screenshots of UI bugs.
Mistral also released Mistral Vibe CLI (December 2025): terminal-based coding assistant powered by Devstral. Natural language commands, file manipulation, Git integration, command execution. Open-source (Apache 2.0). Tired of copying errors into ChatGPT? Try Vibe.
Edge Devices: Ministral 3
Released December 2025. Three sizes: 3B, 8B, 14B parameters. Three variants: Base (pre-trained), Instruct (chat-optimized), Reasoning (logic tasks).
Run on phones, drones, robots – anything without reliable internet. The 8B reasoning variant scores 85% on AIME ’25. Impressive for a model fitting in 16GB of RAM.
Apache 2.0 license. Download weights from Hugging Face, fine-tune on your own data, deploy anywhere.
Ministral models are dense (not mixture-of-experts) – inference is predictable. You don’t get token-to-token latency spikes you see with Large 3. Building a real-time assistant? Smaller + dense often beats larger + sparse.
Audio: Voxtral
Voxtral Realtime: sub-200ms latency, 13 languages, $0.006 per minute. Live call transcription, subtitles, voice agents.
Voxtral Mini Transcribe V2: batch transcription, $0.003 per minute. Handles 3-hour recordings in a single request – supports speaker diarization (who said what).
Both support context biasing: give it a list of 100 words/phrases (names, technical terms), and it prioritizes correct spelling. Great for medical or legal transcription.
Using the API
Create account at console.mistral.ai. Set up payment info. Generate API key. Keep it private.
Install SDK:
pip install mistralai
Basic chat completion:
from mistralai import Mistral
import os
client = Mistral(api_key=os.environ["MISTRAL_API_KEY"])
response = client.chat.complete(
model="mistral-small-latest",
messages=[
{"role": "system", "content": "You are a helpful assistant."},
{"role": "user", "content": "Explain gradient descent in one sentence."}
]
)
print(response.choices[0].message.content)
Streaming response (lower perceived latency):
stream = client.chat.stream(
model="mistral-medium-latest",
messages=[{"role": "user", "content": "Write a Python function to reverse a string."}]
)
for chunk in stream:
print(chunk.data.choices[0].delta.content, end="")
SDK: simple. Docs: solid. Watch for gotchas.
Rate Limits You’ll Hit
Rate limits reset by cumulative billing, not calendar time. You upgrade from Free to Pay-As-You-Go. Limits don’t instantly increase – they bump up after you cross billing thresholds ($5 spent, $50 spent, etc.). Community reports show people stuck at Free tier limits even after adding a card.
429 errors (rate limit exceeded) don’t always include a Retry-After header. Some client libraries don’t auto-retry. You get a raw exception and your job crashes. Build your own retry logic with exponential backoff.
Concurrent request limits are separate from rate limits. You might be under tokens-per-minute cap but hit the concurrency wall – too many simultaneous connections. Error message doesn’t always make this clear. Solution: queue requests or batch them.
Context window + output token math is hard-enforced. Hit the limit? Request fails. No warning. Check token counts before sending large prompts. Mistral uses custom tokenizer (tekken v3) – slightly different token counts than OpenAI’s tiktoken.
Some models (Codestral) occasionally enter “infinite thinking loop” and time out. Community-reported bug. No official workaround yet. Happens? Adjust prompt or switch models.
When to Pick Mistral
Want cheaper inference than OpenAI? Okay doing integration work yourself?
Need to run models on your own infrastructure? Compliance, data residency, or paranoia.
Building something where open-weight matters – fine-tuning on proprietary data, academic research, or you just don’t trust closed APIs.
Need multilingual support out of the box. Mistral models handle 40+ languages without special prompts.
Skip Mistral if you need plug-and-play with zero dev overhead, or you want the absolute best reasoning model regardless of cost (then use OpenAI o1 or Anthropic Claude).
Next Step
Pick one model. Run it. Measure cost and latency for YOUR task – not a benchmark. Best model? The one that hits your accuracy target at lowest cost per request.
Start with Mistral Medium 3 for general tasks. Switch to Devstral if you’re coding. Try Ministral if you need offline or low-latency. Upgrade to Large 3 only if Medium fails your accuracy bar.
Track spend daily. Mistral’s dashboard shows token usage and costs. Set a budget alert. You don’t want a surprise $500 bill because you forgot to cap max_tokens.
FAQ
How do Mistral’s rate limits actually work?
Workspace-level, tier-based. Free tier has conservative limits. Limits increase when cumulative billed amount crosses thresholds – not time-based. Upgrading doesn’t instantly raise limits; you have to spend money first. Check current limits at admin.mistral.ai/plateforme/limits.
Can I run Mistral models locally for free?
Yes. Pick an open-weight model: Ministral 3 (3B, 8B, 14B) or Devstral Small 2 (24B). Apache 2.0 licensed. Download weights from Hugging Face, run with vLLM or Ollama. No API fees – just compute costs.
What’s the difference between Mistral Medium 3 and Mistral Large 3?
Medium 3: $0.40 input, 128K context, general-purpose, multimodal. Large 3: $2.00 input, 256K context, mixture-of-experts (41B active / 675B total params), stronger reasoning. Use Medium unless you need extra context or task genuinely requires frontier reasoning. Most workloads don’t. Large 3 is 5x more expensive for input tokens – test Medium first.