UseRightAI
UseRightAI logo
HomeModelsComparePricingWhat's New
UseRightAI
Cut through AI hype. Pick what works.
UseRightAI logo
Cut through AI hype. Pick what works.

Independent AI model tracker. Live pricing, real benchmarks, zero vendor bias.

X (Twitter)LinkedInUpdatesContact

Compare

ChatGPT vs ClaudeGPT-4o vs Claude SonnetClaude vs GeminiDeepSeek vs ChatGPTMistral vs ClaudeGemini Flash vs GPT-4o MiniLlama vs ChatGPTBuild your own →

Best For

CodingWritingDevelopersProduct ManagersDesignersSalesBest Cheap AIBest Free AI

Pricing & Data

API Token PricingPrice HistoryBenchmark ScoresPrivacy & SafetySubscription PlansCost CalculatorWhich AI is Cheapest?

Company

About UseRightAIContactWhat ChangedAll ModelsDisclosuresPrivacy PolicyTerms of Service

© 2026 UseRightAI. Independent · Free forever · Not affiliated with any AI provider.

Affiliate links are clearly labeled. See disclosures.

Paris, France · Founded 2023

Mistral AI

European AI. Apache licensed. Runs anywhere.

Mistral AI is a Paris-based lab building efficient, Apache 2.0-licensed open-weight models. Mistral Large 2 is preferred for compliance-sensitive European deployments. Codestral 25.01 is the strongest open-weight model for code generation.

Rankings refresh dailyScored on 6 criteriaNo paid rankings
  • Apache 2.0 license — use commercially with no restrictions
  • Preferred by European enterprises for data residency compliance
  • Codestral 25.01 is the strongest open-weight code model in 2026
24 models

All Mistral AI Models

Every Mistral AI model in the directory, ranked by overall capability score.

MistralBudget

Mistral: Mistral Medium 3.1

Mistral Medium 3.1 is a multimodal mid-tier model from Mistral that supersedes Mistral Large 2, offering vision capabilities alongside strong text performance at a significantly reduced price point. It targets the sweet spot between budget models and expensive flagships, with a 128K context window and competitive multilingual support.

Verdict
The best Mistral model for budget-conscious builders who still need multimodal capability and solid multilingual output.
Quality score
70%
Pricing

Mistral AI API Pricing

Per 1 million tokens. Updated when providers change prices.

ModelInput / 1MOutput / 1MContextSpeed
Mistral: Mistral Medium 3.1
Budget
$0.40/1M$2.00/1M131KFast
Mistral: Pixtral Large 2411
Balanced
$2.00/1M$6.00/1M131KBalanced
Mistral: Mistral Small 4
Budget
$0.15/1M$0.60/1M262KFast
Mistral: Mistral Large 3 2512
Budget
$0.50/1M$1.50/1M262KBalanced
Mistral: Mistral Small 3.2 24B

Compare Mistral AI Models

Head-to-head comparisons for the most-searched questions.

Mistral: Mistral Medium 3.1 vs Mistral: Pixtral Large 2411Mistral: Mistral Medium 3.1 vs Mistral: Mistral Small 4Mistral: Pixtral Large 2411 vs Mistral: Mistral Small 4Mistral: Pixtral Large 2411 vs Mistral: Mistral Large 3 2512Mistral: Mistral Small 4 vs Mistral: Mistral Large 3 2512Mistral: Mistral Small 4 vs Mistral: Mistral Small 3.2 24BMistral: Mistral Large 3 2512 vs Mistral: Mistral Small 3.2 24BMistral: Mistral Large 3 2512 vs Mistral Small 3.1Open compare tool →

Newsletter

Get notified when Mistral AI releases new models

Pricing changes, new releases, and ranking shifts — straight to your inbox.

No spam. Useful updates only. Affiliate disclosures always clearly labeled.

Mistral AI FAQ

What is Mistral's best model in 2026?

Mistral Large 2 is Mistral's most capable general model. Codestral 25.01 is Mistral's strongest coding model — competitive with Claude Sonnet 4.6 on specific coding benchmarks. Mistral Small 3.1 is an excellent free option for lightweight tasks.

Why choose Mistral over OpenAI or Anthropic?

Mistral models are Apache 2.0 licensed — you can self-host, fine-tune, and deploy commercially with no usage restrictions. This matters for European data residency, compliance requirements, and teams that need full model ownership.

Are Mistral models open source?

Yes — Mistral Small 3.1, Mistral Large 2, and Codestral are available under Apache 2.0. You can download, modify, and deploy them commercially without needing to pay or seek permission.

Explore other providers

OpenAIAnthropicGooglexAIMetaDeepSeekBrowse all models →
$0.40/1M in
$2.00/1M out
Speed
Fast
Best for cost-sensitive teams needing solid coding, instruction-following, and basic vision tasks without paying flagship prices.
Context
131k tokens
Officially supersedes Mistral Large 2, representing a generational shift in Mistral's lineup toward multimodal capability at lower cost tiers. Available via Mistral API and select cloud providers. No function calling limitations noted at this tier.
BudgetMultimodalMultilingualMid-tierVision
Best for
Cost-sensitive teams needing solid coding, instruction-following, and basic vision tasks without paying flagship prices.
View model
MistralBalanced

Mistral: Pixtral Large 2411

Pixtral Large 2411 is Mistral's flagship multimodal model, adding native image understanding to the Mistral Large 2 foundation. It processes both text and images with strong reasoning across documents, charts, and visual content.

Verdict
A capable and fairly priced multimodal flagship, best suited for Mistral ecosystem users and European compliance requirements.
Quality score
74%
Pricing
$2.00/1M in
$6.00/1M out
Speed
Balanced
Best for teams needing a capable european-hosted multimodal model for document analysis, visual qa, and code generation with image context.
Context
131k tokens
Available via Mistral API (la Plateforme) and supports self-hosted deployment. The '2411' suffix indicates a November 2024 release. Supersedes Mistral Large 2 as the primary flagship. Image input pricing follows the same $2/1M token rate.
MultimodalEuropean AIVisionFlagshipDocument Analysis
Best for
Teams needing a capable European-hosted multimodal model for document analysis, visual QA, and code generation with image context.
View model
MistralBudget

Mistral: Mistral Small 4

Mistral Small 4 is a compact, cost-efficient language model from Mistral AI that punches well above its price class, succeeding Mistral Large 2 in capability while costing a fraction of the price. It features a 256K context window and is optimized for high-throughput, latency-sensitive applications.

Verdict
The best bang-for-buck text model in its class — Mistral Large 2 quality at a fraction of the cost.
Quality score
68%
Pricing
$0.15/1M in
$0.60/1M out
Speed
Fast
Best for teams needing reliable, fast text generation and coding assistance at near-commodity pricing without sacrificing too much quality.
Context
262k tokens
Pricing at $0.15/$0.60 per million tokens makes this one of the most affordable capable models on the market. Available via Mistral's La Plateforme API and compatible with OpenAI-style endpoints. No image input support confirmed at launch.
BudgetFastLong ContextMultilingualCoding
Best for
Teams needing reliable, fast text generation and coding assistance at near-commodity pricing without sacrificing too much quality.
View model
MistralBudget

Mistral: Mistral Large 3 2512

Mistral Large 3 2512 is Mistral's flagship dense model updated in December 2025, offering strong multilingual reasoning and coding capabilities at a significantly reduced price point compared to its predecessor. It targets enterprise workloads that need high-quality outputs without paying top-tier frontier model prices.

Verdict
The best price-per-quality ratio in the non-mini flagship tier, especially for multilingual and long-context enterprise tasks.
Quality score
69%
Pricing
$0.50/1M in
$1.50/1M out
Speed
Balanced
Best for multilingual enterprise tasks, code generation, and long-document analysis where cost efficiency matters more than absolute state-of-the-art performance.
Context
262k tokens
Pricing of $0.50 input / $1.50 output per 1M tokens places it firmly in the budget-flagship category. Available via Mistral API (La Plateforme) and major cloud providers. December 2025 update ('2512') improves instruction following over the earlier 2407 release.
Budget flagshipMultilingualLong contextEnterpriseCode
Best for
Multilingual enterprise tasks, code generation, and long-document analysis where cost efficiency matters more than absolute state-of-the-art performance.
View model
MistralBudget

Mistral: Mistral Small 3.2 24B

Mistral Small 3.2 24B is a compact 24-billion parameter model from Mistral that punches well above its weight class, superseding Mistral Large 2 at a fraction of the cost. It handles coding, instruction-following, and multilingual tasks with strong efficiency for its size.

Verdict
The best budget coding model available today, offering frontier-adjacent performance at commodity pricing.
Quality score
68%
Pricing
$0.07/1M in
$0.20/1M out
Speed
Fast
Best for high-volume production workloads where cost matters but quality can't be sacrificed entirely — especially code generation and structured output tasks.
Context
128k tokens
Mistral Small 3.2 is available as an open-weight model, making it deployable on-premises or via self-hosted infrastructure — a key differentiator over GPT-4o Mini and Claude Haiku for privacy-sensitive use cases.
BudgetCodingEfficientOpen-weightMultilingual
Best for
High-volume production workloads where cost matters but quality can't be sacrificed entirely — especially code generation and structured output tasks.
View model
MistralBudget

Mistral Small 3.1

Mistral's ultra-budget multimodal model — exceptionally cheap with vision support, built for high-volume lightweight tasks where cost is the primary constraint.

Verdict
Ultra-cheap multimodal model for massive-volume, low-complexity pipelines.
Quality score
57%
Pricing
$0.03/1M in
$0.11/1M out
Speed
Very fast
Best for ultra-high-volume classification, summarisation, and lightweight vision tasks
Context
128k tokens
At $0.10/1M input, the cost question disappears. The only question is whether the task complexity exceeds what Mistral Small can handle.
BudgetMultimodalUltra cheapMistral
Best for
Ultra-high-volume classification, summarisation, and lightweight vision tasks
View model
MistralBudget

Mistral: Mistral Medium 3

Mistral Medium 3 is a mid-tier model from Mistral AI that punches above its weight class, officially superseding Mistral Large 2 while costing a fraction of the price. It targets teams needing capable multilingual and coding performance without flagship-level spend.

Verdict
The most capable budget model Mistral has shipped — a smart default for high-volume teams that need real performance without flagship pricing.
Quality score
67%
Pricing
$0.40/1M in
$2.00/1M out
Speed
Fast
Best for cost-conscious teams running high-volume coding, summarization, or multilingual tasks at enterprise scale.
Context
131k tokens
Priced at $0.40 input / $2.00 output per 1M tokens. Officially supersedes Mistral Large 2, making it an easy drop-in upgrade for existing Mistral users. Available via Mistral's API and La Plateforme.
BudgetMultilingualCodingHigh VolumeMid-Tier
Best for
Cost-conscious teams running high-volume coding, summarization, or multilingual tasks at enterprise scale.
View model
MistralBalanced

Mistral Large 2

Balanced enterprise model with consistent reasoning, good speed, and a dependable middle-ground — especially for European teams with data residency requirements.

Verdict
Best balanced generalist for EU teams with data residency needs.
Quality score
67%
Pricing
$2.00/1M in
$6.00/1M out
Speed
Balanced
Best for balanced team usage with eu data residency requirements
Context
128k tokens
The EU hosting angle is the real differentiator here — for teams outside Europe, other models perform better.
EU hostingBalancedTeam default
Best for
Balanced team usage with EU data residency requirements
View model
MistralBudget

Mistral: Devstral Medium

Devstral Medium is Mistral's code-focused model optimized for software development tasks, offering strong code generation and debugging capabilities at a budget-friendly price point. It targets developers who need reliable coding assistance without paying flagship model rates.

Verdict
A genuinely specialized, budget-friendly coding model that earns its place in any developer's API toolkit.
Quality score
60%
Pricing
$0.40/1M in
$2.00/1M out
Speed
Balanced
Best for developers seeking capable code generation, debugging, and code review at a fraction of the cost of gpt-4-class models.
Context
131k tokens
Pricing is notably aggressive at ~$0.40 input / $2.00 output per 1M tokens. Available via Mistral's La Plateforme API. Part of the Devstral family, which is distinct from Mistral's general-purpose Mistral Medium line.
Code-focusedBudgetDeveloper toolsMistral128K context
Best for
Developers seeking capable code generation, debugging, and code review at a fraction of the cost of GPT-4-class models.
View model
MistralBudget

Codestral 25.01

Coding-specialist model designed for fast engineering assistance at a budget-conscious price point.

Verdict
Best budget-focused coding specialist for high-volume developer teams.
Quality score
57%
Pricing
$0.90/1M in
$2.70/1M out
Speed
Very fast
Best for affordable high-volume coding support
Context
256k tokens
Ideal for teams running thousands of daily coding prompts where premium model costs add up quickly.
Coding specialistBudgetFast
Best for
Affordable high-volume coding support
View model
MistralBudget

Mistral: Mistral Nemo

Mistral Nemo is a compact 12B-parameter open-weight model developed in collaboration with NVIDIA, designed to deliver strong multilingual and instruction-following performance at an extremely low cost. It fits into a 128K context window and is optimized for deployment efficiency without sacrificing too much reasoning depth.

Verdict
A dirt-cheap multilingual model perfect for bulk text tasks, but don't expect frontier-level reasoning.
Quality score
55%
Pricing
$0.02/1M in
$0.03/1M out
Speed
Fast
Best for teams needing a cheap, fast, multilingual workhorse for classification, summarization, or light coding tasks at scale.
Context
131k tokens
Mistral Nemo is open-weight (Apache 2.0 license), so self-hosting is an option for teams that want to eliminate API costs entirely. Pricing via API is through Mistral's La Plateforme. The model uses a Tekken tokenizer which is more efficient than older Mistral tokenizers, especially for non-English text.
budgetmultilingualopen-weight12Befficient
Best for
Teams needing a cheap, fast, multilingual workhorse for classification, summarization, or light coding tasks at scale.
View model
MistralBudget

Mistral: Devstral 2 2512

Devstral 2 2512 is Mistral's second-generation code-specialized model, built specifically for software development tasks with a 256K context window. It targets developers needing a cost-efficient coding assistant without sacrificing meaningful capability.

Verdict
A purpose-built coding workhorse that punches well above its price tag for development teams running high-volume or agentic pipelines.
Quality score
55%
Pricing
$0.40/1M in
$2.00/1M out
Speed
Fast
Best for budget-conscious developers who need a capable coding model for agentic workflows, code generation, and repository-scale context at a fraction of flagship pricing.
Context
262k tokens
The December 2025 (2512) release date suggests this is a recent iteration. Pricing at $0.40 input / $2.00 output is notably competitive for a code-specialist model with 256K context. Verify availability and rate limits via Mistral API or partner providers.
Code-specialistBudgetLong contextAgenticMistral
Best for
Budget-conscious developers who need a capable coding model for agentic workflows, code generation, and repository-scale context at a fraction of flagship pricing.
View model
MistralBudget

Mistral: Devstral Small 1.1

Devstral Small 1.1 is Mistral's code-specialized small model, purpose-built for software engineering tasks including code generation, debugging, and repository-level reasoning. It succeeds Devstral Small 1.0 with improved instruction following and agentic coding capabilities at a fraction of flagship model costs.

Verdict
The best dollar-for-dollar coding model for agentic pipelines that doesn't need to do anything else.
Quality score
54%
Pricing
$0.10/1M in
$0.30/1M out
Speed
Fast
Best for developers who need a cheap, fast coding assistant for agentic workflows, code review, and multi-file repo tasks without paying flagship prices.
Context
131k tokens
Available via Mistral API and can be self-hosted via open weights. Pricing is among the lowest available for a code-specialized model. Designed to work within coding agent frameworks like SWE-agent and OpenHands.
code-specialistbudgetagenticopen-source-friendlySWE-bench
Best for
Developers who need a cheap, fast coding assistant for agentic workflows, code review, and multi-file repo tasks without paying flagship prices.
View model
MistralBudget

Mistral: Codestral 2508

Codestral 2508 is Mistral's latest dedicated code model, succeeding Codestral 25.01 with improved code generation, completion, and reasoning across 80+ programming languages. It offers a massive 256K context window at a budget-friendly price point aimed squarely at developer tooling and IDE integrations.

Verdict
The most cost-effective specialized code model for production developer tooling with serious context capacity.
Quality score
53%
Pricing
$0.30/1M in
$0.90/1M out
Speed
Fast
Best for high-volume code generation, completion, and refactoring tasks where cost efficiency and long-context handling matter most.
Context
256k tokens
Available via Mistral's La Plateforme API. Also accessible through Continue.dev, Cursor, and other IDE integrations that support the Codestral endpoint. FIM (fill-in-the-middle) mode is specifically supported for autocomplete use cases. Output price rounds to ~$0.90/1M tokens.
Code SpecialistBudgetLong ContextFIMIDE Integration
Best for
High-volume code generation, completion, and refactoring tasks where cost efficiency and long-context handling matter most.
View model
MistralBudget

Mistral: Ministral 3 8B 2512

Ministral 3B is Mistral's ultra-compact edge model designed for low-latency, cost-sensitive deployments. It punches above its weight for a sub-4B parameter model, handling instruction following, summarization, and lightweight reasoning at near-negligible cost.

Verdict
The go-to model for bulk processing tasks where cost and speed trump quality.
Quality score
50%
Pricing
$0.15/1M in
$0.15/1M out
Speed
Very fast
Best for high-volume, latency-sensitive applications where cost per token matters more than top-tier quality.
Context
262k tokens
The '8B 2512' in the model name likely refers to a specific versioned release; despite the naming, this is based on Mistral's 3B architecture. Confirm parameter count and capabilities with Mistral's official documentation before production use.
budgetedgefastlong-contextcompact
Best for
High-volume, latency-sensitive applications where cost per token matters more than top-tier quality.
View model
MistralBudget

Mistral: Mistral Small 3

Mistral Small 3 is a compact, budget-oriented language model from Mistral AI that punches above its weight class for everyday NLP tasks. It supersedes Mistral Large 2 in efficiency while targeting cost-sensitive deployments that don't require frontier-level reasoning.

Verdict
A lean, fast, affordable workhorse for text tasks — ideal for scale, not for depth.
Quality score
55%
Pricing
$0.05/1M in
$0.08/1M out
Speed
Very fast
Best for high-volume, cost-sensitive applications like customer support automation, content drafting, and lightweight code assistance.
Context
33k tokens
Pricing is exceptionally competitive at $0.05/$0.08 per 1M tokens. Available via Mistral's La Plateforme API and various third-party providers. GDPR-friendly EU-based hosting is a notable advantage for European enterprise customers. No image input or output support.
BudgetFastMultilingualLightweightHigh-volume
Best for
High-volume, cost-sensitive applications like customer support automation, content drafting, and lightweight code assistance.
View model
MistralBudget

Mistral: Ministral 3 14B 2512

Ministral 3B is Mistral's compact edge-optimized model designed for high-throughput, low-latency tasks at an extremely competitive price point. Despite its small size, it supports a 262K context window, making it unusually capable for a sub-$0.20/1M token model.

Verdict
An ultra-cheap, fast model with a surprisingly large context window, but quality limitations make it a pipeline tool rather than a general assistant.
Quality score
48%
Pricing
$0.20/1M in
$0.20/1M out
Speed
Very fast
Best for high-volume, cost-sensitive workflows like document triage, classification, summarization, and lightweight coding assistance where budget is the primary constraint.
Context
262k tokens
Model name suggests a December 2025 revision ('2512'). Pricing is symmetric at $0.20/1M for both input and output, which simplifies cost modeling. Confirm availability on your target API platform as Mistral model availability varies by provider.
budgetedgesmall modellong contexthigh throughput
Best for
High-volume, cost-sensitive workflows like document triage, classification, summarization, and lightweight coding assistance where budget is the primary constraint.
View model
MistralBalanced

Mistral: Mixtral 8x22B Instruct

Mixtral 8x22B Instruct is Mistral's flagship sparse mixture-of-experts model, routing tokens through 2 of 8 expert networks (39B active parameters out of 141B total) for efficient high-quality inference. It excels at multilingual tasks, code generation, and instruction-following with strong European language support.

Verdict
A capable MoE workhorse with strong multilingual chops, but its short context window and rising competition have eroded its value proposition.
Quality score
59%
Pricing
$2.00/1M in
$6.00/1M out
Speed
Balanced
Best for teams needing strong multilingual capabilities and solid coding performance at a mid-tier price point without relying on openai or anthropic infrastructure.
Context
66k tokens
Available via Mistral API and as open weights (Apache 2.0 license) for self-hosting. The open-weight option is a key differentiator for privacy-sensitive or on-premise deployments. API pricing at $2/$6 per million tokens is mid-range but faces pressure from newer, cheaper alternatives.
MoEMultilingualOpen-weightMid-tierInstruct
Best for
Teams needing strong multilingual capabilities and solid coding performance at a mid-tier price point without relying on OpenAI or Anthropic infrastructure.
View model
MistralBudget

Mistral: Voxtral Small 24B 2507

Voxtral Small 24B is Mistral's audio-capable language model, designed for speech transcription, voice understanding, and spoken language tasks at a budget-friendly price point. It supersedes Mistral Small 3.1 with native audio input support built on a 24B parameter base.

Verdict
A purpose-built budget audio model that excels at voice tasks but stumbles on context length and general-purpose depth.
Quality score
47%
Pricing
$0.10/1M in
$0.30/1M out
Speed
Fast
Best for transcribing, analyzing, and responding to audio input cost-effectively without needing a separate speech-to-text pipeline.
Context
32k tokens
Voxtral Small is audio-in capable but does not support image input. The 32K context window is notably short for a 2025 model. Pricing is via Mistral's API; availability through third-party providers may vary. Check whether your use case requires audio input — the text-only version of Mistral Small 3.1 may be more appropriate for pure text workloads.
Audio AIBudgetMultilingualSpeechMistral
Best for
Transcribing, analyzing, and responding to audio input cost-effectively without needing a separate speech-to-text pipeline.
View model
MistralBalanced

Mistral: Mixtral 8x7B Instruct

Mixtral 8x7B Instruct is Mistral's sparse mixture-of-experts model that routes tokens through 2 of 8 expert networks, achieving strong performance while activating only ~13B parameters per forward pass. It excels at instruction-following, multilingual tasks, and code generation at a competitive price point.

Verdict
A historically significant open-weight model that's been surpassed by newer alternatives but still earns its place in self-hosted and multilingual pipelines.
Quality score
53%
Pricing
$0.54/1M in
$0.54/1M out
Speed
Fast
Best for developers and teams needing a capable open-weight model for coding, multilingual tasks, and general instruction-following without flagship model pricing.
Context
33k tokens
Pricing is symmetric at $0.54/1M for both input and output. As an open-weight model, costs can drop significantly if self-hosted. The 32K context window is a hard ceiling — plan accordingly for document-heavy workflows.
Open-weightMoEMultilingualCost-effectiveSelf-hostable
Best for
Developers and teams needing a capable open-weight model for coding, multilingual tasks, and general instruction-following without flagship model pricing.
View model
MistralBudget

Mistral: Ministral 3 3B 2512

Ministral 3B is Mistral's ultra-compact 3-billion parameter edge model designed for lightweight inference, on-device deployment, and cost-sensitive applications. It delivers surprisingly capable text understanding and generation at a fraction of the cost of larger models.

Verdict
The cheapest viable option for simple NLP tasks, but don't expect small-flagship performance.
Quality score
41%
Pricing
$0.10/1M in
$0.10/1M out
Speed
Very fast
Best for high-volume, low-latency tasks where cost and speed matter more than frontier-level reasoning.
Context
131k tokens
Priced at a flat $0.10/1M for both input and output, making cost estimation predictable. The '2512' suffix indicates a December 2025 release version. Best suited for batch processing, classification, or extraction pipelines where volume is high and task complexity is low.
3BEdgeUltra-budgetMistralLightweight
Best for
High-volume, low-latency tasks where cost and speed matter more than frontier-level reasoning.
View model
MistralBudget

Mistral: Saba

Mistral Saba is a compact, budget-oriented language model from Mistral designed for efficient text tasks with a focus on Arabic and South Asian languages alongside English. It targets cost-sensitive deployments where multilingual support is more important than raw reasoning depth.

Verdict
A bargain multilingual model built for Arabic and South Asian languages, but too constrained for demanding workloads.
Quality score
45%
Pricing
$0.20/1M in
$0.60/1M out
Speed
Fast
Best for low-cost multilingual applications requiring arabic, hindi, or urdu language support
Context
33k tokens
Pricing reflects Mistral API rates and may vary by reseller. The model's name 'Saba' references Arabic linguistic heritage, signaling its intended multilingual focus. No vision or tool-use capabilities documented at launch.
BudgetMultilingualArabicCompactEfficient
Best for
Low-cost multilingual applications requiring Arabic, Hindi, or Urdu language support
View model
MistralBudget

Mistral: Mistral Small Creative

Mistral Small Creative is a fine-tuned variant of Mistral Small optimized for creative writing tasks, offering a budget-friendly option for generative content at under $0.10/1M input tokens. It targets storytelling, copywriting, and imaginative text generation at a fraction of the cost of flagship models.

Verdict
A lean, cheap creative writing workhorse — ideal for volume content generation but not for quality-critical storytelling.
Quality score
36%
Pricing
$0.10/1M in
$0.30/1M out
Speed
Fast
Best for budget-conscious creative writing tasks like short stories, marketing copy, and brainstorming where cost matters more than peak quality.
Context
33k tokens
Context window of 32,768 tokens is notably smaller than competing budget models. Pricing is approximate ($0.10 input / $0.30 output per 1M tokens). Availability is through Mistral's API (La Plateforme) and may also be accessible via third-party providers. Confirm fine-tune scope before deploying for non-creative tasks.
Creative WritingBudgetFastShort-formMistral
Best for
Budget-conscious creative writing tasks like short stories, marketing copy, and brainstorming where cost matters more than peak quality.
View model
MistralBudget

Mistral: Mistral 7B Instruct v0.1

Mistral 7B Instruct v0.1 is a 7-billion-parameter instruction-tuned model from Mistral AI, one of the earliest open-weight models to challenge larger proprietary models on efficiency. It handles general text tasks at extremely low cost but is constrained by a very small context window of under 3K tokens.

Verdict
A historically significant but now outdated budget model crippled by an unusably small context window.
Quality score
26%
Pricing
$0.11/1M in
$0.19/1M out
Speed
Very fast
Best for ultra-low-cost simple text tasks like classification, short summarization, or lightweight chatbot responses where context length is not a concern.
Context
3k tokens
This is v0.1, the original release — not to be confused with v0.2 or v0.3 which substantially improve context length and quality. The listed context window of ~2,824 tokens is unusually small even among budget models. Marked as superseding Mistral Large 2 in the spec, which appears to be a data error — this model does not supersede Mistral Large 2 in capability or positioning.
budgetopen-weightsmall modellegacyfast
Best for
Ultra-low-cost simple text tasks like classification, short summarization, or lightweight chatbot responses where context length is not a concern.
View model
Budget
$0.07/1M
$0.20/1M
128K
Fast
Mistral Small 3.1
Budget
$0.03/1M$0.11/1M128KVery fast
Mistral: Mistral Medium 3
Budget
$0.40/1M$2.00/1M131KFast
Mistral Large 2
Balanced
$2.00/1M$6.00/1M128KBalanced
Mistral: Devstral Medium
Budget
$0.40/1M$2.00/1M131KBalanced
Codestral 25.01
Budget
$0.90/1M$2.70/1M256KVery fast
Mistral: Mistral Nemo
Budget
$0.02/1M$0.03/1M131KFast
Mistral: Devstral 2 2512
Budget
$0.40/1M$2.00/1M262KFast
Mistral: Devstral Small 1.1
Budget
$0.10/1M$0.30/1M131KFast
Mistral: Codestral 2508
Budget
$0.30/1M$0.90/1M256KFast
Mistral: Ministral 3 8B 2512
Budget
$0.15/1M$0.15/1M262KVery fast
Mistral: Mistral Small 3
Budget
$0.05/1M$0.08/1M33KVery fast
Mistral: Ministral 3 14B 2512
Budget
$0.20/1M$0.20/1M262KVery fast
Mistral: Mixtral 8x22B Instruct
Balanced
$2.00/1M$6.00/1M66KBalanced
Mistral: Voxtral Small 24B 2507
Budget
$0.10/1M$0.30/1M32KFast
Mistral: Mixtral 8x7B Instruct
Balanced
$0.54/1M$0.54/1M33KFast
Mistral: Ministral 3 3B 2512
Budget
$0.10/1M$0.10/1M131KVery fast
Mistral: Saba
Budget
$0.20/1M$0.60/1M33KFast
Mistral: Mistral Small Creative
Budget
$0.10/1M$0.30/1M33KFast
Mistral: Mistral 7B Instruct v0.1
Budget
$0.11/1M$0.19/1M3KVery fast
Compare all providers →