Model Provider Input $/1M Output $/1M Tier Best for
Mistral: Mistral Nemo Mistral $0.020 $0.030 Budget Teams needing a cheap, fast, multilingual workhorse for classification, summarization, or light coding tasks at scale.
Meta: Llama 3.1 8B Instruct Meta $0.020 $0.050 Budget High-throughput applications where cost and speed matter more than frontier-level quality, such as chatbots, content classification, and text summarization.
Meta: Llama 3 8B Instruct Meta $0.040 $0.040 Budget High-volume, cost-sensitive applications where speed and price matter more than peak accuracy.
Google: Gemma 2 9B Google $0.030 $0.090 Budget Lightweight text tasks, classification, and summarization where cost matters more than frontier-level quality.
Mistral: Ministral 3 3B 2512 Mistral $0.100 $0.100 Budget High-volume, low-latency tasks where cost and speed matter more than frontier-level reasoning.
Meta: Llama 3.2 1B Instruct Meta $0.027 $0.200 Budget Ultra-low-cost text classification, simple Q&A, and high-volume automation pipelines where cost per token is critical.
Mistral: Mistral Small 3.2 24B Mistral $0.075 $0.200 Budget High-volume production workloads where cost matters but quality can't be sacrificed entirely — especially code generation and structured output tasks.
Mistral: Ministral 3 8B 2512 Mistral $0.150 $0.150 Budget High-volume, latency-sensitive applications where cost per token matters more than top-tier quality.
Mistral: Mistral 7B Instruct v0.1 Mistral $0.110 $0.190 Budget Ultra-low-cost simple text tasks like classification, short summarization, or lightweight chatbot responses where context length is not a concern.
Meta: Llama Guard 4 12B Meta $0.180 $0.180 Budget Automated content safety screening and policy enforcement in LLM-powered applications