Price Comparison
LLM Price Per Token Comparison (2026)
Compare AI model pricing per token. 50+ models from OpenAI, Anthropic, Google, Meta, Mistral, and DeepSeek side by side. Click any column header to sort.
53 models
| Model | Provider | Input / 1M Tokens ↑ | Output / 1M Tokens | Context Window | Best For |
|---|---|---|---|---|---|
| Mistral 7B | Mistral | $0.03 | $0.03 | 32K | Edge deployment |
| Gemini 1.5 Flash-8B | $0.04 | $0.15 | 1M | Ultra budget | |
| Llama 3.1 8B | Meta | $0.05 | $0.05 | 128K | Lightweight open |
| Llama 3 8B | Meta | $0.05 | $0.05 | 8K | Legacy small |
| Gemma 2 9B | $0.05 | $0.05 | 8K | Small open Google | |
| Phi-4 | Meta | $0.07 | $0.07 | 16K | Small model tasks |
| Gemini 1.5 Flash | $0.07 | $0.30 | 1M | Budget long context | |
| Gemini 2.0 Flash | $0.10 | $0.40 | 1M | Budget tasks | |
| Mistral Small | Mistral | $0.10 | $0.30 | 128K | Efficient tasks |
| Cohere Embed v3 | Cohere | $0.10 | $0.00 | 512 | Embeddings |
| DeepSeek V2.5 | DeepSeek | $0.14 | $0.28 | 128K | Budget general |
| DeepSeek Coder V2 | DeepSeek | $0.14 | $0.28 | 128K | Budget code |
| Gemma 2 27B | $0.14 | $0.14 | 8K | Open weight Google | |
| GPT-4o-mini | OpenAI | $0.15 | $0.60 | 128K | High volume |
| Gemini 2.5 Flash | $0.15 | $0.60 | 1M | Speed + value | |
| Mistral Nemo | Mistral | $0.15 | $0.15 | 128K | Open weight |
| Cohere Command R | Cohere | $0.15 | $0.60 | 128K | Efficient RAG |
| Qwen 2.5 Coder 32B | Meta | $0.16 | $0.16 | 128K | Code open source |
| Llama 4 Scout | Meta | $0.17 | $0.30 | 512K | Long context open |
| Llama 3.3 70B | Meta | $0.18 | $0.18 | 128K | Open source |
| Llama 3.1 70B | Meta | $0.18 | $0.18 | 128K | Balanced open |
| Llama 4 Maverick | Meta | $0.19 | $0.49 | 256K | Quality open source |
| Qwen QwQ 32B | Meta | $0.20 | $0.20 | 128K | Reasoning open |
| Llama 3 70B | Meta | $0.23 | $0.23 | 8K | Legacy open |
| Mixtral 8x7B | Mistral | $0.24 | $0.24 | 32K | MoE efficiency |
| Claude 3 Haiku | Anthropic | $0.25 | $1.25 | 200K | Legacy fast |
| DeepSeek V3 | DeepSeek | $0.27 | $1.10 | 64K | Cost-effective coding |
| Qwen 2.5 72B | Meta | $0.29 | $0.29 | 128K | Multilingual open |
| Mistral Codestral | Mistral | $0.30 | $0.90 | 256K | Code generation |
| Cohere Command Light | Cohere | $0.30 | $0.60 | 4K | Budget Cohere |
| GPT-3.5 Turbo | OpenAI | $0.50 | $1.50 | 16K | Simple tasks |
| DeepSeek R1 | DeepSeek | $0.55 | $2.19 | 64K | Reasoning |
| Mixtral 8x22B | Mistral | $0.65 | $0.65 | 64K | Large MoE |
| Claude Haiku 3.5 | Anthropic | $0.80 | $4.00 | 200K | Fast responses |
| Cohere Command | Cohere | $1.00 | $2.00 | 4K | Legacy Cohere |
| o3-mini | OpenAI | $1.10 | $4.40 | 200K | Reasoning at scale |
| o4-mini | OpenAI | $1.10 | $4.40 | 200K | Efficient reasoning |
| Gemini 2.5 Pro | $1.25 | $10.00 | 1M | Long context | |
| Gemini 1.5 Pro | $1.25 | $5.00 | 2M | Max context | |
| Mistral Large | Mistral | $2.00 | $6.00 | 128K | Enterprise |
| GPT-4o | OpenAI | $2.50 | $10.00 | 128K | General purpose |
| Cohere Command R+ | Cohere | $2.50 | $10.00 | 128K | RAG applications |
| Mistral Medium | Mistral | $2.70 | $8.10 | 32K | Legacy enterprise |
| Claude Sonnet 4 | Anthropic | $3.00 | $15.00 | 200K | Balanced performance |
| Claude 3 Sonnet | Anthropic | $3.00 | $15.00 | 200K | Legacy balanced |
| Llama 3.1 405B | Meta | $3.00 | $3.00 | 128K | Max open source |
| GPT-5 | OpenAI | $5.00 | $15.00 | 200K | Complex reasoning |
| o3 | OpenAI | $10.00 | $40.00 | 200K | Advanced reasoning |
| GPT-4 Turbo | OpenAI | $10.00 | $30.00 | 128K | Legacy apps |
| Claude Opus 4 | Anthropic | $15.00 | $75.00 | 200K | Complex analysis |
| Claude 3 Opus | Anthropic | $15.00 | $75.00 | 200K | Legacy complex |
| GPT-4 | OpenAI | $30.00 | $60.00 | 8K | Legacy compatibility |
| GPT-4 32K | OpenAI | $60.00 | $120.00 | 32K | Long legacy prompts |
Track your actual spend across all these models
CostLayer monitors your real API usage across every provider and model, so you always know exactly what you are spending.
Get Started