Live Model List: For the complete, up-to-date list of models with current pricing, call the List Models endpoint or visit the Console.
Featured Models
Anthropic Claude
| Model | Context | Max Output | Input ($/M) | Output ($/M) |
|---|---|---|---|---|
anthropic/claude-opus-4.5 | 200K | 32K | $15.00 | $75.00 |
anthropic/claude-sonnet-4.5 | 200K | 16K | $3.00 | $15.00 |
anthropic/claude-haiku-3.5 | 200K | 8K | $0.80 | $4.00 |
OpenAI GPT
| Model | Context | Max Output | Input ($/M) | Output ($/M) |
|---|---|---|---|---|
openai/gpt-4o | 128K | 16K | $2.50 | $10.00 |
openai/gpt-4o-mini | 128K | 16K | $0.15 | $0.60 |
openai/o1 | 200K | 100K | $15.00 | $60.00 |
openai/o1-mini | 128K | 65K | $3.00 | $12.00 |
Google Gemini
| Model | Context | Max Output | Input ($/M) | Output ($/M) |
|---|---|---|---|---|
google/gemini-2.0-flash | 1M | 8K | $0.10 | $0.40 |
google/gemini-1.5-pro | 2M | 8K | $1.25 | $5.00 |
google/gemini-1.5-flash | 1M | 8K | $0.075 | $0.30 |
xAI Grok
| Model | Context | Max Output | Input ($/M) | Output ($/M) |
|---|---|---|---|---|
xai/grok-2 | 128K | 8K | $2.00 | $10.00 |
xai/grok-2-vision | 32K | 8K | $2.00 | $10.00 |
DeepSeek
| Model | Context | Max Output | Input ($/M) | Output ($/M) |
|---|---|---|---|---|
deepseek/deepseek-chat | 64K | 8K | $0.14 | $0.28 |
deepseek/deepseek-reasoner | 64K | 8K | $0.55 | $2.19 |
Meta Llama
| Model | Context | Max Output | Input ($/M) | Output ($/M) |
|---|---|---|---|---|
meta/llama-3.3-70b | 128K | 8K | $0.60 | $0.60 |
meta/llama-3.1-405b | 128K | 8K | $3.00 | $3.00 |
Mistral
| Model | Context | Max Output | Input ($/M) | Output ($/M) |
|---|---|---|---|---|
mistral/mistral-large | 128K | 8K | $2.00 | $6.00 |
mistral/mistral-small | 32K | 8K | $0.20 | $0.60 |
mistral/codestral | 32K | 8K | $0.30 | $0.90 |
All Providers
Case.dev provides access to models from:Anthropic
Claude family
OpenAI
GPT & o1 family
Gemini family
xAI
Grok family
Meta
Llama family
Mistral
Mistral family
DeepSeek
DeepSeek family
Cohere
Command family
Perplexity
Sonar family
Groq
Fast inference
Together
Open models
Fireworks
Fast inference
Pricing
All prices shown are per million tokens ($/M). Cache pricing is available for supported models.- Input: Cost per million input tokens
- Output: Cost per million output tokens
- Cache Read: Cost to read from prompt cache (typically 90% discount)
- Cache Write: Cost to write to prompt cache