Model reference · Synced 2025-04-29
Llama 3.1 70B
Llama 3.1 70B is an AI model from Nvidia. 128K context window. Capabilities: reasoning, tool calling, open weights. Available on 7 providers. Cheapest listing: $0 input / $0 output per 1M tokens.
Quick facts
- Cheapest input: $0 per 1M tokens (Nvidia)
- Cheapest output: $0 per 1M tokens
- Context window: 128K tokens
- Max output: 4K tokens
- Release date: 2024-07-16
- Capabilities: reasoning, tool calling, open weights
- Provider count: 7
Provider pricing
Same model, different providers, different prices. Cheapest first.
| Provider | Input / 1M | Output / 1M | Context | Listed |
|---|---|---|---|---|
| Nvidia | $0 | $0 | 128K | 2024-07-16 |
| Deep Infra | $0.4 | $0.4 | 131K | 2024-07-23 |
| Vercel AI Gateway | $0.4 | $0.4 | 131K | 2024-07-23 |
| Amazon Bedrock | $0.72 | $0.72 | 128K | 2024-07-23 |
| LLM Gateway | $0.72 | $0.72 | 128K | 2024-07-23 |
| Weights & Biases | $0.8 | $0.8 | 128K | 2024-07-23 |
| Synthetic | $0.9 | $0.9 | 128K | 2024-07-23 |
Prices synced daily from models.dev + provider docs.
How to use this model
If you're picking Llama 3.1 70B for a project, the three things that matter most:
- Compare it side-by-side with one or two alternatives in the live comparison tool. Pricing differences matter more than benchmarks at scale.
- Pick the cheapest provider that meets your latency / SLA need. Big spread across providers for the same weights.
- Re-evaluate every 3 months. Frontier prices drop fast; a model that's cheapest today may not be in a quarter.
Related models
FAQ
How much does Llama 3.1 70B cost? $0 input / $0 output per 1M tokens at the cheapest listing. See the table above for other providers.
What is the context window? 128K tokens.
Which providers offer it? Weights & Biases, Deep Infra, Amazon Bedrock, Synthetic, Nvidia, Vercel AI Gateway, LLM Gateway.
Where do these numbers come from? models.dev + provider documentation, synced daily. About the data.