Live pricing snapshot
Input / 1K
$0.0012
Prompt tokens
Output / 1K
$0.0012
Completion tokens
Input / 1M
$1.2000
Large-volume planning
Catalog models
144
Current pricing catalog size
llama 3.1 70b cost
Llama 3.1 70B request volume can scale faster than expected. Use this calculator page to estimate spend before overruns.
Model routing choices are easier when Llama 3.1 70B cost is visible at request level.
Input / 1K
$0.0012
Prompt tokens
Output / 1K
$0.0012
Completion tokens
Input / 1M
$1.2000
Large-volume planning
Catalog models
144
Current pricing catalog size
Real UI snapshot from AI Cost Board used in production workflows.

Provider-level drilldown for spend and token economics.
Estimated mode. Input capped at 100,000 chars.
Pricing updated: Mar 5, 2026, 04:00 AM
Input Cost
$0.0000
Output Cost
$0.000077
Total Cost
$0.000077
Price basis: 120 cents / 1M input tokens and 120 cents / 1M output tokens.
Use this free tool without login.
If you want ongoing tracking by project/provider, continue in the dashboard.
Normalize with the same prompt/output profile for every model. This page uses live input/output rates for Llama and converts them into per-1K and per-1M views.
Yes. Use the embedded calculator or pricing table first, then multiply by expected request volume and retry behavior.
Track retries, latency, and error rate. Production spend is pricing multiplied by operational behavior.
Use AI Cost Board to monitor cost per team, project, model, and provider with budget alerts and anomaly detection.
Move from one-off estimates to project-level cost, token, latency, and error tracking with alerts.
Start free tracking