AI economics · Free calculator

AI Model Cost Calculator

Estimate monthly API costs for GPT-5, Claude Sonnet, Gemini Pro, and other LLMs. Model input/output tokens, request volume, and break-even pricing for your AI feature or product.

Scenarios
Common scenarios

Tap a persona to auto-load realistic numbers for that scenario, then tweak the sliders.

100,000
1,200

1 token ≈ 4 chars. A typical chat prompt with system prompt is 800–2000.

400
$1.25

GPT-5: $1.25, Claude Sonnet 4: $3, Gemini 2.5 Flash: $0.30.

$10.00

GPT-5: $10, Claude Sonnet 4: $15, Gemini 2.5 Flash: $2.50.

50%

Prompt-caching cuts repeated input cost 50–90%.

Formula used

LLM cost formula

Output is 5–15× more expensive than input. Cap max_tokens aggressively and stream early to keep costs down.

(input tokens × input $/1M) + (output tokens × output $/1M)
GPT-5 input
$1.25/M
GPT-5 output
$10/M
Cache savings
Up to 90%
Backlink-friendly embed

Embed this calculator

Free to embed on any site. Inputs preserved, link back to RevenueLab. Each format trades polish for SEO juice.

<iframe src="https://revenuelab.fyi/embed/ai-model-cost-calculator?requests=100000&inputTokens=1200&outputTokens=400&inputPrice=1.25&outputPrice=10&cachingDiscount=50" width="100%" height="680" style="border:0;border-radius:12px;max-width:100%" loading="lazy" title="AI Model Cost Calculator"></iframe>
<p style="font:12px/1.4 system-ui;color:#666;margin:6px 0 0">Calculator by <a href="https://revenuelab.fyi/ai-model-cost-calculator?requests=100000&inputTokens=1200&outputTokens=400&inputPrice=1.25&outputPrice=10&cachingDiscount=50" target="_blank" rel="noopener">RevenueLab</a></p>

Easiest to install — passes referral traffic and a referring-domain signal.

Cite this calculator

Writing about this topic? Grab a citation — every link helps keep these tools free.

APA
RevenueLab. (2026). AI Model Cost Calculator. Retrieved from https://revenuelab.fyi/ai-model-cost-calculator
HTML
<p>Source: <a href="https://revenuelab.fyi/ai-model-cost-calculator" target="_blank" rel="noopener">AI Model Cost Calculator — RevenueLab</a> (2026).</p>
Markdown
Source: [AI Model Cost Calculator — RevenueLab](https://revenuelab.fyi/ai-model-cost-calculator) (2026).

Where AI costs actually leak

Most teams under-estimate output tokens by 3–5×. A 'short answer' from an unconstrained model averages 400–800 tokens. The fix is hard system-prompt limits ('respond in <= 80 words') and aggressive max_tokens, not just choosing a cheaper model.

  • Prompt caching cuts repeated system-prompt cost by 50–90%.
  • Batch API endpoints are 50% cheaper for non-realtime work.
  • Cheaper models often need 1.5× more retries — net cost can be higher.

Pricing your AI feature

If a user runs ~100 requests/mo at $0.02 cost/request, you need to charge $7+/mo for a 70% gross margin. Free tiers should hard-cap requests, not tokens — easier to communicate, harder to abuse.

FAQ

How much does GPT-5 cost per request?

At ~1200 input / 400 output tokens (typical chat), GPT-5 costs about $0.005 per request before caching, or $0.003 with 70% input caching. Roughly half the price of GPT-4o.

Is Claude or Gemini cheaper than GPT-5?

Gemini 2.5 Flash is the cheapest tier ($0.30 in / $2.50 out per 1M). Claude Sonnet 4.5 is the most expensive of the three ($3 / $15). GPT-5 sits in the middle.

How do I lower my AI bill?

Three biggest levers: (1) enable prompt caching, (2) cap max_tokens hard, (3) use a cheaper model for routing/classification and the big model only for final output.