LLM Cost Calculator

Compare API costs across providers with real 2025 pricing. Calculate monthly spend or per-conversation cost instantly.

Usage Parameters

Models (price per 1M tokens: input/output)

Cost Breakdown

Model Provider Total Tokens Cost

Frequently Asked Questions

How much does it cost to run an LLM?
LLM costs vary dramatically by model and method. API costs range from $0.25/1M tokens (Claude Haiku) to $75/1M tokens (Claude Opus output). Self-hosting on GPU instances costs $1,000-$10,000+/month depending on model size and hardware. Use our calculator to compare exact costs for your use case.
What is the cheapest LLM API?
For hosted APIs, Groq's Llama 3 70B ($0.59/$0.79 per 1M tokens) and Claude Haiku ($0.25/$1.25) are among the cheapest options while maintaining high quality. The "cheapest" depends on your quality requirements — our calculator helps you find the best cost-performance ratio.
GPT-4 vs Claude pricing comparison?
GPT-4o costs $5/$15 per 1M tokens while Claude 3.5 Sonnet costs $3/$15. GPT-4 Turbo is $10/$30 vs Claude Opus at $15/$75. For most use cases, Claude Sonnet offers better value than GPT-4o with comparable quality.
When should I self-host an LLM instead of using an API?
Self-hosting makes sense when your monthly API spend exceeds $2,000-$5,000, you need data privacy guarantees, or you require low latency at high volume. Use our break-even calculator to find the exact crossover point for your usage pattern.
How many GPUs do I need to run Llama 3 70B?
Llama 3 70B requires about 140GB of GPU memory in FP16, so you need 2x A100 80GB or 2x H100 80GB GPUs. With quantization (INT8), you can fit it on a single A100 80GB. Smaller models like Llama 3 8B fit on a single A10G 24GB.
Is this LLM cost calculator free?
Yes, KickLLM's calculator is completely free with real pricing data from all major providers. We update prices regularly to ensure accuracy. No sign-up required.