LLM Cost Calculator
Estimate token-based API cost for a single request or a repeated workload. This page is an estimate tool, not a billing export or invoice reconciliation view.
AI Cost Planning
How this LLM cost calculator works
The first release stays simple on purpose: you bring the token counts, the page applies the selected model's local pricing table, and the result scales to a daily or monthly workload when needed.
Notes and boundaries
LLM cost calculator FAQ
- How is LLM cost calculated?
- The calculator multiplies the selected model's per-million-token rates by your input, output, and cached-input token counts, then scales the result by request volume when repeated mode is enabled.
- What is the difference between input and output tokens?
- Input tokens are the prompt tokens you send to the model. Output tokens are the tokens the model returns in its response.
- What are cached tokens?
- Cached tokens are prompt tokens billed at a discounted cached-input rate when the provider supports prompt caching or an equivalent cache mechanism.
- Why might this estimate differ from my actual bill?
- Real bills can differ because of tokenizer differences, retries, tool calls, grounding, long-context tiers, regional processing, provider-side pricing changes, and differences between aggregated pricing data and the provider's final billing logic.
- Which providers and models are included?
- The page currently focuses on OpenAI, Anthropic, and Google, and aims to cover a broader set of their priced text and reasoning models while still excluding categories like image, audio, embedding, and other non-comparable SKUs.
Related AI planning tools
These tools sit next to LLM cost math in the planned AI cluster and help users move from pricing to comparison and token planning.