AI Model Pricing Calculator
Compare API costs across all major AI models. Input your token counts and see per-request and monthly cost estimates.
Compare API costs across all major AI models. Input your token counts and see per-request and monthly cost estimates.
AI model pricing varies significantly across providers, models, and usage patterns. Models are typically priced per token (or per million tokens) with separate rates for input and output tokens. Factors like context window size, model capability, and whether you use batch or real-time processing all affect the total cost. Comparing these prices across providers helps teams make informed decisions about which model to use for their specific use case.
This tool provides an up-to-date comparison of pricing for popular AI models including GPT-4, Claude, Gemini, Llama, and others, with sortable columns and cost calculators.
AI API pricing is based on token usage, charged per million tokens. Most providers charge separately for input tokens (your prompt) and output tokens (the model's response). The total cost per request is the sum of input and output token costs.
The cheapest model depends on your use case. For simple tasks, smaller models like GPT-4.1 nano or Gemini 2.0 Flash offer the lowest per-token costs. For complex tasks requiring high quality, you may need more capable (and expensive) models. Use the calculator above to compare costs for your specific usage.
Prices shown are standard API rates and were last updated on the date shown at the bottom of the calculator. AI providers frequently adjust pricing, so check the provider's official pricing page for the most current rates. Volume discounts, batching, and prompt caching can also reduce costs.