LLM Pricing Comparison — find the right AI model for your project.

Answer three quick questions — what you want to do, how heavy your usage is, and which models you’d like to compare. We’ll show you the real monthly cost and feature differences, in plain language.

  • Tracks 109+ AI models, refreshed monthly.
  • Plain-language labels — no technical jargon.
  • Share your results via link.

What the data shows

As of April 2026, the cheapest production-grade AI model in Buzzi.ai's pricing database is devstral-small-2505 at $0.00 per million input tokens. We track 97 models across pricing, quality benchmarks, context window and data-residency — refreshed daily.

How it works

Three quick questions. A real cost number in return.

No sign-up, no spreadsheet, no jargon. Built for founders, product teams, and engineers who need an answer in under a minute.

  1. 01

    Pick a scenario.

    Tell us what you want AI to do — chat, code, extract data, understand images, reason, or bulk processing. We filter the model list to what matters.

  2. 02

    Set your usage.

    Share a rough sense of volume and message length. No tokens, no math — plain English with anchors like "a side project" or "production scale."

  3. 03

    Compare real costs.

    Every model card shows your personalized monthly cost. Side-by-side bars surface the cheapest pick and how much you’d save by switching.

What we track

One database. Every provider worth watching.

97 production-ready models across 21 providers — pricing, context window, benchmarks, regions, and compliance on every row. Refreshed each morning from official pricing pages, cross-checked against third-party aggregators.

Production-ready models

97

tracked

Providers covered

21

worldwide

Quality benchmarks

1

per model

Refresh cadence

Daily

price sync

Priced today

The latest flagship model from every major lab.

Prices are per 1 million tokens. Cached and batched rates apply when you reuse prompts or accept a delay. Click a row to open the full model page.

ProviderModelContextInput /1MOutput /1M
OpenAIGPT-4.5128K$75.00$150.00
OpenAIGPT-5 pro400K$15.00$120.00
OpenAIo1200K$15.00$60.00
AnthropicClaude Opus 4.7200K$5.00$25.00
AnthropicClaude Opus 4.6200K$5.00$25.00
AnthropicClaude Sonnet 4.6200K$3.00$15.00
GoogleGemini 3 Pro2M$2.00$12.00
GoogleGemini 2 Pro2M$3.50$10.50
GoogleGemini 2.5 Pro2M$1.25$10.00
MetaLLaMA 3.1 405B128K$3.00$3.00
MetaLLaMA 4 Maverick1M$1.00$3.00
MetaLLaMA 3.2 90B128K$0.6$1.80
MistralMistral Large128K$2.00$6.00
MistralMistral Medium32K$0.4$2.00
MistralMixtral 8x22B66K$1.20$1.20
Alibaba (Qwen)Qwen 2.5131K$0.5$1.50
Alibaba (Qwen)qwen3.5-9b262K$0.4$1.50
Alibaba (Qwen)qwen3.5-4b$0.3$1.50
DeepSeekDeepSeek R1128K$0.55$2.19
DeepSeekDeepSeek V3.2128K$0.27$1.10
DeepSeekDeepSeek V3128K$0.27$1.10
AmazonTitan Text Premier32K$0.5$1.50
Amazonnova-micro-v1128K$0.35$1.40
AmazonTitan Text Express8K$0.2$0.6
xAIGrok 4.1256K$3.00$15.00
xAIGrok 4256K$3.00$15.00
xAIGrok 3131K$3.00$15.00
NVIDIAnemotron-nano-9b-v2131K$0.4$1.60
NVIDIANemotron-4128K$1.00$1.00
MiniMaxMiniMax-Text-011M$0.2$1.10
MiniMaxMiniMax-011M$0.2$1.10
Moonshot AIKimi K22M$0.15$2.50

Top 3 priced models per provider by list price. Prices refreshed daily from each provider’s public pricing page.

Beyond sticker price

Five calculators that sit behind the main flow.

Once you’ve narrowed down, dig deeper — migration math, real-prompt costs, curated stacks, lifecycle risk, and compliance.

  • Switch cost calculator

    Before you migrate, see how migration engineering hours weigh against the monthly savings over 12 months.

  • Prompt cost

    Paste a real prompt and reply. Get a per-provider cost at today’s rates, tokenized with the right family coefficient.

  • Model stacks

    Editorial picks for budget, balanced, and frontier use — curated by our applied-AI team, refreshed monthly.

  • Lifecycle timeline

    Which models are sunsetting, when, and what the provider is pushing customers toward.

  • Compliance matrix

    Regions and certifications (SOC 2, HIPAA, GDPR, FedRAMP) per provider, in one grid.

FAQ

Questions we get asked most.

Pricing freshness, sourcing, cache and batch discounts, embedding, alerts — all the things teams ask before picking a model.

Get instant answers from our AI agent

As of April 2026, the lowest input $/1M on our comparison is devstral-small-2505. Real-world cost depends on your cache hit rate and batch eligibility.
We mirror pricing from official provider pricing pages and docs. Each model row has a "last verified" timestamp and a link to the source so you can check yourself.
A nightly snapshot cron diffs against the previous day. When a change is detected we log it and email subscribed users within 24 hours.
Models that offer cached input pricing get a separate column. The volume calculator multiplies your cache hit rate by the cached price and the rest by the standard input price.
Providers that support async batch endpoints usually list a reduced price. If a model row has a batch price, you can set the "batch eligible" slider to model cost savings for that workload share.
Our top recommendation: pick two candidates from the filtered shortlist, estimate break-even with the switch-cost calculator, and run your real prompts through "Compare my prompt" for a grounded test. Top 3 this month: devstral-small-2505, GPT-5 nano, Gemini 2.0 Flash-Lite.
Yes. The comparison, calculators, and public JSON API are free. Signing in with Google enables the "Compare my prompt", saved comparisons, and price alerts features.
We list the top open-weight models (Meta Llama, Mistral, DeepSeek) when a pay-per-token API exists. Self-host cost modeling is not included since it depends on your GPU inventory.
Yes — the /embed route renders a minimal iframe with attribution. Use the embed builder on the main page to generate the snippet.
After signing in you can subscribe to any model. When the nightly snapshot detects a price change or deprecation, you get an email within 24 hours.
Each task has a weighted score over benchmarks relevant to that task plus a price pillar. We publish the exact weights on the methodology page.
No. The ranking is not pay-to-play. Providers pay us nothing.

Once a month

Get the LLM Market Pulse in your inbox.

New models, quiet deprecations, price moves — digested into one short email each month. No spam, unsubscribe anytime.

Deploying at scale?

Need help picking a model for your use case?

A 30-minute call with a Buzzi applied-AI lead. We’ll look at your volume, your data, and your constraints and recommend a stack you can actually ship.