The LLM Price Index
Every major large language model, normalized to dollars per million tokens. Scored on value, cheapness, and frontier capability. Independent, automated, refreshed daily.
What this is
The LLM Price Index is a live, independently maintained price comparison for every large language model offered through the OpenRouter catalog. Input and output pricing is normalized to dollars per million tokens, blended 3:1 to produce a single comparable figure, and scored 0–10 on three axes: value, cheapness, and frontier capability. The cheapest paid model right now is Gemma 3n 4B at $0.025 per million tokens (3:1 blended).
Mercury
5.6/10Mercury is the first diffusion large language model (dLLM). Applying a breakthrough discrete diffusion approach, the model runs 5-10x faster than even…
Palmyra X5
5.5/10Palmyra X5 is Writer's most advanced model, purpose-built for building and scaling AI agents across the enterprise. It delivers industry-leading speed and…
ERNIE 4.5 21B A3B
5.5/10A sophisticated text-based Mixture-of-Experts (MoE) model featuring 21B total parameters with 3B activated per token, delivering exceptional multimodal understanding and generation through…
GLM 4.6
5.5/10Compared with GLM-4.5, this generation brings several key improvements: Longer context window: The context window has been expanded from 128K to 200K…
Kimi K2 0905
5.4/10Kimi K2 0905 is the September update of [Kimi K2 0711](moonshotai/kimi-k2). It is a large-scale Mixture-of-Experts (MoE) language model developed by Moonshot…
DeepSeek V3 0324
5.3/10DeepSeek V3, a 685B-parameter, mixture-of-experts model, is the latest iteration of the flagship chat model family from the DeepSeek team. It succeeds…
Olmo 3.1 32B Instruct
5.3/10Olmo 3.1 32B Instruct is a large-scale, 32-billion-parameter instruction-tuned language model engineered for high-performance conversational AI, multi-turn dialogue, and practical instruction following.…
Relace Search
5.3/10The relace-search model uses 4-12 `view_file` and `grep` tools in parallel to explore a codebase and return relevant files to the user…
GLM 5
5.1/10GLM-5 is Z.ai’s flagship open-source foundation model engineered for complex systems design and long-horizon agent workflows. Built for expert developers, it delivers…
GLM 4.5
5.0/10GLM-4.5 is our latest flagship foundation model, purpose-built for agent-based applications. It leverages a Mixture-of-Experts (MoE) architecture and supports a context length…
Mistral Small 3
5.0/10Mistral Small 3 is a 24B-parameter language model optimized for low-latency performance across common AI tasks. Released under the Apache 2.0 license,…
Aion-2.0
5.0/10Aion-2.0 is a variant of DeepSeek V3.2 optimized for immersive roleplaying and storytelling. It is particularly strong at introducing tension, crises, and…
Rocinante 12B
4.9/10Rocinante 12B is designed for engaging storytelling and rich prose. Early testers have reported: - Expanded vocabulary with unique and expressive word…
Llama 3.1 70B Instruct
4.8/10Meta's latest class of model (Llama 3.1) launched with a variety of sizes & flavors. This 70B instruct-tuned version is optimized for…
GPT-4o-mini Search Preview
4.8/10GPT-4o mini Search Preview is a specialized model for web search in Chat Completions. It is trained to understand and execute web…
Kimi K2 0711
4.8/10Kimi K2 Instruct is a large-scale Mixture-of-Experts (MoE) language model developed by Moonshot AI, featuring 1 trillion total parameters with 32 billion…
Qwen2.5 72B Instruct
4.7/10Qwen2.5 72B is the latest series of Qwen large language models. Qwen2.5 brings the following improvements upon Qwen2: - Significantly more knowledge…
DeepSeek V3
4.7/10DeepSeek-V3 is the latest model from the DeepSeek team, building upon the instruction following and coding abilities of the previous versions. Pre-trained…
Llama 3.2 1B Instruct
4.7/10Llama 3.2 1B is a 1-billion-parameter language model focused on efficiently performing natural language tasks, such as summarization, dialogue, and multilingual text…
Saba
4.7/10Mistral Saba is a 24B-parameter language model specifically designed for the Middle East and South Asia, delivering accurate and contextually relevant responses…
Cogito v2.1 671B
4.7/10Cogito v2.1 671B MoE represents one of the strongest open models globally, matching performance of frontier closed and open models. This model…
Morph V3 Large
4.6/10Morph's high-accuracy apply model for complex code edits. ~4,500 tokens/sec with 98% accuracy for precise code transformations. The model requires the prompt…
Jamba Large 1.7
4.6/10Jamba Large 1.7 is the latest model in the Jamba open family, offering improvements in grounding, instruction-following, and overall efficiency. Built on…
MythoMax 13B
4.6/10One of the highest performing and most popular fine-tunes of Llama 2 13B, with rich descriptions and roleplay. #merge
Cydonia 24B V4.1
4.5/10Uncensored and creative writing model based on Mistral Small 3.2 24B with good recall, prompt adherence, and intelligence.
UnslopNemo 12B
4.5/10UnslopNemo v4.1 is the latest addition from the creator of Rocinante, designed for adventure writing and role-play scenarios.
Relace Apply 3
4.4/10Relace Apply 3 is a specialized code-patching LLM that merges AI-suggested edits straight into your source files. It can apply updates from…
Llama 3.1 Nemotron 70B Instruct
4.3/10NVIDIA's Llama 3.1 Nemotron 70B is a language model designed for generating precise and useful responses. Leveraging [Llama 3.1 70B](/models/meta-llama/llama-3.1-70b-instruct) architecture and…
Mistral Large 2411
4.2/10Mistral Large 2 2411 is an update of [Mistral Large 2](/mistralai/mistral-large) released together with [Pixtral Large 2411](/mistralai/pixtral-large-2411) It provides a significant upgrade…
Reka Flash 3
4.2/10Reka Flash 3 is a general-purpose, instruction-tuned large language model with 21 billion parameters, developed by Reka. It excels at general chat,…
Grok 3
4.1/10Grok 3 is the latest model from xAI. It's their flagship model that excels at enterprise use cases like data extraction, coding,…
Gemma 2 9B
4.1/10Gemma 2 9B by Google is an advanced, open-source language model that sets a new standard for efficiency and performance in its…
Switchpoint Router
4.0/10Switchpoint AI's router instantly analyzes your request and directs it to the optimal AI from an ever-evolving library. As the world of…
Command R+ (08-2024)
4.0/10command-r-plus-08-2024 is an update of the [Command R+](/models/cohere/command-r-plus) with roughly 50% higher throughput and 25% lower latencies as compared to the previous…
Command A
4.0/10Command A is an open-weights 111B parameter model with a 256k context window focused on delivering great performance across agentic, multilingual, and…
ERNIE 4.5 300B A47B
3.9/10ERNIE-4.5-300B-A47B is a 300B parameter Mixture-of-Experts (MoE) language model developed by Baidu as part of the ERNIE 4.5 series. It activates 47B…
Mixtral 8x7B Instruct
3.9/10Mixtral 8x7B Instruct is a pretrained generative Sparse Mixture of Experts, by Mistral AI, for chat and instruction use. Incorporates 8 experts…
MiniMax M2-her
3.9/10MiniMax M2-her is a dialogue-first large language model built for immersive roleplay, character-driven chat, and expressive multi-turn conversations. Designed to stay consistent…
Hermes 2 Pro - Llama-3 8B
3.8/10Hermes 2 Pro is an upgraded, retrained version of Nous Hermes 2, consisting of an updated and cleaned version of the OpenHermes…
Mixtral 8x22B Instruct
3.6/10Mistral's official instruct fine-tuned version of [Mixtral 8x22B](/models/mistralai/mixtral-8x22b). It uses 39B active parameters out of 141B, offering unparalleled cost efficiency for its…
Grok 3 Beta
3.6/10Grok 3 is the latest model from xAI. It's their flagship model that excels at enterprise use cases like data extraction, coding,…
GPT-4o Search Preview
3.6/10GPT-4o Search Previewis a specialized model for web search in Chat Completions. It is trained to understand and execute web search queries.
GPT-4 Turbo Preview
3.5/10The preview GPT-4 model with improved instruction following, JSON mode, reproducible outputs, parallel function calling, and more. Training data: up to Dec…
GPT-4 Turbo (older v1106)
3.5/10The latest GPT-4 Turbo model with vision capabilities. Vision requests can now use JSON mode and function calling. Training data: up to…
Qwen-Max
3.5/10Qwen-Max, based on Qwen2.5, provides the best inference performance among [Qwen models](/qwen), especially for complex multi-step tasks. It's a large-scale MoE model…
GPT-3.5 Turbo (older v0613)
3.5/10GPT-3.5 Turbo is OpenAI's fastest model. It can understand and generate natural language or code, and is optimized for chat and traditional…
GPT-3.5 Turbo
3.5/10GPT-3.5 Turbo is OpenAI's fastest model. It can understand and generate natural language or code, and is optimized for chat and traditional…
Mistral 7B Instruct v0.1
3.4/10A 7.3B parameter model that outperforms Llama 2 13B on all benchmarks, with optimizations for speed and context length.
Free · No spam
Price drops, new models, deprecations
Every Tuesday: which models got cheaper, which launched, which got pulled. Five minutes. No filler.
Three axes, one overall score
- Value (35%) — capability per dollar. Context length, vision, tools, and structured-output support divided by log-scaled blended price.
- Cheapness (35%) — raw affordability. Free models score 10. Paid models use an inverse log curve anchored at $0.01 / Mtok.
- Frontier (30%) — how close to the state of the art. Recent releases, long context windows, and premium pricing all contribute.
Blended price formula
- Most production workloads are input-heavy, so the index uses a 3:1 blended price:
(input × 0.75) + (output × 0.25). - All prices are normalized to dollars per million tokens. OpenRouter publishes per-token figures which we multiply by 1,000,000 before display.
Where does the pricing data come from?
Every model and price on this page is sourced from OpenRouter's public models API, which aggregates pricing directly from model providers including Anthropic, OpenAI, Google, Mistral, Meta, xAI, DeepSeek, and dozens of others. The pipeline re-fetches and re-scores the entire catalog once per day.
Why normalize to $/million tokens?
Model providers publish prices in inconsistent units — per 1K tokens, per million tokens, per character, sometimes per request. Comparing them directly is error-prone. Dollars per million tokens is the industry's most common reporting unit and makes cross-provider comparisons immediate and honest.
What does "3:1 blended" mean?
Most production LLM workloads are input-heavy — context, RAG retrievals, system prompts — while output is comparatively short. A 3:1 input:output ratio is the informal industry convention for producing a single number that reflects typical cost: (input × 0.75) + (output × 0.25). Your actual ratio may differ; always check both input and output columns for workloads with long generations.
What's the cheapest LLM right now?
The cheapest paid model as of the latest scan is Gemma 3n 4B from Google at $0.025 per million tokens (3:1 blended). Sort by "Cheapest" above for the full ranking. Many providers also offer free-tier variants of their models, which score a perfect 10 on the cheapness axis.
Is this affiliated with OpenRouter or any provider?
No. MegaOne AI is independent. OpenRouter is used as a public data source because their models API is the most complete and up-to-date LLM catalog available, but this directory is not operated by OpenRouter and we rate all models — including ones that compete with one another.
How often does the price index update?
A full re-fetch, re-score, and daily snapshot runs once per 24 hours. Snapshots are written to a history table so we can build price-over-time charts and detect drops. New models typically appear within 24 hours of being added to OpenRouter.
Is it free to use?
Yes. Browsing, filtering, sorting, and searching the entire price index is free. The weekly email briefing is free. There is no account required and no paywall.