
โก Loading model data...
Side by side
Compare any AI models on one screen using live benchmarks, pricing, output speed, and context window data โ up to 4 models at once, with a full benchmark breakdown when you need to go deeper.
Select up to 4 models to analyze quality, performance, pricing, and benchmarks.
Quick picks from top 10 models by Quality Index
Highest Quality
gpt-oss-120B (low)
24.47
Fastest Output
gpt-oss-120B (low)
49 tok/s
Best Value
gpt-oss-120B (low)
$0.10/M
Quality, Speed, Context, and Value at a glance
| Metric | gpt-oss-120B (low) |
|---|---|
| Creator | OpenAI |
| Quality Index | 24.47๐ |
| Price per Million Tokens | $0.10๐ฐ |
| Output Speed | 49 tok/sโก |
| Context Window | 131K๐ |
| Latency (TTFT) | 0.56s๐ |
| Provider | Novita18 more providers |
Estimated cost based on 1 million tokens per day usage
gpt-oss-120B (low)
$3
/month
Use coding, reasoning, math, and tool-use benchmarks to see where a model is actually strong instead of relying on a single overall score. A model that leads in quality may still be wrong for your workflow if your primary constraint is latency or cost.
The same model can be cheap on one host and expensive on another, or fast on one provider and unusable on the next. If the model looks promising, move to provider comparison before you commit.
Model rankings
Browse the latest ranking pages for overall models, coding, open source, Ollama, long context, and agentic workflows.
Live ranking of the best overall AI models by quality, price, speed, and context window.
Current coding leaderboard using LiveCodeBench, Terminal-Bench, and SciCode.
Top open-weight models for self-hosting, Ollama, and low-cost API use.
Best local AI models by hardware tier for self-hosting on Macs, RTX GPUs, and workstations.
Ollama-first picks for coding, chat, reasoning, and low-friction local inference.
Best long-context models for large documents, codebases, and retrieval-heavy workflows.
Rankings for tool use, multi-step execution, and autonomous agent workflows.