Leaderboard
Sort and filter 100+ AI models by quality, price, speed, and context window — all from a single live view. Updated April 2026.
Model | Creator | Quality | Best Price | Top Speed | Max Context | Providers | Actions | |
|---|---|---|---|---|---|---|---|---|
gpt-oss-120B (low) OSS19 providers | OpenAI | 24.47 | $0.10/M | 1843 tok/s | 131K | 19 | Compare |
If you are choosing between frontier models, sort by Quality Index first. Then apply filters for cost, output speed, or context length depending on whether your bottleneck is budget, latency, or document size.
This leaderboard helps you shortlist models quickly. For side-by-side decisions, move into Compare or use the LLM Selector to match models to a use case.
Claude Opus 4.7 leads the overall quality ranking right now. The best model for you depends on your use case — coding, cost, speed, and context length all shift the answer.
Sort by Quality Index for overall strength, then filter by price, speed, or context window to match your constraints. Move to the Compare page to put 2–4 finalists head to head.
Open-weight models like DeepSeek, Qwen, and Llama consistently lead on quality-to-price. Use the price axis in the scatter chart to see which models deliver the most quality per dollar.
Data is pulled from Artificial Analysis and refreshed automatically. New models appear as soon as they have benchmark scores and provider endpoints.
Model rankings
Browse the latest ranking pages for overall models, coding, open source, Ollama, long context, and agentic workflows.
Current coding leaderboard using LiveCodeBench, Terminal-Bench, and SciCode.
Top open-weight models for self-hosting, Ollama, and low-cost API use.
Best local AI models by hardware tier for self-hosting on Macs, RTX GPUs, and workstations.
Ollama-first picks for coding, chat, reasoning, and low-friction local inference.
Best long-context models for large documents, codebases, and retrieval-heavy workflows.
Rankings for tool use, multi-step execution, and autonomous agent workflows.