
⥠Loading model data...
The top 30 models ranked for tool use and agents. Updated live from Artificial Analysis data. Last updated: April 2026
Want the full deep dive? See the dedicated agentic ai models guide.
Top overall
GPT-5.5 (xhigh)
QI 60.24
Best value
DeepSeek V4 Flash
$0.01/M
Fastest
Llama 3.1 Instruct 8B
2435 tok/s
Largest context
Grok 4.20 0309 v2 (Reasoning)
2.0M
| # | Model | Quality | Price/M | Speed | Context |
|---|---|---|---|---|---|
| 1 | GPT-5.5 (xhigh)OpenAI | 60.24 | $11.3 | 66 tok/s | 1.1M |
| 2 | GPT-5.5 (high)OpenAI | 58.87 | $11.3 | 55 tok/s | 1.1M |
| 3 | GPT-5.5 (medium)OpenAI | 56.71 | $11.3 | 54 tok/s | 1.1M |
| 4 | GPT-5.5 (low)OpenAI | 50.78 | $11.3 | 54 tok/s | 1.1M |
| 5 | GPT-5.5OpenAI | 40.94 | $11.3 | 56 tok/s | 1.1M |
| 6 | Claude Opus 4.7Anthropic | 57.28 | $10.0 | 111 tok/s | 1.0M |
| 7 | Gemini 3.1 Pro PreviewGoogle | 57.18 | $4.5 | 130 tok/s | 1.0M |
| 8 | GPT-5.4 (xhigh)OpenAI | 56.8 | $5.6 | 98 tok/s | 1.1M |
| 9 | GPT-5.4 (low)OpenAI | 47.94 | $5.6 | 64 tok/s | 1.1M |
| 10 | GPT-5.4OpenAI | 35.39 | $5.6 | 68 tok/s | 1.1M |
| 11 | Kimi K2.6Kimi | 53.9 | $1.1 | 215 tok/s | 262K |
| 12 | MiMo-V2.5-ProXiaomi | 53.83 | $1.2 | 64 tok/s | 1.1M |
| 13 | GPT-5.3 Codex (xhigh)OpenAI | 53.56 | $4.8 | 100 tok/s | 400K |
| 14 | Grok 4.3xAI | 53.2 | $1.6 | 92 tok/s | 1.0M |
| 15 | Claude Opus 4.6Anthropic | 52.95 | $10.9 | 65 tok/s | 1.0M |
| 16 | Qwen3.6 Max PreviewAlibaba | 51.81 | $2.9 | 38 tok/s | 256K |
| 17 | Claude Sonnet 4.6Anthropic | 51.72 | $6.6 | 100 tok/s | 1.0M |
| 18 | DeepSeek V4 ProDeepSeek | 51.51 | $1.4 | 168 tok/s | 1.0M |
| 19 | GLM-5.1 (Reasoning)Z AI | 51.41 | $1.7 | 182 tok/s | 205K |
| 20 | GLM-5.1Z AI | 43.82 | $2.1 | 149 tok/s | 205K |
| 21 | GPT-5.2 (xhigh)OpenAI | 51.28 | $4.8 | 97 tok/s | 400K |
| 22 | GPT-5.2 (medium)OpenAI | 46.64 | $4.8 | 71 tok/s | 400K |
| 23 | GPT-5.2OpenAI | 33.57 | $4.8 | 74 tok/s | 400K |
| 24 | Qwen3.6 PlusAlibaba | 49.98 | $1.1 | 53 tok/s | 1.0M |
| 25 | GLM-5 (Reasoning)Z AI | 49.77 | $1.2 | 236 tok/s | 203K |
| 26 | GLM-5Z AI | 40.57 | $0.97 | 235 tok/s | 203K |
| 27 | Claude Opus 4.5 (Reasoning)Anthropic | 49.73 | $10.0 | 70 tok/s | 200K |
| 28 | Claude Opus 4.5Anthropic | 43.09 | $10.0 | 58 tok/s | 200K |
| 29 | MiniMax-M2.7MiniMax | 49.62 | $0.52 | 454 tok/s | 205K |
| 30 | Grok 4.20 0309 v2 (Reasoning)xAI | 49.33 | $3.0 | 247 tok/s | 2.0M |
Showing top 30 of 95 ranked models
View all in Explore âEach guide goes deeper than the quick filters, with methodology, benchmarks, and picks per scenario.
Model rankings
Browse the latest ranking pages for overall models, coding, open source, Ollama, long context, and agentic workflows.
Current coding leaderboard using LiveCodeBench, Terminal-Bench, and SciCode.
Top open-weight models for self-hosting, Ollama, and low-cost API use.
Best local AI models by hardware tier for self-hosting on Macs, RTX GPUs, and workstations.
Ollama-first picks for coding, chat, reasoning, and low-friction local inference.
Best long-context models for large documents, codebases, and retrieval-heavy workflows.
Rankings for tool use, multi-step execution, and autonomous agent workflows.
Every model is scored using the Artificial Analysis Intelligence Index â a composite of GPQA Diamond, AIME 2025, LiveCodeBench, MMLU-Pro, and other benchmarks, weighted into a single 0-100 quality score. Speed, price, and context window are tracked live across providers.
The overall ranking is a starting point. For production decisions, narrow by use case using the category pages above, then compare finalists head-to-head on Compare.
GPT-5.5 (xhigh) leads on overall quality right now, but the best model depends on your priorities. Coding, cost, speed, and context length all shift the answer. Use the category rankings above to find the right fit.
DeepSeek V4 Flash currently offers one of the best quality-to-cost ratios. Open-source models on providers like Groq or Together can be even cheaper at strong quality levels.
Start with overall quality index, then narrow by what matters for your workload: cost per million tokens, output speed, context window, or a specific capability like coding or tool use. Use our Compare tool to put finalists head to head.
Llama 3.1 Instruct 8B leads on output speed right now at 2435 tokens/second. Speed matters most for real-time applications and agentic workflows with many sequential steps.
Grok 4.20 0309 v2 (Reasoning) has the biggest context window in this ranking at 2.0M. For a dedicated long-context comparison, see our largest context window page.
Data is pulled from Artificial Analysis and refreshed automatically. New models appear as soon as they have benchmark scores and provider endpoints. The ranking reflects the live state of the leaderboard.