Artificial Analysis aggregate coding score.
The Coding Index is the source dataset's aggregate coding signal. Use it as a broad programming capability view before drilling into individual code benchmarks such as LiveCodeBench and SciCode.
Test type: Aggregate coding evaluation that combines code-focused benchmark results exposed by Artificial Analysis.
410 models have this metric.
Current leader: GPT-5.5 (xhigh)
Project links
Scores come from the Artificial Analysis LLM snapshot committed in this app.
Top models ranked by Coding.
| Rank | Model | Creator | Value | Speed | Blended Price |
|---|---|---|---|---|---|
| #1 | GPT-5.5 (xhigh) | OpenAI | 59.1 | 66.1 tok/s | $11.25/M |
| #2 |
| OpenAI |
| 58.5 |
| 59.3 tok/s |
| $11.25/M |
| #3 | GPT-5.4 (xhigh) | OpenAI | 57.3 | 93.5 tok/s | $5.63/M |
| #4 | GPT-5.5 (medium) | OpenAI | 56.2 | 57.5 tok/s | $11.25/M |
| #5 | Gemini 3.1 Pro Preview | 55.5 | 131.2 tok/s | $4.50/M |
| #6 | Claude Opus 4.7 (Non-reasoning, High Effort) | Anthropic | 53.1 | 43 tok/s | $10.00/M |
| #7 | GPT-5.3 Codex (xhigh) | OpenAI | 53.1 | 87.1 tok/s | $4.81/M |
| #8 | Claude Opus 4.7 (Adaptive Reasoning, Max Effort) | Anthropic | 52.5 | 51.8 tok/s | $10.00/M |
| #9 | GPT-5.5 (low) | OpenAI | 52.1 | 56.8 tok/s | $11.25/M |
| #10 | GPT-5.4 mini (xhigh) | OpenAI | 51.5 | 158.9 tok/s | $1.69/M |
| #11 | Claude Sonnet 4.6 (Adaptive Reasoning, Max Effort) | Anthropic | 50.9 | 68 tok/s | $6.00/M |
| #12 | GPT-5.2 (xhigh) | OpenAI | 48.7 | 71.8 tok/s | $4.81/M |
| #13 | GPT-5.5 (Non-reasoning) | OpenAI | 48.6 | 51.3 tok/s | $11.25/M |
| #14 | Claude Opus 4.6 (Adaptive Reasoning, Max Effort) | Anthropic | 48.1 | 49.9 tok/s | $10.00/M |
| #15 | Claude Opus 4.5 (Reasoning) | Anthropic | 47.8 | 57 tok/s | $10.00/M |
| #16 | Claude Opus 4.6 (Non-reasoning, High Effort) | Anthropic | 47.6 | 42 tok/s | $10.00/M |
| #17 | DeepSeek V4 Pro (Reasoning, Max Effort) | DeepSeek | 47.5 | 34.3 tok/s | $2.18/M |
| #18 | Muse Spark | Meta | 47.5 | n/a | - |
| #19 | Kimi K2.6 | Kimi | 47.1 | 29.1 tok/s | $1.71/M |
| #20 | Gemini 2.5 Pro Preview (Mar' 25) | 46.7 | n/a | - |
| #21 | Gemini 3 Pro Preview (high) | 46.5 | 128.7 tok/s | $4.50/M |
| #22 | Claude Sonnet 4.6 (Non-reasoning, High Effort) | Anthropic | 46.4 | 48.3 tok/s | $6.00/M |
| #23 | GPT-5.4 (low) | OpenAI | 45.6 | 59.1 tok/s | $5.63/M |
| #24 | KAT Coder Pro V2 | KwaiKAT | 45.6 | 110.7 tok/s | $0.525/M |
| #25 | MiMo-V2.5-Pro | Xiaomi | 45.5 | 59.9 tok/s | $1.50/M |
| #26 | Qwen3.6 Max Preview | Alibaba | 44.9 | 33.2 tok/s | $2.93/M |
| #27 | GPT-5.1 (high) | OpenAI | 44.7 | 123.3 tok/s | $3.44/M |
| #28 | GLM-5 (Reasoning) | Z AI | 44.2 | 64.5 tok/s | $1.55/M |
| #29 | GPT-5.2 (medium) | OpenAI | 44.2 | n/a | $4.81/M |
| #30 | GPT-5.4 nano (xhigh) | OpenAI | 43.9 | 160.3 tok/s | $0.463/M |
| #31 | GLM-5.1 (Reasoning) | Z AI | 43.4 | 45.7 tok/s | $2.15/M |
| #32 | DeepSeek V4 Pro (Reasoning, High Effort) | DeepSeek | 43.3 | 32.9 tok/s | $2.18/M |
| #33 | Claude Sonnet 4.6 (Non-reasoning, Low Effort) | Anthropic | 43.0 | 51.5 tok/s | $6.00/M |
| #34 | GPT-5.2 Codex (xhigh) | OpenAI | 43.0 | 87.7 tok/s | $4.81/M |
| #35 | Claude Opus 4.5 (Non-reasoning) | Anthropic | 42.9 | 50.3 tok/s | $10.00/M |
| #36 | Qwen3.6 Plus | Alibaba | 42.9 | 53.1 tok/s | $1.13/M |
| #37 | Gemini 3 Flash Preview (Reasoning) | 42.6 | 193.2 tok/s | $1.13/M |
| #38 | Grok 4.20 0309 (Reasoning) | xAI | 42.2 | 87.8 tok/s | $3.00/M |
| #39 | MiMo-V2.5 | Xiaomi | 42.1 | n/a | - |
| #40 | MiniMax-M2.7 | MiniMax | 41.9 | 43.9 tok/s | $0.525/M |
| #41 | MiMo-V2-Pro | Xiaomi | 41.4 | n/a | - |
| #42 | Qwen3.5 397B A17B (Reasoning) | Alibaba | 41.3 | 50.4 tok/s | $1.35/M |
| #43 | GPT-5.4 (Non-reasoning) | OpenAI | 41.0 | 57.2 tok/s | $5.63/M |
| #44 | Grok 4 | xAI | 40.5 | 50.3 tok/s | $6.00/M |
| #45 | Grok 4.20 0309 v2 (Reasoning) | xAI | 40.5 | 89.3 tok/s | $3.00/M |
| #46 | DeepSeek V4 Flash (Reasoning, High Effort) | DeepSeek | 39.8 | n/a | $0.175/M |
| #47 | Kimi K2.5 (Reasoning) | Kimi | 39.5 | 31.6 tok/s | $1.20/M |
| #48 | Gemini 3 Pro Preview (low) | 39.4 | n/a | $4.50/M |
| #49 | GLM-5 (Non-reasoning) | Z AI | 39.0 | 59.6 tok/s | $1.55/M |
| #50 | GPT-5 (medium) | OpenAI | 39.0 | 82.3 tok/s | $3.44/M |
| #51 | GPT-5 Codex (high) | OpenAI | 38.9 | 166.8 tok/s | $3.44/M |
| #52 | DeepSeek V4 Flash (Reasoning, Max Effort) | DeepSeek | 38.7 | 77.4 tok/s | $0.175/M |
| #53 | Gemma 4 31B (Reasoning) | 38.7 | 34.8 tok/s | - |
| #54 | Claude 4.5 Sonnet (Reasoning) | Anthropic | 38.6 | 43.8 tok/s | $6.00/M |
| #55 | DeepSeek V4 Pro (Non-reasoning) | DeepSeek | 38.4 | n/a | - |
| #56 | Kimi K2.6 (Non-reasoning) | Kimi | 38.4 | n/a | - |
| #57 | o3 | OpenAI | 38.4 | 72.7 tok/s | $3.50/M |
| #58 | DeepSeek V3.2 Speciale | DeepSeek | 37.9 | n/a | - |
| #59 | Gemini 3 Flash Preview (Non-reasoning) | 37.8 | 178.3 tok/s | $1.13/M |
| #60 | GPT-5.4 mini (medium) | OpenAI | 37.5 | 159.2 tok/s | $1.69/M |
| #61 | MiniMax-M2.5 | MiniMax | 37.4 | 79.7 tok/s | $0.525/M |
| #62 | Qwen3.5 397B A17B (Non-reasoning) | Alibaba | 37.4 | 52.5 tok/s | $1.35/M |
| #63 | MiMo-V2-Omni-0327 | Xiaomi | 36.9 | n/a | - |
| #64 | GLM-5-Turbo | Z AI | 36.8 | n/a | - |
| #65 | MiMo-V2.5-Pro (Non-reasoning) | Xiaomi | 36.8 | n/a | - |
| #66 | DeepSeek V3.2 (Reasoning) | DeepSeek | 36.7 | n/a | $0.315/M |
| #67 | GPT-5.1 Codex (high) | OpenAI | 36.6 | 162.7 tok/s | $3.44/M |
| #68 | Claude 4.1 Opus (Reasoning) | Anthropic | 36.5 | 35.8 tok/s | $30.00/M |
| #69 | Hy3-preview (Reasoning) | Tencent | 36.5 | 86.4 tok/s | - |
| #70 | Qwen3.6 27B (Reasoning) | Alibaba | 36.5 | 64.1 tok/s | $1.35/M |
| #71 | GPT-5.1 Codex mini (high) | OpenAI | 36.4 | 207.2 tok/s | $0.688/M |
| #72 | GLM-4.7 (Reasoning) | Z AI | 36.3 | 90.3 tok/s | $1.00/M |
| #73 | GLM 5V Turbo (Reasoning) | Z AI | 36.2 | n/a | - |
| #74 | GPT-5 (high) | OpenAI | 36.0 | 84.2 tok/s | $3.44/M |
| #75 | GLM-5.1 (Non-reasoning) | Z AI | 35.8 | 41.5 tok/s | $2.15/M |
| #76 | MiMo-V2-Omni | Xiaomi | 35.5 | n/a | - |
| #77 | GPT-5 mini (high) | OpenAI | 35.3 | 85.7 tok/s | $0.688/M |
| #78 | DeepSeek V4 Flash (Non-reasoning) | DeepSeek | 35.1 | n/a | - |
| #79 | Qwen3.6 35B A3B (Reasoning) | Alibaba | 35.1 | 191.8 tok/s | $0.557/M |
| #80 | GPT-5.4 nano (medium) | OpenAI | 35.0 | 153.4 tok/s | $0.463/M |