EBEasy BenchmarksLLM model index
Workspace
Overview
Benchmarks
Benchmarks list
Overall Index
Coding
Math
MMLU-Pro
Speed
Value
Models
All models
GPT-5.5 (xhigh)
GPT-5.5 (high)
Claude Opus 4.7 (Adaptive Reasoning, Max Effort)
Gemini 3.1 Pro Preview
GPT-5.4 (xhigh)
Artificial Analysis data
Back

Artificial Analysis Coding Index

Artificial Analysis aggregate coding score.

The Coding Index is the source dataset's aggregate coding signal. Use it as a broad programming capability view before drilling into individual code benchmarks such as LiveCodeBench and SciCode.

Test type: Aggregate coding evaluation that combines code-focused benchmark results exposed by Artificial Analysis.

Coverage

410 models have this metric.

59.1

Current leader: GPT-5.5 (xhigh)

Project links

Scores come from the Artificial Analysis LLM snapshot committed in this app.

Artificial Analysis methodology

Top Coding Models

Top models ranked by Coding.

Leaderboard

RankModelCreatorValueSpeedBlended Price
#1GPT-5.5 (xhigh)OpenAI59.166.1 tok/s$11.25/M
#2
GPT-5.5 (high)
OpenAI
58.5
59.3 tok/s
$11.25/M
#3GPT-5.4 (xhigh)OpenAI57.393.5 tok/s$5.63/M
#4GPT-5.5 (medium)OpenAI56.257.5 tok/s$11.25/M
#5Gemini 3.1 Pro PreviewGoogle55.5131.2 tok/s$4.50/M
#6Claude Opus 4.7 (Non-reasoning, High Effort)Anthropic53.143 tok/s$10.00/M
#7GPT-5.3 Codex (xhigh)OpenAI53.187.1 tok/s$4.81/M
#8Claude Opus 4.7 (Adaptive Reasoning, Max Effort)Anthropic52.551.8 tok/s$10.00/M
#9GPT-5.5 (low)OpenAI52.156.8 tok/s$11.25/M
#10GPT-5.4 mini (xhigh)OpenAI51.5158.9 tok/s$1.69/M
#11Claude Sonnet 4.6 (Adaptive Reasoning, Max Effort)Anthropic50.968 tok/s$6.00/M
#12GPT-5.2 (xhigh)OpenAI48.771.8 tok/s$4.81/M
#13GPT-5.5 (Non-reasoning)OpenAI48.651.3 tok/s$11.25/M
#14Claude Opus 4.6 (Adaptive Reasoning, Max Effort)Anthropic48.149.9 tok/s$10.00/M
#15Claude Opus 4.5 (Reasoning)Anthropic47.857 tok/s$10.00/M
#16Claude Opus 4.6 (Non-reasoning, High Effort)Anthropic47.642 tok/s$10.00/M
#17DeepSeek V4 Pro (Reasoning, Max Effort)DeepSeek47.534.3 tok/s$2.18/M
#18Muse SparkMeta47.5n/a-
#19Kimi K2.6Kimi47.129.1 tok/s$1.71/M
#20Gemini 2.5 Pro Preview (Mar' 25)Google46.7n/a-
#21Gemini 3 Pro Preview (high)Google46.5128.7 tok/s$4.50/M
#22Claude Sonnet 4.6 (Non-reasoning, High Effort)Anthropic46.448.3 tok/s$6.00/M
#23GPT-5.4 (low)OpenAI45.659.1 tok/s$5.63/M
#24KAT Coder Pro V2KwaiKAT45.6110.7 tok/s$0.525/M
#25MiMo-V2.5-ProXiaomi45.559.9 tok/s$1.50/M
#26Qwen3.6 Max PreviewAlibaba44.933.2 tok/s$2.93/M
#27GPT-5.1 (high)OpenAI44.7123.3 tok/s$3.44/M
#28GLM-5 (Reasoning)Z AI44.264.5 tok/s$1.55/M
#29GPT-5.2 (medium)OpenAI44.2n/a$4.81/M
#30GPT-5.4 nano (xhigh)OpenAI43.9160.3 tok/s$0.463/M
#31GLM-5.1 (Reasoning)Z AI43.445.7 tok/s$2.15/M
#32DeepSeek V4 Pro (Reasoning, High Effort)DeepSeek43.332.9 tok/s$2.18/M
#33Claude Sonnet 4.6 (Non-reasoning, Low Effort)Anthropic43.051.5 tok/s$6.00/M
#34GPT-5.2 Codex (xhigh)OpenAI43.087.7 tok/s$4.81/M
#35Claude Opus 4.5 (Non-reasoning)Anthropic42.950.3 tok/s$10.00/M
#36Qwen3.6 PlusAlibaba42.953.1 tok/s$1.13/M
#37Gemini 3 Flash Preview (Reasoning)Google42.6193.2 tok/s$1.13/M
#38Grok 4.20 0309 (Reasoning)xAI42.287.8 tok/s$3.00/M
#39MiMo-V2.5Xiaomi42.1n/a-
#40MiniMax-M2.7MiniMax41.943.9 tok/s$0.525/M
#41MiMo-V2-ProXiaomi41.4n/a-
#42Qwen3.5 397B A17B (Reasoning)Alibaba41.350.4 tok/s$1.35/M
#43GPT-5.4 (Non-reasoning)OpenAI41.057.2 tok/s$5.63/M
#44Grok 4xAI40.550.3 tok/s$6.00/M
#45Grok 4.20 0309 v2 (Reasoning)xAI40.589.3 tok/s$3.00/M
#46DeepSeek V4 Flash (Reasoning, High Effort)DeepSeek39.8n/a$0.175/M
#47Kimi K2.5 (Reasoning)Kimi39.531.6 tok/s$1.20/M
#48Gemini 3 Pro Preview (low)Google39.4n/a$4.50/M
#49GLM-5 (Non-reasoning)Z AI39.059.6 tok/s$1.55/M
#50GPT-5 (medium)OpenAI39.082.3 tok/s$3.44/M
#51GPT-5 Codex (high)OpenAI38.9166.8 tok/s$3.44/M
#52DeepSeek V4 Flash (Reasoning, Max Effort)DeepSeek38.777.4 tok/s$0.175/M
#53Gemma 4 31B (Reasoning)Google38.734.8 tok/s-
#54Claude 4.5 Sonnet (Reasoning)Anthropic38.643.8 tok/s$6.00/M
#55DeepSeek V4 Pro (Non-reasoning)DeepSeek38.4n/a-
#56Kimi K2.6 (Non-reasoning)Kimi38.4n/a-
#57o3OpenAI38.472.7 tok/s$3.50/M
#58DeepSeek V3.2 SpecialeDeepSeek37.9n/a-
#59Gemini 3 Flash Preview (Non-reasoning)Google37.8178.3 tok/s$1.13/M
#60GPT-5.4 mini (medium)OpenAI37.5159.2 tok/s$1.69/M
#61MiniMax-M2.5MiniMax37.479.7 tok/s$0.525/M
#62Qwen3.5 397B A17B (Non-reasoning)Alibaba37.452.5 tok/s$1.35/M
#63MiMo-V2-Omni-0327Xiaomi36.9n/a-
#64GLM-5-TurboZ AI36.8n/a-
#65MiMo-V2.5-Pro (Non-reasoning)Xiaomi36.8n/a-
#66DeepSeek V3.2 (Reasoning)DeepSeek36.7n/a$0.315/M
#67GPT-5.1 Codex (high)OpenAI36.6162.7 tok/s$3.44/M
#68Claude 4.1 Opus (Reasoning)Anthropic36.535.8 tok/s$30.00/M
#69TEHy3-preview (Reasoning)Tencent36.586.4 tok/s-
#70Qwen3.6 27B (Reasoning)Alibaba36.564.1 tok/s$1.35/M
#71GPT-5.1 Codex mini (high)OpenAI36.4207.2 tok/s$0.688/M
#72GLM-4.7 (Reasoning)Z AI36.390.3 tok/s$1.00/M
#73GLM 5V Turbo (Reasoning)Z AI36.2n/a-
#74GPT-5 (high)OpenAI36.084.2 tok/s$3.44/M
#75GLM-5.1 (Non-reasoning)Z AI35.841.5 tok/s$2.15/M
#76MiMo-V2-OmniXiaomi35.5n/a-
#77GPT-5 mini (high)OpenAI35.385.7 tok/s$0.688/M
#78DeepSeek V4 Flash (Non-reasoning)DeepSeek35.1n/a-
#79Qwen3.6 35B A3B (Reasoning)Alibaba35.1191.8 tok/s$0.557/M
#80GPT-5.4 nano (medium)OpenAI35.0153.4 tok/s$0.463/M