Difficult broad knowledge and reasoning benchmark score.
Humanity's Last Exam is a broad expert benchmark from CAIS and Scale AI. The public project describes 2,500 difficult questions across many subjects, with closed-ended answers for automatic grading and held-out questions to monitor overfitting.
Test type: Closed-ended expert reasoning and knowledge benchmark with automatic grading.
474 models have this metric.
Current leader: Gemini 3.1 Pro Preview
Project links
This app ranks the HLE score exposed by the Artificial Analysis snapshot.
Top models ranked by HLE.
| Rank | Model | Creator | Value | Speed | Blended Price |
|---|---|---|---|---|---|
| #1 | Gemini 3.1 Pro Preview | 44.7% | 131.2 tok/s | $4.50/M | |
| #2 |
| OpenAI |
| 44.3% |
| 66.1 tok/s |
| $11.25/M |
| #3 | GPT-5.5 (high) | OpenAI | 43.0% | 59.3 tok/s | $11.25/M |
| #4 | GPT-5.4 (xhigh) | OpenAI | 41.6% | 93.5 tok/s | $5.63/M |
| #5 | GPT-5.5 (medium) | OpenAI | 40.6% | 57.5 tok/s | $11.25/M |
| #6 | GPT-5.3 Codex (xhigh) | OpenAI | 39.9% | 87.1 tok/s | $4.81/M |
| #7 | Muse Spark | Meta | 39.9% | n/a | - |
| #8 | Claude Opus 4.7 (Adaptive Reasoning, Max Effort) | Anthropic | 39.6% | 51.8 tok/s | $10.00/M |
| #9 | Gemini 3 Pro Preview (high) | 37.2% | 128.7 tok/s | $4.50/M |
| #10 | Claude Opus 4.6 (Adaptive Reasoning, Max Effort) | Anthropic | 36.7% | 49.9 tok/s | $10.00/M |
| #11 | DeepSeek V4 Pro (Reasoning, Max Effort) | DeepSeek | 35.9% | 34.3 tok/s | $2.18/M |
| #12 | Kimi K2.6 | Kimi | 35.9% | 29.1 tok/s | $1.71/M |
| #13 | GPT-5.2 (xhigh) | OpenAI | 35.4% | 71.8 tok/s | $4.81/M |
| #14 | Gemini 3 Flash Preview (Reasoning) | 34.7% | 193.2 tok/s | $1.13/M |
| #15 | MiMo-V2.5-Pro | Xiaomi | 33.8% | 59.9 tok/s | $1.50/M |
| #16 | DeepSeek V4 Pro (Reasoning, High Effort) | DeepSeek | 33.5% | 32.9 tok/s | $2.18/M |
| #17 | GPT-5.2 Codex (xhigh) | OpenAI | 33.5% | 87.7 tok/s | $4.81/M |
| #18 | KAT-Coder-Pro V1 | KwaiKAT | 33.4% | 117.1 tok/s | $0.525/M |
| #19 | Grok 4.20 0309 v2 (Reasoning) | xAI | 32.2% | 89.3 tok/s | $3.00/M |
| #20 | DeepSeek V4 Flash (Reasoning, Max Effort) | DeepSeek | 32.1% | 77.4 tok/s | $0.175/M |
| #21 | Claude Opus 4.7 (Non-reasoning, High Effort) | Anthropic | 31.2% | 43 tok/s | $10.00/M |
| #22 | GPT-5.5 (low) | OpenAI | 31.0% | 56.8 tok/s | $11.25/M |
| #23 | Claude Sonnet 4.6 (Adaptive Reasoning, Max Effort) | Anthropic | 30.0% | 68 tok/s | $6.00/M |
| #24 | Grok 4.20 0309 (Reasoning) | xAI | 30.0% | 87.8 tok/s | $3.00/M |
| #25 | Kimi K2.5 (Reasoning) | Kimi | 29.4% | 31.6 tok/s | $1.20/M |
| #26 | GPT-5.4 (low) | OpenAI | 28.9% | 59.1 tok/s | $5.63/M |
| #27 | Qwen3.6 Max Preview | Alibaba | 28.9% | 33.2 tok/s | $2.93/M |
| #28 | Claude Opus 4.5 (Reasoning) | Anthropic | 28.4% | 57 tok/s | $10.00/M |
| #29 | MiMo-V2-Pro | Xiaomi | 28.3% | n/a | - |
| #30 | MiniMax-M2.7 | MiniMax | 28.1% | 43.9 tok/s | $0.525/M |
| #31 | GLM-5.1 (Reasoning) | Z AI | 28.0% | 45.7 tok/s | $2.15/M |
| #32 | DeepSeek V4 Flash (Reasoning, High Effort) | DeepSeek | 27.8% | n/a | $0.175/M |
| #33 | Gemini 3 Pro Preview (low) | 27.6% | n/a | $4.50/M |
| #34 | Qwen3.5 397B A17B (Reasoning) | Alibaba | 27.3% | 50.4 tok/s | $1.35/M |
| #35 | GLM-5 (Reasoning) | Z AI | 27.2% | 64.5 tok/s | $1.55/M |
| #36 | GPT-5.4 mini (xhigh) | OpenAI | 26.6% | 158.9 tok/s | $1.69/M |
| #37 | GPT-5 (high) | OpenAI | 26.5% | 84.2 tok/s | $3.44/M |
| #38 | GPT-5.1 (high) | OpenAI | 26.5% | 123.3 tok/s | $3.44/M |
| #39 | GPT-5.4 nano (xhigh) | OpenAI | 26.5% | 160.3 tok/s | $0.463/M |
| #40 | Qwen3 Max Thinking | Alibaba | 26.2% | 34.3 tok/s | $2.40/M |
| #41 | DeepSeek V3.2 Speciale | DeepSeek | 26.1% | n/a | - |
| #42 | Qwen3.6 Plus | Alibaba | 25.7% | 53.1 tok/s | $1.13/M |
| #43 | GLM-5.1 (Non-reasoning) | Z AI | 25.6% | 41.5 tok/s | $2.15/M |
| #44 | GPT-5 Codex (high) | OpenAI | 25.6% | 166.8 tok/s | $3.44/M |
| #45 | Hy3-preview (Reasoning) | Tencent | 25.5% | 86.4 tok/s | - |
| #46 | GLM-5-Turbo | Z AI | 25.4% | n/a | - |
| #47 | MiMo-V2.5 | Xiaomi | 25.2% | n/a | - |
| #48 | GLM-4.7 (Reasoning) | Z AI | 25.1% | 90.3 tok/s | $1.00/M |
| #49 | GPT-5.2 (medium) | OpenAI | 24.9% | n/a | $4.81/M |
| #50 | Grok 4.20 0309 v2 (Non-reasoning) | xAI | 24.2% | 86.6 tok/s | $3.00/M |
| #51 | Grok 4 | xAI | 23.9% | 50.3 tok/s | $6.00/M |
| #52 | GPT-5 (medium) | OpenAI | 23.5% | 82.3 tok/s | $3.44/M |
| #53 | GPT-5.1 Codex (high) | OpenAI | 23.4% | 162.7 tok/s | $3.44/M |
| #54 | Qwen3.5 122B A10B (Reasoning) | Alibaba | 23.4% | 139.9 tok/s | $1.10/M |
| #55 | Gemma 4 31B (Reasoning) | 22.7% | 34.8 tok/s | - |
| #56 | Step 3.5 Flash 2603 | StepFun | 22.6% | 132.3 tok/s | - |
| #57 | Grok 4.20 0309 (Non-reasoning) | xAI | 22.5% | 77.1 tok/s | $3.00/M |
| #58 | Kimi K2 Thinking | Kimi | 22.3% | 99 tok/s | $1.08/M |
| #59 | DeepSeek V3.2 (Reasoning) | DeepSeek | 22.2% | n/a | $0.315/M |
| #60 | MiniMax-M2.1 | MiniMax | 22.2% | 84.8 tok/s | $0.525/M |
| #61 | Qwen3.5 27B (Reasoning) | Alibaba | 22.2% | 87 tok/s | $0.825/M |
| #62 | Qwen3.6 27B (Reasoning) | Alibaba | 21.6% | 64.1 tok/s | $1.35/M |
| #63 | Gemini 2.5 Pro | 21.1% | 120.2 tok/s | $3.44/M |
| #64 | MiMo-V2-Flash (Reasoning) | Xiaomi | 21.1% | 118.8 tok/s | $0.150/M |
| #65 | MiMo-V2-Omni-0327 | Xiaomi | 20.4% | n/a | - |
| #66 | Qwen3.6 35B A3B (Reasoning) | Alibaba | 20.2% | 191.8 tok/s | $0.557/M |
| #67 | MiMo-V2-Flash (Feb 2026) | Xiaomi | 20.0% | 120.6 tok/s | $0.150/M |
| #68 | o3 | OpenAI | 20.0% | 72.7 tok/s | $3.50/M |
| #69 | MiMo-V2-Omni | Xiaomi | 19.9% | n/a | - |
| #70 | GPT-5 mini (high) | OpenAI | 19.7% | 85.7 tok/s | $0.688/M |
| #71 | Qwen3.5 35B A3B (Reasoning) | Alibaba | 19.7% | 137.7 tok/s | $0.688/M |
| #72 | NVIDIA Nemotron 3 Super 120B A12B (Reasoning) | NVIDIA | 19.2% | 162.5 tok/s | $0.412/M |
| #73 | MiniMax-M2.5 | MiniMax | 19.1% | 79.7 tok/s | $0.525/M |
| #74 | Step 3.5 Flash | StepFun | 19.1% | 123.6 tok/s | $0.150/M |
| #75 | Qwen3.5 397B A17B (Non-reasoning) | Alibaba | 18.8% | 52.5 tok/s | $1.35/M |
| #76 | Claude Opus 4.6 (Non-reasoning, High Effort) | Anthropic | 18.6% | 42 tok/s | $10.00/M |
| #77 | gpt-oss-120B (high) | OpenAI | 18.5% | 212.3 tok/s | $0.263/M |
| #78 | GPT-5 (low) | OpenAI | 18.4% | 65.8 tok/s | $3.44/M |
| #79 | Gemma 4 26B A4B (Reasoning) | 18.3% | n/a | $0.198/M |
| #80 | Kimi K2.6 (Non-reasoning) | Kimi | 18.2% | n/a | - |