EBEasy BenchmarksLLM model index
Workspace
Overview
Benchmarks
Benchmarks list
Overall Index
Coding
Math
MMLU-Pro
Speed
Value
Models
All models
GPT-5.5 (xhigh)
GPT-5.5 (high)
Claude Opus 4.7 (Adaptive Reasoning, Max Effort)
Gemini 3.1 Pro Preview
GPT-5.4 (xhigh)
Artificial Analysis data
Back

Anthropic

Claude 4 Opus (Reasoning)

Claude 4 Opus (Reasoning) is an Anthropic Claude Opus profile, representing the high-capability end of the Claude lineup for difficult coding, agentic, research, and computer-use tasks. This page compares the exact benchmarked variant against other Claude and non-Claude models on capability, runtime, and price.

Introducing Claude Opus

Operational Metrics

Output Speed36.8 tok/s
First Token7.03s
Blended Price$30.00/M

Model Metadata

Queryable facts extracted from the upstream model payload.

ReleaseMay 22, 2025
Context Windown/a
Modalitiesn/a
API fields: release_date

Strength: MMLU-Pro

Rank #11 across 345 models.

87.3%

Strength: MATH-500

Rank #14 across 201 models.

98.2%

Strength: AIME

Rank #30 across 194 models.

75.7%

Watch Area: Output $

Rank #320 across 325 models.

$75.00/M

Watch Area: Blended $

Rank #319 across 325 models.

$30.00/M

Watch Area: Input $

Rank #317 across 325 models.

$15.00/M

Strength Profile

Percentile score by analysis domain.

* Cost is inverted: lower input, output, and blended prices rank higher.

Benchmark Percentiles

Higher bars mean stronger relative placement.

All Benchmarks

MetricDomainValueRank
Artificial Analysis Intelligence Indexoverall39.0#80
Artificial Analysis Coding Indexcoding34.0#88
Artificial Analysis Math Indexmath73.3#85
MMLU-Proreasoning87.3%#11
GPQA
reasoning
79.6%
#104
Humanity's Last Examreasoning11.7%#126
LiveCodeBenchcoding63.6%#105
SciCodecoding, reasoning39.8%#101
MATH-500math98.2%#14
AIMEmath75.7%#30
Output Speedspeed36.8 tok/s#258
Time to First Tokenspeed7.03s#247
Blended Pricecost$30.00/M#319
Input Pricecost$15.00/M#317
Output Pricecost$75.00/M#320
Value Indexcost, overall1.3#314