EBEasy BenchmarksLLM model index
Workspace
Overview
Benchmarks
Benchmarks list
Overall Index
Coding
Math
MMLU-Pro
Speed
Value
Models
All models
GPT-5.5 (xhigh)
GPT-5.5 (high)
Claude Opus 4.7 (Adaptive Reasoning, Max Effort)
Gemini 3.1 Pro Preview
GPT-5.4 (xhigh)
Artificial Analysis data
Back

Anthropic

Claude 4 Opus (Non-reasoning)

Claude 4 Opus (Non-reasoning) is an Anthropic Claude Opus profile, representing the high-capability end of the Claude lineup for difficult coding, agentic, research, and computer-use tasks. This page compares the exact benchmarked variant against other Claude and non-Claude models on capability, runtime, and price.

Introducing Claude Opus

Operational Metrics

Output Speed36.6 tok/s
First Token1.40s
Blended Price$30.00/M

Model Metadata

Queryable facts extracted from the upstream model payload.

ReleaseMay 22, 2025
Context Windown/a
Modalitiesn/a
API fields: release_date

Strength: MMLU-Pro

Rank #20 across 345 models.

86.0%

Strength: SciCode

Rank #79 across 472 models.

40.9%

Strength: Overall

Rank #119 across 500 models.

33.0

Watch Area: Value

Rank #318 across 323 models.

1.1

Watch Area: Output $

Rank #319 across 325 models.

$75.00/M

Watch Area: Blended $

Rank #318 across 325 models.

$30.00/M

Strength Profile

Percentile score by analysis domain.

* Cost is inverted: lower input, output, and blended prices rank higher.

Benchmark Percentiles

Higher bars mean stronger relative placement.

All Benchmarks

MetricDomainValueRank
Artificial Analysis Intelligence Indexoverall33.0#119
Artificial Analysis Math Indexmath36.3#171
MMLU-Proreasoning86.0%#20
GPQAreasoning70.1%#197
Humanity's Last Exam
reasoning
5.9%
#241
LiveCodeBenchcoding54.2%#136
SciCodecoding, reasoning40.9%#79
MATH-500math94.1%#49
AIMEmath56.3%#53
Output Speedspeed36.6 tok/s#259
Time to First Tokenspeed1.40s#197
Blended Pricecost$30.00/M#318
Input Pricecost$15.00/M#316
Output Pricecost$75.00/M#319
Value Indexcost, overall1.1#318