EBEasy BenchmarksLLM model index
Workspace
Overview
Benchmarks
Benchmarks list
Overall Index
Coding
Math
MMLU-Pro
Speed
Value
Models
All models
GPT-5.5 (xhigh)
GPT-5.5 (high)
Claude Opus 4.7 (Adaptive Reasoning, Max Effort)
Gemini 3.1 Pro Preview
GPT-5.4 (xhigh)
Artificial Analysis data
Back

Anthropic

Claude 4 Sonnet (Reasoning)

Claude 4 Sonnet (Reasoning) is an Anthropic Claude Sonnet model, a balanced Claude tier commonly aimed at strong coding, reasoning, and agent workflows with practical latency and cost. The local profile shows whether this specific reasoning mode is a better fit for quality, speed, or value-sensitive use cases.

Introducing Claude Sonnet

Operational Metrics

Output Speed50.3 tok/s
First Token8.96s
Blended Price$6.00/M

Model Metadata

Queryable facts extracted from the upstream model payload.

ReleaseMay 22, 2025
Context Windown/a
Modalitiesn/a
API fields: release_date

Strength: MATH-500

Rank #4 across 201 models.

99.1%

Strength: MMLU-Pro

Rank #40 across 345 models.

84.2%

Strength: AIME

Rank #26 across 194 models.

77.3%

Watch Area: Blended $

Rank #292 across 325 models.

$6.00/M

Watch Area: Input $

Rank #292 across 325 models.

$3.00/M

Watch Area: Output $

Rank #290 across 325 models.

$15.00/M

Strength Profile

Percentile score by analysis domain.

* Cost is inverted: lower input, output, and blended prices rank higher.

Benchmark Percentiles

Higher bars mean stronger relative placement.

All Benchmarks

MetricDomainValueRank
Artificial Analysis Intelligence Indexoverall38.7#82
Artificial Analysis Coding Indexcoding34.1#87
Artificial Analysis Math Indexmath74.3#81
MMLU-Proreasoning84.2%#40
GPQA
reasoning
77.7%
#123
Humanity's Last Examreasoning9.6%#159
LiveCodeBenchcoding65.5%#93
SciCodecoding, reasoning40.0%#95
MATH-500math99.1%#4
AIMEmath77.3%#26
Output Speedspeed50.3 tok/s#227
Time to First Tokenspeed8.96s#257
Blended Pricecost$6.00/M#292
Input Pricecost$3.00/M#292
Output Pricecost$15.00/M#290
Value Indexcost, overall6.5#272