EBEasy BenchmarksLLM model index
Workspace
Overview
Benchmarks
Benchmarks list
Overall Index
Coding
Math
MMLU-Pro
Speed
Value
Models
All models
GPT-5.5 (xhigh)
GPT-5.5 (high)
Claude Opus 4.7 (Adaptive Reasoning, Max Effort)
Gemini 3.1 Pro Preview
GPT-5.4 (xhigh)
Artificial Analysis data
Back

DeepSeek

DeepSeek R1 (Jan '25)

DeepSeek R1 (Jan '25) is a DeepSeek model profile, from a family known for strong reasoning, coding, and cost-conscious API models such as R1 and V-series releases. This page separates the public release context from measured benchmark performance so you can inspect its quality and efficiency directly.

DeepSeek release notes

Operational Metrics

Output Speedn/a
First Tokenn/a
Blended Price$2.36/M

Model Metadata

Queryable facts extracted from the upstream model payload.

ReleaseJan 20, 2025
Context Windown/a
Modalitiesn/a
API fields: release_date

Strength: MMLU-Pro

Rank #37 across 345 models.

84.4%

Strength: MATH-500

Rank #31 across 201 models.

96.6%

Strength: AIME

Rank #44 across 194 models.

68.3%

Watch Area: Value

Rank #261 across 323 models.

8.0

Watch Area: Input $

Rank #253 across 325 models.

$1.68/M

Watch Area: Blended $

Rank #228 across 325 models.

$2.36/M

Strength Profile

Percentile score by analysis domain.

* Cost is inverted: lower input, output, and blended prices rank higher.

Benchmark Percentiles

Higher bars mean stronger relative placement.

All Benchmarks

MetricDomainValueRank
Artificial Analysis Intelligence Indexoverall18.8#248
Artificial Analysis Coding Indexcoding15.9#232
Artificial Analysis Math Indexmath68.0#101
MMLU-Proreasoning84.4%#37
GPQA
reasoning
70.8%
#195
Humanity's Last Examreasoning9.3%#165
LiveCodeBenchcoding61.7%#111
SciCodecoding, reasoning35.7%#181
MATH-500math96.6%#31
AIMEmath68.3%#44
Blended Pricecost$2.36/M#228
Input Pricecost$1.68/M#253
Output Pricecost$4.70/M#225
Value Indexcost, overall8.0#261