EBEasy BenchmarksLLM model index
Workspace
Overview
Benchmarks
Benchmarks list
Overall Index
Coding
Math
MMLU-Pro
Speed
Value
Models
All models
GPT-5.5 (xhigh)
GPT-5.5 (high)
Claude Opus 4.7 (Adaptive Reasoning, Max Effort)
Gemini 3.1 Pro Preview
GPT-5.4 (xhigh)
Artificial Analysis data
Back

Blended Price

Price per 1M tokens using the Artificial Analysis 3:1 blended ratio.

Blended Price is a cost metric for comparing providers with different input and output token prices. Artificial Analysis calculates it with a 3:1 input-to-output token ratio so a single number can summarize typical API cost.

Test type: Provider price normalization. Lower values rank better.

Coverage

325 models have this metric.

$0.020/M

Current leader: Qwen3.5 0.8B (Non-reasoning)

Project links

Prices come from Artificial Analysis pricing data in the committed snapshot.

Artificial Analysis methodology

Top Blended $ Models

Top models ranked by Blended $.

Leaderboard

RankModelCreatorValueSpeedBlended Price
#1Qwen3.5 0.8B (Non-reasoning)Alibaba$0.020/M273.6 tok/s$0.020/M
#2
Qwen3.5 0.8B (Reasoning)
Alibaba
$0.020/M
n/a
$0.020/M
#3Gemma 3n E4B InstructGoogle$0.025/M15.3 tok/s$0.025/M
#4Qwen3.5 2B (Non-reasoning)Alibaba$0.040/M227 tok/s$0.040/M
#5Qwen3.5 2B (Reasoning)Alibaba$0.040/Mn/a$0.040/M
#6LFM2 24B A2BLiquid AI$0.052/M196.9 tok/s$0.052/M
#7Qwen3.5 4B (Non-reasoning)Alibaba$0.060/M200.2 tok/s$0.060/M
#8Qwen3.5 4B (Reasoning)Alibaba$0.060/M204.8 tok/s$0.060/M
#9Nova MicroAmazon$0.061/M332.1 tok/s$0.061/M
#10Granite 4.1 8BIBM$0.063/M134.6 tok/s$0.063/M
#11Llama 3 Instruct 8BMeta$0.070/M82.2 tok/s$0.070/M
#12NVIDIA Nemotron Nano 9B V2 (Reasoning)NVIDIA$0.070/M121.6 tok/s$0.070/M
#13Granite 3.3 8B (Non-reasoning)IBM$0.085/M410.5 tok/s$0.085/M
#14NVIDIA Nemotron Nano 9B V2 (Non-reasoning)NVIDIA$0.086/M153.3 tok/s$0.086/M
#15Qwen2.5 TurboAlibaba$0.087/M77.7 tok/s$0.087/M
#16NVIDIA Nemotron 3 Nano 30B A3B (Reasoning)NVIDIA$0.096/M154.8 tok/s$0.096/M
#17gpt-oss-20B (high)OpenAI$0.100/M242.3 tok/s$0.100/M
#18Llama 2 Chat 7BMeta$0.100/M99.7 tok/s$0.100/M
#19Llama 3.1 Instruct 8BMeta$0.100/M164.4 tok/s$0.100/M
#20Llama 3.2 Instruct 1BMeta$0.100/M97.7 tok/s$0.100/M
#21Ministral 3 3BMistral$0.100/M287.6 tok/s$0.100/M
#22Nova LiteAmazon$0.105/M186.8 tok/s$0.105/M
#23Granite 4.0 H SmallIBM$0.107/M238.9 tok/s$0.107/M
#24gpt-oss-20B (low)OpenAI$0.108/M249.7 tok/s$0.108/M
#25Qwen3.5 9B (Reasoning)Alibaba$0.113/M62.9 tok/s$0.113/M
#26Apertus 8B InstructSwiss AI Initiative$0.125/Mn/a$0.125/M
#27Olmo 3 7B InstructAllen Institute for AI$0.125/Mn/a$0.125/M
#28GPT-5 nano (high)OpenAI$0.138/M136 tok/s$0.138/M
#29GPT-5 nano (medium)OpenAI$0.138/M150.3 tok/s$0.138/M
#30GPT-5 nano (minimal)OpenAI$0.138/M139.1 tok/s$0.138/M
#31Devstral Small (Jul '25)Mistral$0.150/M194.2 tok/s$0.150/M
#32Ling 2.6 FlashInclusionAI$0.150/M206 tok/s$0.150/M
#33Llama 3.2 Instruct 3BMeta$0.150/M52.2 tok/s$0.150/M
#34MiMo-V2-Flash (Feb 2026)Xiaomi$0.150/M120.6 tok/s$0.150/M
#35MiMo-V2-Flash (Non-reasoning)Xiaomi$0.150/M116.7 tok/s$0.150/M
#36MiMo-V2-Flash (Reasoning)Xiaomi$0.150/M118.8 tok/s$0.150/M
#37Ministral 3 8BMistral$0.150/M157.6 tok/s$0.150/M
#38Mistral Small 3Mistral$0.150/M135.9 tok/s$0.150/M
#39Mistral Small 3.1Mistral$0.150/M138.8 tok/s$0.150/M
#40Mistral Small 3.2Mistral$0.150/M153.8 tok/s$0.150/M
#41Solar MiniUpstage$0.150/M41.7 tok/s$0.150/M
#42Step 3.5 FlashStepFun$0.150/M123.6 tok/s$0.150/M
#43GLM-4.7-Flash (Non-reasoning)Z AI$0.152/M89.6 tok/s$0.152/M
#44GLM-4.7-Flash (Reasoning)Z AI$0.152/M110.5 tok/s$0.152/M
#45DeepSeek V4 Flash (Reasoning, High Effort)DeepSeek$0.175/Mn/a$0.175/M
#46DeepSeek V4 Flash (Reasoning, Max Effort)DeepSeek$0.175/M77.4 tok/s$0.175/M
#47Gemini 2.5 Flash-Lite (Non-reasoning)Google$0.175/M239.9 tok/s$0.175/M
#48Gemini 2.5 Flash-Lite (Reasoning)Google$0.175/M243.6 tok/s$0.175/M
#49Gemini 2.5 Flash-Lite Preview (Sep '25) (Non-reasoning)Google$0.175/Mn/a$0.175/M
#50Gemini 2.5 Flash-Lite Preview (Sep '25) (Reasoning)Google$0.175/Mn/a$0.175/M
#51GPT-4.1 nanoOpenAI$0.175/M125.2 tok/s$0.175/M
#52Llama Nemotron Super 49B v1.5 (Non-reasoning)NVIDIA$0.175/M51.3 tok/s$0.175/M
#53Llama Nemotron Super 49B v1.5 (Reasoning)NVIDIA$0.175/M50.8 tok/s$0.175/M
#54Qwen3 0.6B (Non-reasoning)Alibaba$0.188/M204.8 tok/s$0.188/M
#55Qwen3 1.7B (Non-reasoning)Alibaba$0.188/M139 tok/s$0.188/M
#56Qwen3 4B (Non-reasoning)Alibaba$0.188/M104.4 tok/s$0.188/M
#57Gemma 4 26B A4B (Reasoning)Google$0.198/Mn/a$0.198/M
#58Hermes 4 - Llama-3.1 70B (Non-reasoning)Nous Research$0.198/M83.5 tok/s$0.198/M
#59Hermes 4 - Llama-3.1 70B (Reasoning)Nous Research$0.198/M78.6 tok/s$0.198/M
#60Ministral 3 14BMistral$0.200/M121.6 tok/s$0.200/M
#61Phi-4Microsoft$0.219/M41.6 tok/s$0.219/M
#62Llama 3.2 Instruct 11B (Vision)Meta$0.245/M77.4 tok/s$0.245/M
#63Ling-flash-2.0InclusionAI$0.247/M87.3 tok/s$0.247/M
#64Ring-flash-2.0InclusionAI$0.247/M91 tok/s$0.247/M
#65Jamba 1.5 MiniAI21 Labs$0.250/Mn/a$0.250/M
#66Jamba 1.6 MiniAI21 Labs$0.250/M184.5 tok/s$0.250/M
#67Mistral 7B InstructMistral$0.250/M156.9 tok/s$0.250/M
#68Gemini 2.0 Flash (Feb '25)Google$0.263/Mn/a$0.263/M
#69GPT-4o miniOpenAI$0.263/M59.9 tok/s$0.263/M
#70gpt-oss-120B (high)OpenAI$0.263/M212.3 tok/s$0.263/M
#71gpt-oss-120B (low)OpenAI$0.263/M216.3 tok/s$0.263/M
#72Mistral Small 4 (Non-reasoning)Mistral$0.263/M139.5 tok/s$0.263/M
#73Mistral Small 4 (Reasoning)Mistral$0.263/M149.5 tok/s$0.263/M
#74Grok 4 Fast (Non-reasoning)xAI$0.275/M77.4 tok/s$0.275/M
#75Grok 4 Fast (Reasoning)xAI$0.275/M76.2 tok/s$0.275/M
#76Grok 4.1 Fast (Non-reasoning)xAI$0.275/M112.1 tok/s$0.275/M
#77Grok 4.1 Fast (Reasoning)xAI$0.275/M140.9 tok/s$0.275/M
#78Qwen3.5 Omni FlashAlibaba$0.275/M190.4 tok/s$0.275/M
#79Llama 4 ScoutMeta$0.292/M109.2 tok/s$0.292/M
#80Hermes 3 - Llama-3.1 70BNous Research$0.300/M28.8 tok/s$0.300/M