DeepSeek R1 — pricing, providers, and benchmarks

Parameters
671B
Context window
131k tokens
License
mit
Released
2025-01-20

DeepSeek R1, released in January 2025 under the MIT license, was the first open-weights model to credibly match OpenAI's o1 series on reasoning benchmarks like AIME and MATH. It's a 671B-parameter MoE that uses extended chain-of-thought reasoning at inference time, which means it's more expensive per token than its non-reasoning siblings — typically $0.40–$0.55 per 1M input tokens and $2.00–$2.50 per 1M output tokens, since reasoning chains generate substantially more output tokens than a standard completion. Best for problems where verifiable correctness matters more than per-call cost: math, formal proofs, code synthesis with strict tests, and multi-step planning. For anything not requiring deep reasoning, V3.2 is a better cost choice.

Provider pricing

Sorted by total monthly cost for 100M input + 10M output tokens.

ProviderInput / 1MOutput / 1MMonthly costContext
DeepInfra$0.4000$2.0000$60.00131k
OpenRouter$0.5000$2.3000$73.00131k
Fireworks AI$0.5500$2.5000$80.00131k

Frequently asked questions

How much does it cost to run DeepSeek R1 for 100M tokens?

Running DeepSeek R1 with 100M input and 10M output tokens per month costs approximately $60.00 on DeepInfra, the cheapest available provider as of the latest pricing data. Costs vary significantly depending on your input/output ratio and whether you use prompt caching.

What is the cheapest provider for DeepSeek R1?

DeepInfra currently offers DeepSeek R1 at the lowest total cost for a standard workload. Prices change frequently — check the table above for the latest data.

What context window does DeepSeek R1 support?

DeepSeek R1 supports a context window of 131,072 tokens. Individual providers may cap this lower — see the pricing table for per-provider context limits.