LIVE
ANTHROPICOpus 4.7 benchmarks published2m ago
CLAUDEOK142ms
OPUS 4.7$15 / $75per Mtok
CHATGPTOK89ms
HACKERNEWSWhy has not AI improved design quality the way it improved dev speed?14m ago
MMLU-PROleader Opus 4.788.4
GEMINIDEGRADED312ms
MISTRALMistral Medium 3 released6m ago
GPT-4o$5 / $15per Mtok
ARXIVCompositional reasoning in LRMs22m ago
BEDROCKOK178ms
GEMINI 2.5$3.50 / $10.50per Mtok
THE VERGEFrontier Model Forum expansion announced38m ago
SWE-BENCHleader Claude Opus 4.772.1%
MISTRALOK104ms
ANTHROPICOpus 4.7 benchmarks published2m ago
CLAUDEOK142ms
OPUS 4.7$15 / $75per Mtok
CHATGPTOK89ms
HACKERNEWSWhy has not AI improved design quality the way it improved dev speed?14m ago
MMLU-PROleader Opus 4.788.4
GEMINIDEGRADED312ms
MISTRALMistral Medium 3 released6m ago
GPT-4o$5 / $15per Mtok
ARXIVCompositional reasoning in LRMs22m ago
BEDROCKOK178ms
GEMINI 2.5$3.50 / $10.50per Mtok
THE VERGEFrontier Model Forum expansion announced38m ago
SWE-BENCHleader Claude Opus 4.772.1%
MISTRALOK104ms

DeepSeek is the Chinese AI lab that keeps closing the gap with frontier proprietary models while releasing everything under the MIT license. Their V4 family, launched in April 2026, includes V4 Pro (1.6 trillion parameters, 49B active) and V4 Flash (284B total, 13B active), both with native 1M token context windows. V4 Pro scored 80.6% on SWE-bench Verified, within 0.2 points of Claude Opus 4.6. At $1.74 per million input tokens for Pro and $0.14 for Flash, they offer near-frontier performance at a fraction of proprietary pricing.

Founded

2023

Headquarters

Hangzhou, China

CEO

Liang Wenfeng

Models

2 active

Key Products

DeepSeek V4 ProDeepSeek V4 FlashDeepSeek API

Strengths

  • MIT open source license
  • Near-frontier benchmarks
  • Ultra-competitive pricing
  • Native 1M context
  • Strong coding performance

DeepSeek Models

ModelInput / 1MOutput / 1MContextCapabilities
DeepSeek V4 Pro$1.74$3.481Mtext, vision, code, reasoning
DeepSeek V4 Flash$0.14$0.281Mtext, vision, code

Prices per 1M tokens in USD. See the full pricing guide for detailed analysis.

Benchmark Scores

ModelMMLU-ProHumanEvalGPQA DiamondMATHSWE-bench
DeepSeek V4 Pro91.594.873.192.463.8
DeepSeek V4 Flash85.289.458.782.148.9
DeepSeek V388.191.263.585.951.4

Other Providers

Visit DeepSeekCompare ModelsAll Models