insidejob
open source DeepSeek

DeepSeek V4

1T
Context 128K tokens
Max output 32K tokens
Architecture Mixture-of-Experts, 1T total parameters
Pricing (per 1M tokens) $0.28 in / $1.1 out

Benchmark scores

72.5 SWE-bench Verified
84 GPQA Diamond
1445 LM Arena Elo
Available via: DeepSeek APIOpenRouterSelf-hosted

DeepSeek V4 is the largest freely available AI model — 1 trillion parameters at a price that undercuts every competitor by an order of magnitude.

Benchmarks

BenchmarkScoreNotes
SWE-bench Verified72.5%Competitive with models 10x the price
GPQA Diamond84.0%Approaching frontier scores
LM Arena Elo1445Top-tier for open models

Pricing

Per 1M tokensvs Claude Opus
Input$0.2818x cheaper
Output$1.1023x cheaper

The price/performance ratio is staggering. For workloads that don’t need the absolute best quality, DeepSeek V4 delivers 85-90% of frontier performance at 5% of the cost.

Architecture

Builds on the V3 and R1 series, focusing on:

  • Frontier reasoning quality via massive MoE architecture
  • Improved long-context efficiency
  • Enhanced tool-use and function calling
  • Strong agentic workload performance

Variants

ModelParamsCost (input/M)Use case
DeepSeek V41T$0.28Full quality
DeepSeek V4 Lite~200B~$0.10Budget, matches frontier on limited compute

Strengths

  • Price/performance king — frontier-adjacent quality at commodity pricing
  • Open weights — self-host for zero marginal cost
  • 1T parameters deliver genuine reasoning depth
  • V4 Lite variant for ultra-budget deployments

Weaknesses

  • 128K context window — much smaller than Claude (1M) or Llama (10M)
  • Self-hosting requires massive GPU cluster
  • Chinese company — some enterprises have compliance concerns
  • Trails frontier models by 8-10 points on top benchmarks