DeepSeek is a Chinese open-weight model family known for cost-efficient Mixture-of-Experts architectures. DeepSeek V3 and the R1 reasoning model match or exceed closed frontier models on coding and math at a fraction of the inference cost.
See all models from DeepSeekModels in family
6
Open weight
6
API only
0
Avg score
64.5
Top benchmark
94.2
AIME 2026
Total HF downloads
18.7M
Primary modality
Text
Context window
128K – 1.0M
First release
Dec 2024
Latest release
Apr 2026
Every release in the DeepSeek family, ranked by composite score across benchmarks, popularity, efficiency, and versatility.
| # | Model | Modality | Score | Params | Released |
|---|---|---|---|---|---|
| 1 | text | AA73.6 | 671B | Jan 2025 | |
| 2 | text | BB69.6 | 671B | Dec 2024 | |
| 3 | text | BB66.1 | 685B | Nov 2025 | |
| 4 | text | BB64.9 | 671B | Jul 2025 | |
| 5 | text | BB58.5 | 1.6T | Apr 2026 | |
| 6 | text | CC54.1 | 284B | Apr 2026 |
When each release shipped, newest first. Useful for tracking version cadence.
Apr 24
Apr 24
Nov 30
Jul 31
Jan 19
Dec 25
Composite grades across this family. Higher is better, blending benchmarks, popularity, and efficiency.
Models with downloadable weights, ranked by composite score.
| # | Model | Modality | Score | Params | Released |
|---|---|---|---|---|---|
| 1 | text | AA73.6 | 671B | Jan 2025 | |
| 2 | text | BB69.6 | 671B | Dec 2024 | |
| 3 | text | BB66.1 | 685B | Nov 2025 | |
| 4 | text | BB64.9 | 671B | Jul 2025 | |
| 5 | text | BB58.5 | 1.6T | Apr 2026 | |
| 6 | text | CC54.1 | 284B | Apr 2026 |
Spin up an instance in the cloud, or pick local hardware that fits.
Full Directory
Open the full directory to filter by hardware, capability, license, and benchmark score.
Or Browse by Provider
See every model from a lab side by side, with aggregate stats. Useful when you want a cross-family view of one provider.