1.4M Downloads Updated 4 months ago
A strong Mixture-of-Experts (MoE) language model with 671B total parameters with 37B activated for each token.
671b
5 models
deepseek-v3:latest
5da0e2d4a9e0 • 404GB •
4K context window •
Text input •
4 months ago
Text input •
4 months ago
5da0e2d4a9e0 · 4 months ago
deepseek-v3:671b
latest
5da0e2d4a9e0 • 404GB •
4K context window •
Text input •
4 months ago
Text input •
4 months ago
5da0e2d4a9e0 · 4 months ago
deepseek-v3:671b-q4_K_M
5da0e2d4a9e0 • 404GB •
4K context window •
Text input •
4 months ago
Text input •
4 months ago
5da0e2d4a9e0 · 4 months ago
deepseek-v3:671b-q8_0
96061c74c1a5 • 713GB •
4K context window •
Text input •
4 months ago
Text input •
4 months ago
96061c74c1a5 · 4 months ago