Local LLM model page
DeepSeek V3 (671B MoE)
671B MoE with 37B active params. The original massive DeepSeek. 2.4M downloads. Server-grade only.
Parameters
671B (37B active)
Minimum RAM
512 GB
Model size
360 GB
Quantization
Q4_K_M
Can DeepSeek V3 (671B MoE) run locally?
DeepSeek V3 (671B MoE) is best suited for server-grade or multi-GPU systems. LocalClaw recommends Q4_K_M as the default quantization, with at least 512 GB RAM.
Search term for LM Studio or compatible runtimes: deepseek-v3
Hugging Face repository: deepseek-ai/DeepSeek-V3-GGUF
chatcodequality
Strengths
- Original massive DeepSeek MoE
- Only 37B active parameters despite 671B total
- Top-tier quality
- 2.4M downloads
Limitations
- Requires 512GB+ RAM
- Server-grade hardware only
- Complex setup
Best use cases
- Maximum quality outputs
- Research
- Enterprise deployment
- Frontier AI tasks
Benchmarks
Speed: 1/10
Quality: 10/10
Coding: 10/10
Reasoning: 10/10
Technical details
Developer: DeepSeek AI
License: DeepSeek License
Context window: 131,072 tokens
Architecture: Mixture of Experts (MoE) — 671B total, ~37B active
Released: 2024-12