Local LLM model page

DeepSeek V3 (671B MoE)

671B MoE with 37B active params. The original massive DeepSeek. 2.4M downloads. Server-grade only.

Parameters
671B (37B active)
Minimum RAM
512 GB
Model size
360 GB
Quantization
Q4_K_M

Can DeepSeek V3 (671B MoE) run locally?

DeepSeek V3 (671B MoE) is best suited for server-grade or multi-GPU systems. LocalClaw recommends Q4_K_M as the default quantization, with at least 512 GB RAM.

Search term for LM Studio or compatible runtimes: deepseek-v3

Hugging Face repository: deepseek-ai/DeepSeek-V3-GGUF

chatcodequality

Strengths

  • Original massive DeepSeek MoE
  • Only 37B active parameters despite 671B total
  • Top-tier quality
  • 2.4M downloads

Limitations

  • Requires 512GB+ RAM
  • Server-grade hardware only
  • Complex setup

Best use cases

  • Maximum quality outputs
  • Research
  • Enterprise deployment
  • Frontier AI tasks

Benchmarks

Speed: 1/10

Quality: 10/10

Coding: 10/10

Reasoning: 10/10

Technical details

Developer: DeepSeek AI

License: DeepSeek License

Context window: 131,072 tokens

Architecture: Mixture of Experts (MoE) — 671B total, ~37B active

Released: 2024-12