Local LLM model page
DeepSeek R1 0528 (671B MoE)
Updated flagship DeepSeek R1 with improved reasoning chains and fewer hallucinations. Major upgrade to chain-of-thought quality. MIT licensed. Server-grade only.
Parameters
671B (37B active)
Minimum RAM
512 GB
Model size
360 GB
Quantization
Q4_K_M
Can DeepSeek R1 0528 (671B MoE) run locally?
DeepSeek R1 0528 (671B MoE) is best suited for server-grade or multi-GPU systems. LocalClaw recommends Q4_K_M as the default quantization, with at least 512 GB RAM.
Search term for LM Studio or compatible runtimes: deepseek-r1-0528
Hugging Face repository: deepseek-ai/DeepSeek-R1-0528-GGUF
reasoningcodequality
Strengths
- Significantly improved reasoning chains vs original R1
- Fewer hallucinations and better self-correction
- MIT license — fully open
- 128K context
- State-of-the-art on math and coding benchmarks
Limitations
- Requires 512GB+ RAM (server-grade only)
- Slower than non-reasoning models
- Still verbose in reasoning chains
Best use cases
- Server-grade reasoning deployment
- Mathematical research
- Complex coding tasks
- Scientific analysis
- Enterprise on-premise AI
Benchmarks
Speed: 1/10
Quality: 10/10
Coding: 10/10
Reasoning: 10/10
Technical details
Developer: DeepSeek AI
License: MIT
Context window: 131,072 tokens
Architecture: Mixture of Experts (MoE) — 671B total, 37B active per token. Updated from original R1.
Released: 2025-05