Local LLM model page
DeepSeek R1 Distill (70B)
Massive reasoning distill at 70B. Good for pure reasoning but outperformed by GLM 4.5 Air and newer flagship models overall.
Parameters
70B
Minimum RAM
48 GB
Model size
40 GB
Quantization
Q4_K_M
Can DeepSeek R1 Distill (70B) run locally?
DeepSeek R1 Distill (70B) is best suited for high-end workstations with 64 GB RAM. LocalClaw recommends Q4_K_M as the default quantization, with at least 48 GB RAM.
Search term for LM Studio or compatible runtimes: deepseek-r1-distill-llama-70b
Hugging Face repository: lmstudio-community/DeepSeek-R1-Distill-Llama-70B-GGUF
reasoningquality
Strengths
- Top-tier reasoning model
- Rivals o1 on math benchmarks
- MIT license
- Strong at complex logic
Limitations
- Requires 48GB+ RAM
- Very slow
- Verbose outputs
Best use cases
- Advanced mathematics
- Scientific reasoning
- Research
- Complex problem solving
Benchmarks
Speed: 2/10
Quality: 9/10
Coding: 8/10
Reasoning: 10/10
Technical details
Developer: DeepSeek AI
License: MIT
Context window: 131,072 tokens
Architecture: Transformer distilled from DeepSeek-R1
Released: 2025-01