Local LLM model page

DeepSeek R1 Distill (8B)

DeepSeek's reasoning model distilled to 8B. Shows its thought process step-by-step. Mind-blowing for logic.

Parameters
8B
Minimum RAM
8 GB
Model size
5.5 GB
Quantization
Q5_K_M

Can DeepSeek R1 Distill (8B) run locally?

DeepSeek R1 Distill (8B) is best suited for entry-level laptops and desktops. LocalClaw recommends Q5_K_M as the default quantization, with at least 8 GB RAM.

Search term for LM Studio or compatible runtimes: deepseek-r1-distill-llama-8b

Hugging Face repository: lmstudio-community/DeepSeek-R1-Distill-Llama-8B-GGUF

chatreasoningstandard

Strengths

  • Amazing reasoning for 8B
  • Shows chain-of-thought process
  • MIT license
  • Runs on 8GB RAM

Limitations

  • Verbose thinking tokens
  • Distilled — not as strong as full R1
  • Can overthink simple tasks

Best use cases

  • Math reasoning
  • Logic puzzles
  • Step-by-step explanations
  • Education
  • Analysis

Benchmarks

Speed: 7/10

Quality: 7/10

Coding: 7/10

Reasoning: 9/10

Technical details

Developer: DeepSeek AI

License: MIT

Context window: 131,072 tokens

Architecture: Transformer distilled from DeepSeek-R1 (Llama 3.1 8B base)

Released: 2025-01