Local LLM model page
Phi-4 Mini (3.8B)
Microsoft's latest small miracle. Punches way above its weight in reasoning & code.
Parameters
3.8B
Minimum RAM
4 GB
Model size
2.5 GB
Quantization
Q5_K_M
Can Phi-4 Mini (3.8B) run locally?
Phi-4 Mini (3.8B) is best suited for entry-level laptops and desktops. LocalClaw recommends Q5_K_M as the default quantization, with at least 4 GB RAM.
Search term for LM Studio or compatible runtimes: phi-4-mini-instruct
Hugging Face repository: lmstudio-community/Phi-4-mini-instruct-GGUF
chatcodelightspeed
Strengths
- MIT license
- Punches way above weight class
- 128K context at 3.8B
- Great at reasoning & code for its size
Limitations
- English-only
- Limited factual knowledge
- Can hallucinate
Best use cases
- Quick coding help
- Reasoning on-the-go
- Edge deployment
- Education
Benchmarks
Speed: 9/10
Quality: 6/10
Coding: 7/10
Reasoning: 6/10
Technical details
Developer: Microsoft Research
License: MIT
Context window: 131,072 tokens
Architecture: Transformer decoder-only, 128K context
Released: 2025-02