Local LLM model page

Phi-4 (14B)

Microsoft's full Phi-4. Compact powerhouse with exceptional reasoning and coding for its size. MIT licensed.

Parameters
14B
Minimum RAM
16 GB
Model size
9 GB
Quantization
Q5_K_M

Can Phi-4 (14B) run locally?

Phi-4 (14B) is best suited for mainstream Macs and PCs with 16 GB RAM. LocalClaw recommends Q5_K_M as the default quantization, with at least 16 GB RAM.

Search term for LM Studio or compatible runtimes: phi-4-instruct

Hugging Face repository: lmstudio-community/Phi-4-Instruct-GGUF

chatcodepowerreasoning

Strengths

  • Exceptional reasoning for 14B
  • MIT license
  • Strong coding
  • State-of-the-art at its size

Limitations

  • 16K context only
  • English-only
  • Smaller than Phi-4 Mini context

Best use cases

  • Coding assistance
  • Math and reasoning
  • Research
  • Technical writing

Benchmarks

Speed: 7/10

Quality: 8/10

Coding: 9/10

Reasoning: 8/10

Technical details

Developer: Microsoft Research

License: MIT

Context window: 16,384 tokens

Architecture: Transformer decoder-only

Released: 2025-01