Local LLM model page

Qwen 2.5 (72B)

Alibaba's massive 72B. Among the best open models globally. Exceptional multilingual + coding + reasoning.

Parameters
72B
Minimum RAM
64 GB
Model size
42 GB
Quantization
Q4_K_M

Can Qwen 2.5 (72B) run locally?

Qwen 2.5 (72B) is best suited for high-end workstations with 64 GB RAM. LocalClaw recommends Q4_K_M as the default quantization, with at least 64 GB RAM.

Search term for LM Studio or compatible runtimes: qwen2.5-72b-instruct

Hugging Face repository: lmstudio-community/Qwen2.5-72B-Instruct-GGUF

chatcodequalitygeneral

Strengths

  • Among the best open 72B models globally
  • Exceptional multilingual
  • Top coding and reasoning
  • 128K context

Limitations

  • Requires 64GB+ RAM
  • Very slow on consumer hardware
  • Restrictive license vs Apache 2.0

Best use cases

  • Maximum quality outputs
  • Research
  • Enterprise AI
  • Complex analysis

Benchmarks

Speed: 2/10

Quality: 10/10

Coding: 9/10

Reasoning: 9/10

Technical details

Developer: Alibaba Cloud (Qwen Team)

License: Qwen License

Context window: 131,072 tokens

Architecture: Transformer decoder-only

Released: 2024-09