Local LLM model page
Qwen 2.5 (72B)
Alibaba's massive 72B. Among the best open models globally. Exceptional multilingual + coding + reasoning.
Parameters
72B
Minimum RAM
64 GB
Model size
42 GB
Quantization
Q4_K_M
Can Qwen 2.5 (72B) run locally?
Qwen 2.5 (72B) is best suited for high-end workstations with 64 GB RAM. LocalClaw recommends Q4_K_M as the default quantization, with at least 64 GB RAM.
Search term for LM Studio or compatible runtimes: qwen2.5-72b-instruct
Hugging Face repository: lmstudio-community/Qwen2.5-72B-Instruct-GGUF
chatcodequalitygeneral
Strengths
- Among the best open 72B models globally
- Exceptional multilingual
- Top coding and reasoning
- 128K context
Limitations
- Requires 64GB+ RAM
- Very slow on consumer hardware
- Restrictive license vs Apache 2.0
Best use cases
- Maximum quality outputs
- Research
- Enterprise AI
- Complex analysis
Benchmarks
Speed: 2/10
Quality: 10/10
Coding: 9/10
Reasoning: 9/10
Technical details
Developer: Alibaba Cloud (Qwen Team)
License: Qwen License
Context window: 131,072 tokens
Architecture: Transformer decoder-only
Released: 2024-09