Local LLM model page
QwQ (32B)
Early Qwen reasoning model. Superseded by GLM-4 32B and Qwen 3 32B for most tasks. Still decent for pure math.
Parameters
32B
Minimum RAM
24 GB
Model size
19 GB
Quantization
Q4_K_M
Can QwQ (32B) run locally?
QwQ (32B) is best suited for power-user machines with 32 GB RAM. LocalClaw recommends Q4_K_M as the default quantization, with at least 24 GB RAM.
Search term for LM Studio or compatible runtimes: qwq-32b-preview
Hugging Face repository: lmstudio-community/QwQ-32B-Preview-GGUF
reasoningpower
Strengths
- o1-class reasoning
- Shows chain-of-thought process
- Apache 2.0
- Strong math/logic
Limitations
- Verbose outputs (thinking tokens)
- Slower due to reasoning overhead
- Needs 24GB+ RAM
Best use cases
- Complex math problems
- Logical reasoning
- Scientific analysis
- Strategic planning
Benchmarks
Speed: 4/10
Quality: 7/10
Coding: 6/10
Reasoning: 8/10
Technical details
Developer: Alibaba Cloud (Qwen Team)
License: Apache 2.0
Context window: 131,072 tokens
Architecture: Reasoning-focused Transformer
Released: 2024-11