Local LLM model page
Cogito (32B)
Hybrid reasoning at 32B. Outperforms larger models on reasoning tasks. Strong general purpose.
Parameters
32B
Minimum RAM
24 GB
Model size
19 GB
Quantization
Q4_K_M
Can Cogito (32B) run locally?
Cogito (32B) is best suited for power-user machines with 32 GB RAM. LocalClaw recommends Q4_K_M as the default quantization, with at least 24 GB RAM.
Search term for LM Studio or compatible runtimes: cogito-32b
Hugging Face repository: deepcogito/cogito-v1-preview-qwen-32B-GGUF
chatreasoningpowerquality
Strengths
- Outperforms larger models on reasoning
- Apache 2.0
- Excellent general purpose
- Strong coding
Limitations
- Needs 24GB+ RAM
- English-only
- New model
Best use cases
- Advanced reasoning
- Professional coding
- Complex analysis
- Research
Benchmarks
Speed: 4/10
Quality: 9/10
Coding: 8/10
Reasoning: 10/10
Technical details
Developer: Deep Cogito
License: Apache 2.0
Context window: 131,072 tokens
Architecture: Hybrid reasoning Transformer (Qwen 32B base)
Released: 2025-04