Local LLM model page

Cogito (32B)

Hybrid reasoning at 32B. Outperforms larger models on reasoning tasks. Strong general purpose.

Parameters
32B
Minimum RAM
24 GB
Model size
19 GB
Quantization
Q4_K_M

Can Cogito (32B) run locally?

Cogito (32B) is best suited for power-user machines with 32 GB RAM. LocalClaw recommends Q4_K_M as the default quantization, with at least 24 GB RAM.

Search term for LM Studio or compatible runtimes: cogito-32b

Hugging Face repository: deepcogito/cogito-v1-preview-qwen-32B-GGUF

chatreasoningpowerquality

Strengths

  • Outperforms larger models on reasoning
  • Apache 2.0
  • Excellent general purpose
  • Strong coding

Limitations

  • Needs 24GB+ RAM
  • English-only
  • New model

Best use cases

  • Advanced reasoning
  • Professional coding
  • Complex analysis
  • Research

Benchmarks

Speed: 4/10

Quality: 9/10

Coding: 8/10

Reasoning: 10/10

Technical details

Developer: Deep Cogito

License: Apache 2.0

Context window: 131,072 tokens

Architecture: Hybrid reasoning Transformer (Qwen 32B base)

Released: 2025-04