Local LLM model page
Command R (35B)
Cohere optimized for RAG and long-context. 128K context. Excellent retrieval-augmented generation. 354K downloads.
Parameters
35B
Minimum RAM
24 GB
Model size
20 GB
Quantization
Q4_K_M
Can Command R (35B) run locally?
Command R (35B) is best suited for power-user machines with 32 GB RAM. LocalClaw recommends Q4_K_M as the default quantization, with at least 24 GB RAM.
Search term for LM Studio or compatible runtimes: command-r-35b
Hugging Face repository: lmstudio-community/c4ai-command-r-v01-GGUF
chatgeneralpower
Strengths
- 128K context
- Optimized for RAG
- 10 languages
- Excellent retrieval-augmented generation
Limitations
- Non-commercial license
- Needs 24GB+ RAM
- Superseded by Command A
Best use cases
- RAG applications
- Long document QA
- Multilingual search
- Enterprise chatbot
Benchmarks
Speed: 4/10
Quality: 8/10
Coding: 7/10
Reasoning: 8/10
Technical details
Developer: Cohere
License: CC-BY-NC 4.0
Context window: 131,072 tokens
Architecture: Transformer optimized for RAG
Released: 2024-03