Local LLM model page

Command R (35B)

Cohere optimized for RAG and long-context. 128K context. Excellent retrieval-augmented generation. 354K downloads.

Parameters
35B
Minimum RAM
24 GB
Model size
20 GB
Quantization
Q4_K_M

Can Command R (35B) run locally?

Command R (35B) is best suited for power-user machines with 32 GB RAM. LocalClaw recommends Q4_K_M as the default quantization, with at least 24 GB RAM.

Search term for LM Studio or compatible runtimes: command-r-35b

Hugging Face repository: lmstudio-community/c4ai-command-r-v01-GGUF

chatgeneralpower

Strengths

  • 128K context
  • Optimized for RAG
  • 10 languages
  • Excellent retrieval-augmented generation

Limitations

  • Non-commercial license
  • Needs 24GB+ RAM
  • Superseded by Command A

Best use cases

  • RAG applications
  • Long document QA
  • Multilingual search
  • Enterprise chatbot

Benchmarks

Speed: 4/10

Quality: 8/10

Coding: 7/10

Reasoning: 8/10

Technical details

Developer: Cohere

License: CC-BY-NC 4.0

Context window: 131,072 tokens

Architecture: Transformer optimized for RAG

Released: 2024-03