Local LLM model page

Command A (111B)

Cohere open-weight flagship optimised for agentic workflows and long-context RAG. 256K context, excellent multilingual coverage (23 languages). CC-BY-NC 4.0 — non-commercial.

Parameters
111B
Minimum RAM
96 GB
Model size
68 GB
Quantization
Q4_K_M

Can Command A (111B) run locally?

Command A (111B) is best suited for large-memory workstations. LocalClaw recommends Q4_K_M as the default quantization, with at least 96 GB RAM.

Search term for LM Studio or compatible runtimes: command-a-111b

Hugging Face repository: CohereLabs/command-a-03-2025-GGUF

chatreasoningqualitygeneralpower

Strengths

  • Cohere open-weight flagship optimised for agentic workflows and long-context RAG. 256K context, excellent multilingual coverage (23 languages). CC-BY-NC 4.0 — non-commercial.

Limitations

  • Performance depends heavily on quantization, RAM bandwidth and runtime support.

Best use cases

  • chat
  • reasoning
  • quality
  • general
  • power

Benchmarks

Speed: 2/10

Quality: 10/10

Coding: 9/10

Reasoning: 10/10

Technical details

Developer: cohere

License: See model repository

Context window: Unknown tokens

Architecture: See model card

Released: 2025-03