Local LLM model page

Mistral Small (24B)

Mistral's refined 24B model. Excellent for nuanced conversations and professional writing.

Parameters
24B
Minimum RAM
24 GB
Model size
15 GB
Quantization
Q4_K_M

Can Mistral Small (24B) run locally?

Mistral Small (24B) is best suited for power-user machines with 32 GB RAM. LocalClaw recommends Q4_K_M as the default quantization, with at least 24 GB RAM.

Search term for LM Studio or compatible runtimes: mistral-small-24b-instruct

Hugging Face repository: lmstudio-community/Mistral-Small-24B-Instruct-2501-GGUF

chatgeneralpowerquality

Strengths

  • Excellent quality-to-size ratio
  • Strong European multilingual
  • Apache 2.0
  • Fast inference for 24B

Limitations

  • 32K context limit
  • Needs 16GB+ RAM

Best use cases

  • Enterprise chatbot
  • Multilingual support
  • Professional writing
  • Code generation

Benchmarks

Speed: 5/10

Quality: 8/10

Coding: 7/10

Reasoning: 8/10

Technical details

Developer: Mistral AI

License: Apache 2.0

Context window: 32,768 tokens

Architecture: Transformer decoder-only

Released: 2025-01