🧭 Guided Mode
Answer simple questions about your machine — OS, RAM level, and use case. We handle the rest.
LocalClaw recommends the best open-source LLM for LM Studio based on your RAM, GPU and use case. 100% private — everything runs in your browser. No data collected. Ever.
Answer simple questions about your machine — OS, RAM level, and use case. We handle the rest.
Know your specs? Select RAM, GPU and priorities directly for instant AI model recommendations.
Paste your system diagnostics output. We auto-detect OS, RAM and GPU to find your perfect model.
LM Studio is a free desktop application that lets you run Large Language Models (LLMs) locally on your computer. No internet needed, no data sent anywhere. It provides a chat interface similar to ChatGPT but everything runs on YOUR hardware.
Quantization is a compression technique that reduces model size while preserving most of the quality. Think of it like JPEG compression for images. Q4 = more compressed (smaller, slightly lower quality), Q8 = less compressed (larger, nearly original quality). Q5_K_M is the sweet spot for most users.
Rule of thumb: the model file size + ~2-3 GB for the system. A 5 GB model needs at least 8 GB RAM. On macOS with Apple Silicon, the unified memory makes things more efficient. On Windows/Linux with a GPU, VRAM helps offload the model.
Apple Silicon (M1-M4) uses unified memory, meaning your entire RAM is available for the model. This is incredibly efficient. NVIDIA GPUs are faster for inference but limited by VRAM (typically 8-24 GB). Both are great choices.
Yes! LocalClaw runs entirely in your browser — zero data is collected or sent anywhere. When using LM Studio with recommended models, everything runs locally on your machine. No cloud, no tracking, no API calls.
For 8 GB RAM: Qwen 3 8B and Llama 3.3 8B offer the best quality. For 16 GB: Qwen 3 14B is king. For 32 GB+: Qwen 3 32B and DeepSeek R1 32B rival GPT-4. For coding: Qwen 2.5 Coder 7B. For vision: Gemma 3 12B. For reasoning: DeepSeek R1 series.