Guided Mode
Simple questionnaire. OS, RAM level, use case. We handle the complexity.
Ex: MacBook Air 8 GB → Qwen 3 8B
Stop sending your data to the cloud. Find the perfect open-source model for your hardware.
Simple questionnaire. OS, RAM level, use case. We handle the complexity.
Ex: MacBook Air 8 GB → Qwen 3 8B
Direct input. Select RAM, GPU and priorities for instant logic execution.
Ex: 32 GB RAM + RTX 4090 → DeepSeek R1 32B
Paste diagnostics. Auto-detection of OS/RAM/GPU for precision targeting.
Ex: Paste neofetch → auto-detect & match
Text-to-Speech models that run 100% offline on your hardware. Perfect for voice assistants, audiobooks, accessibility, and creative projects.
Install, update and manage all your local models from a single unified dashboard.
View PricingLM Studio is a free desktop application that lets you run Large Language Models (LLMs) locally on your computer. No internet needed, no data sent anywhere. It provides a chat interface similar to ChatGPT but everything runs on YOUR hardware.
Quantization is a compression technique that reduces model size while preserving most of the quality. Think of it like JPEG compression for images. Q4 = more compressed (smaller, slightly lower quality), Q8 = less compressed (larger, nearly original quality). Q5_K_M is the sweet spot for most users.
Rule of thumb: the model file size plus 2-3 GB for the system. A 5 GB model needs at least 8 GB RAM. On macOS with Apple Silicon, the unified memory makes things more efficient. On Windows/Linux with a GPU, VRAM helps offload the model.
Apple Silicon (M1-M4) uses unified memory, meaning your entire RAM is available for the model. This is incredibly efficient. NVIDIA GPUs are faster for inference but limited by VRAM (typically 8-24 GB). Both are great choices.
Yes! LocalClaw runs entirely in your browser — zero data is collected or sent anywhere. When using LM Studio with recommended models, everything runs locally on your machine. No cloud, no tracking, no API calls.
For 8 GB RAM: Qwen 3 8B and Llama 3.3 8B. For 16 GB: Qwen 3 14B. For 32 GB+: Qwen 3 32B and DeepSeek R1 32B. For coding: Qwen 2.5 Coder 7B. For vision: Gemma 3 12B. For reasoning: DeepSeek R1 series.
OpenClaw is the open-source, self-hosted AI assistant at the heart of the LocalClaw ecosystem. It connects to your local models running in LM Studio or Ollama and provides a unified chat interface on desktop, web, and CLI. It's 100% private — no telemetry, no cloud, no API keys required.
LocalClaw Installer is the native macOS app that manages your local AI setup — install models, handle updates, switch versions, and launch everything with one click. No terminal needed. It's a one-time purchase at $49, no subscription, no recurring fees. Your license is valid forever. See pricing →
Answer a few questions about your hardware and get personalized AI model recommendations — instantly, privately, for free.
Find My Model