Tutorial 20 min January 20, 2026

LM Studio Beginner Guide: From Zero to Your First LLM

Installation, GPU configuration, GGUF model download, and first steps with the local chat interface. Step-by-step tutorial to master LM Studio in 2026.

Prerequisites

  • Windows 10/11, macOS 12+, or Linux Ubuntu 20.04+
  • 8GB RAM minimum (16GB recommended)
  • NVIDIA GPU (optional but recommended) or Apple Silicon
  • 10GB disk space for models
1

Installing LM Studio

LM Studio is available for free for Windows, macOS and Linux. Here's how to install it:

  1. Go to lmstudio.ai
  2. Click "Download" and choose your system
  3. Launch the downloaded installer
  4. Follow the installation wizard (accept default settings)
  5. On first launch, LM Studio will automatically download its components

๐Ÿ’ก Tip: On macOS, you may need to authorize the application in System Preferences > Security & Privacy after first launch.

2

GPU Configuration

For optimal performance, configure GPU acceleration. This step is crucial for generation speed.

Windows with NVIDIA

  1. Open LM Studio and go to the Settings tab (โš™๏ธ)
  2. Click Hardware in the left menu
  3. Under "GPU Acceleration", select cuBLAS (NVIDIA)
  4. Enable "GPU offload" and set the slider to 100%
  5. Verify that your GPU appears in the list

macOS (Apple Silicon)

  1. Go to Settings > Hardware
  2. Select Metal as GPU backend
  3. Enable "GPU offload" โ€” all unified RAM will be used

โš ๏ธ Warning: If you don't have a compatible GPU, LM Studio will use the CPU. It will work but be 5-10x slower.

3

Download your first model

LM Studio integrates a model browser. Here's how to find and download an LLM:

  1. Click on the Search tab (๐Ÿ”) on the left
  2. In the search bar, type a model name (ex: "Qwen 3", "Llama 3.3")
  3. Filter by recommended size according to your RAM:
    • 8GB RAM โ†’ 3-7B models
    • 16GB RAM โ†’ 7-13B models
    • 32GB+ RAM โ†’ 30B+ models
  4. Click on a model (ex: "Qwen3-8B-Q5_K_M.gguf")
  5. Click Download โ€” the model will be saved automatically

๐Ÿ’ก Quick alternative: Use LocalClaw to get a personalized recommendation with direct download link!

4

Launch chat and configure the model

Once the model is downloaded, it's time to chat with your local AI:

  1. Click on the Chat tab (๐Ÿ’ฌ) on the left
  2. At the top right, click Select a model to load
  3. Choose the downloaded model from the list
  4. Wait for loading (progress bar)
  5. Once loaded, the chat window becomes active

Important parameters to know

  • Temperature (0-2): Model creativity. 0.7 = balanced, 1.2+ = more creative/risky
  • Context Length: Conversation memory length (default: 4096 tokens)
  • Max Tokens: Maximum length of responses
5

Your first prompts

Test these prompts to evaluate your installation:

Conversation test:

Explain the theory of relativity to me like I'm 10 years old

Code test:

Write a Python function that calculates the Fibonacci sequence up to n

Reasoning test:

A train leaves Paris at 100km/h, another leaves Lyon at 120km/h...

๐Ÿ’ก Performance indicator: Look at the "tokens/s" counter at the bottom of the screen. Above 20 tok/s is fluid, 50+ tok/s is very responsive.

6

Advanced features

Once comfortable, explore these powerful features:

Server Mode (local API)

LM Studio can serve your model as an OpenAI-compatible API:

  1. Go to the Developer tab (๐Ÿ‘จโ€๐Ÿ’ป)
  2. Click Start Server
  3. Your model is now accessible on http://localhost:1234
  4. Use this URL in any OpenAI-compatible application

RAG โ€” Chat with your documents

LM Studio 0.3+ allows you to "converse" with your files:

  1. In the Chat tab, click the ๐Ÿ“Ž icon
  2. Select a PDF or text file
  3. The model will respond based on the document content

System Prompt

Customize model behavior with a "system prompt":

  1. In chat settings, find "System Prompt"
  2. Example: You are a Python programming expert. Respond concisely and technically.
  3. This will influence the entire conversation

Troubleshooting common issues

โŒ "Out of memory" / Crashes

The model is too large for your RAM. Switch to a smaller version (7B instead of 13B) or a stronger quantization (Q4 instead of Q8).

๐ŸŒ Very slow generation

Verify that the GPU is properly configured in Settings > Hardware. If you don't have a compatible GPU, it's normal to be slow on CPU.

๐Ÿ“ฅ Model won't download

Check your internet connection and available disk space. Try another model to isolate the problem.

Conclusion

Congratulations! You now have a completely local, private, and functional generative AI. No subscription, no data sent to the cloud, just you and your machine.

To go further:

And don't forget: LocalClaw is here to help you choose the perfect model every time you want to explore a new LLM!