Local LLM model page

WizardLM 2 (8x22B)

Microsoft AI's ultra-popular fine-tune of Mixtral 8x22B. Apache 2.0 license. Exceptional instruction following and conversational quality.

Parameters
8x22B (141B total)
Minimum RAM
96 GB
Model size
88 GB
Quantization
Q4_K_M

Can WizardLM 2 (8x22B) run locally?

WizardLM 2 (8x22B) is best suited for large-memory workstations. LocalClaw recommends Q4_K_M as the default quantization, with at least 96 GB RAM.

Search term for LM Studio or compatible runtimes: wizardlm-2-8x22b

Hugging Face repository: lmstudio-community/WizardLM-2-8x22B-GGUF

chatcodepowerqualitygeneral

Strengths

  • Exceptional instruction following
  • Apache 2.0
  • One of the best fine-tunes ever
  • Strong conversational quality

Limitations

  • Requires 96GB+ RAM
  • Very large model
  • Slow inference

Best use cases

  • Maximum quality chat
  • Complex instructions
  • Professional content creation
  • Research

Benchmarks

Speed: 3/10

Quality: 10/10

Coding: 9/10

Reasoning: 9/10

Technical details

Developer: Microsoft AI

License: Apache 2.0

Context window: 65,536 tokens

Architecture: Mixtral 8x22B fine-tuned with WizardLM pipeline

Released: 2024-04