Local LLM model page

LLaVA 1.6 (13B)

Larger LLaVA multimodal. Better image understanding than 7B. Describe, analyze and discuss images.

Parameters
13B
Minimum RAM
12 GB
Model size
8 GB
Quantization
Q4_K_M

Can LLaVA 1.6 (13B) run locally?

LLaVA 1.6 (13B) is best suited for mainstream Macs and PCs with 16 GB RAM. LocalClaw recommends Q4_K_M as the default quantization, with at least 12 GB RAM.

Search term for LM Studio or compatible runtimes: llava-v1.6-13b

Hugging Face repository: lmstudio-community/llava-v1.6-vicuna-13b-GGUF

visionpower

Strengths

  • Better image understanding than 7B
  • Apache 2.0
  • Solid multimodal performance

Limitations

  • Needs 12GB RAM
  • Limited context
  • Weak at coding

Best use cases

  • Image analysis
  • Visual Q&A
  • Image captioning
  • Document understanding

Benchmarks

Speed: 5/10

Quality: 7/10

Coding: 3/10

Reasoning: 6/10

Technical details

Developer: LLaVA Team (UW-Madison / Microsoft)

License: Apache 2.0

Context window: 4,096 tokens

Architecture: CLIP vision encoder + Vicuna 13B language model

Released: 2024-02