Local LLM model page
InternVL3 (8B)
Shanghai AI Lab multimodal model. Strong vision understanding for documents, charts, and photos. MIT licensed. Note: primarily PyTorch/safetensors — community GGUF may vary.
Parameters
8B
Minimum RAM
8 GB
Model size
5 GB
Quantization
Q4_K_M
Can InternVL3 (8B) run locally?
InternVL3 (8B) is best suited for entry-level laptops and desktops. LocalClaw recommends Q4_K_M as the default quantization, with at least 8 GB RAM.
Search term for LM Studio or compatible runtimes: internvl3-8b
Hugging Face repository: OpenGVLab/InternVL3-8B-Instruct
visionstandard
Strengths
- Excellent document understanding
- Strong chart and graph reading
- MIT licensed
- Good balance of vision and language
Limitations
- Not the best for general text-only tasks
- 8GB RAM needed
- Slower than text-only models
Best use cases
- Document analysis and OCR
- Chart understanding
- Photo description
- Multimodal Q&A
- Visual reasoning
Benchmarks
Speed: 7/10
Quality: 7/10
Coding: 5/10
Reasoning: 7/10
Technical details
Developer: Shanghai AI Lab (OpenGVLab)
License: MIT
Context window: 32,768 tokens
Architecture: Vision-Language Model with InternViT-6B vision encoder
Released: 2025-04