Local LLM model page
Apriel Nemotron 15B Thinker
ServiceNow x NVIDIA mid-size reasoner. Half the memory of 32B reasoners with comparable performance on MBPP, BFCL, GPQA. Strong enterprise fit. MIT licensed.
Parameters
15B
Minimum RAM
16 GB
Model size
9.5 GB
Quantization
Q5_K_M
Can Apriel Nemotron 15B Thinker run locally?
Apriel Nemotron 15B Thinker is best suited for mainstream Macs and PCs with 16 GB RAM. LocalClaw recommends Q5_K_M as the default quantization, with at least 16 GB RAM.
Search term for LM Studio or compatible runtimes: apriel-nemotron-15b-thinker
Hugging Face repository: ServiceNow-AI/Apriel-Nemotron-15b-Thinker
reasoningcodepowergeneral
Strengths
- ServiceNow x NVIDIA mid-size reasoner. Half the memory of 32B reasoners with comparable performance on MBPP, BFCL, GPQA. Strong enterprise fit. MIT licensed.
Limitations
- Performance depends heavily on quantization, RAM bandwidth and runtime support.
Best use cases
- reasoning
- code
- power
- general
Benchmarks
Speed: 6/10
Quality: 8/10
Coding: 8/10
Reasoning: 9/10
Technical details
Developer: nemotron
License: See model repository
Context window: Unknown tokens
Architecture: See model card
Released: 2025-05