Local LLM model page

Apriel Nemotron 15B Thinker

ServiceNow x NVIDIA mid-size reasoner. Half the memory of 32B reasoners with comparable performance on MBPP, BFCL, GPQA. Strong enterprise fit. MIT licensed.

Parameters
15B
Minimum RAM
16 GB
Model size
9.5 GB
Quantization
Q5_K_M

Can Apriel Nemotron 15B Thinker run locally?

Apriel Nemotron 15B Thinker is best suited for mainstream Macs and PCs with 16 GB RAM. LocalClaw recommends Q5_K_M as the default quantization, with at least 16 GB RAM.

Search term for LM Studio or compatible runtimes: apriel-nemotron-15b-thinker

Hugging Face repository: ServiceNow-AI/Apriel-Nemotron-15b-Thinker

reasoningcodepowergeneral

Strengths

  • ServiceNow x NVIDIA mid-size reasoner. Half the memory of 32B reasoners with comparable performance on MBPP, BFCL, GPQA. Strong enterprise fit. MIT licensed.

Limitations

  • Performance depends heavily on quantization, RAM bandwidth and runtime support.

Best use cases

  • reasoning
  • code
  • power
  • general

Benchmarks

Speed: 6/10

Quality: 8/10

Coding: 8/10

Reasoning: 9/10

Technical details

Developer: nemotron

License: See model repository

Context window: Unknown tokens

Architecture: See model card

Released: 2025-05