Local LLM model page

Trinity Large Preview (70B MoE)

Arcee AI's massive MoE open model. ~400B total parameters, 70B active per forward pass. Ranks near the top of global usage leaderboards. Exceptional versatility across reasoning, coding and chat. Free and open-source. Apache 2.0.

Parameters
70B (MoE, ~400B total)
Minimum RAM
48 GB
Model size
45 GB
Quantization
Q4_K_M

Can Trinity Large Preview (70B MoE) run locally?

Trinity Large Preview (70B MoE) is best suited for high-end workstations with 64 GB RAM. LocalClaw recommends Q4_K_M as the default quantization, with at least 48 GB RAM.

Search term for LM Studio or compatible runtimes: trinity-large-preview

Hugging Face repository: arcee-ai/Trinity-Large-Preview-GGUF

chatcodereasoningpowerqualitygeneral

Strengths

  • Ranks #2 globally with 114B monthly tokens
  • Exceptional versatility
  • Apache 2.0
  • Free and open-source

Limitations

  • Requires 48GB+ RAM
  • New model
  • MoE complexity

Best use cases

  • Enterprise AI
  • Complex reasoning
  • Coding
  • Research

Benchmarks

Speed: 3/10

Quality: 10/10

Coding: 10/10

Reasoning: 10/10

Technical details

Developer: Arcee AI

License: Apache 2.0

Context window: 131,072 tokens

Architecture: Mixture of Experts (MoE), 70B

Released: 2025-12