Local LLM model page
Trinity Large Preview (70B MoE)
Arcee AI's massive MoE open model. ~400B total parameters, 70B active per forward pass. Ranks near the top of global usage leaderboards. Exceptional versatility across reasoning, coding and chat. Free and open-source. Apache 2.0.
Parameters
70B (MoE, ~400B total)
Minimum RAM
48 GB
Model size
45 GB
Quantization
Q4_K_M
Can Trinity Large Preview (70B MoE) run locally?
Trinity Large Preview (70B MoE) is best suited for high-end workstations with 64 GB RAM. LocalClaw recommends Q4_K_M as the default quantization, with at least 48 GB RAM.
Search term for LM Studio or compatible runtimes: trinity-large-preview
Hugging Face repository: arcee-ai/Trinity-Large-Preview-GGUF
chatcodereasoningpowerqualitygeneral
Strengths
- Ranks #2 globally with 114B monthly tokens
- Exceptional versatility
- Apache 2.0
- Free and open-source
Limitations
- Requires 48GB+ RAM
- New model
- MoE complexity
Best use cases
- Enterprise AI
- Complex reasoning
- Coding
- Research
Benchmarks
Speed: 3/10
Quality: 10/10
Coding: 10/10
Reasoning: 10/10
Technical details
Developer: Arcee AI
License: Apache 2.0
Context window: 131,072 tokens
Architecture: Mixture of Experts (MoE), 70B
Released: 2025-12