Page cover

Trinity-Mini (26B)

Overview

Trinity Mini is a 26B-parameter (3B active) sparse mixture-of-experts language model, engineered for efficient inference over long contexts with robust function calling and multi-step agent workflows.

Key Features

  • Efficient attention mechanism: reduces memory and compute requirements while preserving long-context coherence.

  • 128K-token context window: supports multi-turn interactions and extended document processing.

  • Strong context utilization: fully leverages long inputs for coherent multi-turn reasoning and reliable function/tool calls.

  • High inference efficiency: generates tokens rapidly while minimizing compute, delivering an outstanding price-to-performance ratio.

Deployment Quickstart

To get started deploying Trinity-26B, download the model here and proceed to Quick Deploys

Model Summary

Name

Trinity-Mini-26B

Architecture

Mixture-of-Experts

Parameters

26 Billion Total, 3.5 Billion Active

Experts

128 Experts, 8 Active

Attention Mechanism

Grouped Query Attention (GQA)

Training Tokens

10 trillion

License

Apache 2.0

Recommended Inference Parameters

  • temperature: 0.15

  • top_p: 0.75

  • top_k: 50

  • min_p: 0.06

Last updated