# Trinity-Large-Thinking

**Overview**

Trinity-Large-Thinking is a reasoning-optimized variant of Arcee AI's Trinity-Large family — a 398B-parameter sparse Mixture-of-Experts (MoE) model with approximately 13B active parameters per token. Built on Trinity-Large-Base and post-trained with extended chain-of-thought reasoning and agentic RL, Trinity-Large-Thinking delivers state-of-the-art performance on agentic benchmarks while maintaining strong general capabilities.

Trinity-Large-Thinking generates explicit reasoning traces wrapped in `<think>...</think>` blocks before producing its final response. This thinking process is critical to the model's performance — **thinking tokens must be kept in context** for multi-turn conversations and agentic loops to function correctly.

<figure><img src="https://2302384909-files.gitbook.io/~/files/v0/b/gitbook-x-prod.appspot.com/o/spaces%2FOUJInHjoAOfzq3bL12Af%2Fuploads%2Fa0cyB3WKEOKpC9Gr64jr%2FIMG_1716.jpg?alt=media&#x26;token=889e8991-91e9-49fa-80d1-82e8e89c95fe" alt=""><figcaption></figcaption></figure>

**Key Features**

* **Agentic-first design**: Purpose-built for tool calling, multi-step planning, and agent workflows
* **State-of-the-art agentic performance**: 94.7% on τ²-Bench, 91.9% on PinchBench, 98.2% on LiveCodeBench
* **Native reasoning traces**: Extended chain-of-thought via `<think>...</think>` blocks
* **Compatible with major agent frameworks**: Works out of the box with [OpenClaw](https://github.com/openclaw) and [Hermes Agent](https://github.com/NousResearch/hermes-agent)

**Thinking-in-Context: Important Usage Note**

Trinity-Large-Thinking produces reasoning traces inside `<think>...</think>` blocks before generating its final response.

This means:

1. **Multi-turn conversations**: When building chat applications, include the full assistant response (thinking + answer) in the conversation history for subsequent turns.
2. **Agentic loops**: When using Trinity-Large-Thinking as the backbone of an agent (OpenClaw, Hermes Agent, or custom), ensure your tool-calling loop preserves `<think>` blocks in the message history between steps.
3. **Context window management**: The 512k extended context window accommodates long reasoning chains across many agentic steps. If you must truncate history, prefer removing older turns entirely rather than stripping thinking tokens from recent turns.

For implementation details, pitfalls (`reasoning` vs `reasoning_content`), and Python/TypeScript examples, refer to the [Reasoning Traces](https://docs.arcee.ai/capabilities/reasoning-traces) page.

### Benchmarks

<table><thead><tr><th>Benchmark</th><th width="114.20703125" align="right">Trinity-Large-Thinking</th><th width="117.40625" align="right">Opus-4.6</th><th width="113.6171875" align="right">GLM-5</th><th width="113.8203125" align="right">MiniMax-M2.7</th><th width="124.5" align="right">Kimi-K2.5</th></tr></thead><tbody><tr><td>IFBench</td><td align="right">52.3</td><td align="right">53.1</td><td align="right">72.3</td><td align="right"><strong>75.7</strong></td><td align="right">70.2</td></tr><tr><td>GPQA-Diamond</td><td align="right">76.3</td><td align="right"><strong>89.2</strong></td><td align="right">81.6</td><td align="right">86.2</td><td align="right">86.9</td></tr><tr><td>Tau2-Airline</td><td align="right"><strong>88.0</strong></td><td align="right">82.0</td><td align="right">80.5</td><td align="right">80.0</td><td align="right">80.0</td></tr><tr><td>Tau2-Telecom</td><td align="right">94.7</td><td align="right">92.1</td><td align="right"><strong>98.2</strong></td><td align="right">84.8</td><td align="right">95.9</td></tr><tr><td>PinchBench</td><td align="right">91.9</td><td align="right"><strong>93.3</strong></td><td align="right">86.4</td><td align="right">89.8</td><td align="right">84.8</td></tr><tr><td>AIME25</td><td align="right">96.3</td><td align="right"><strong>99.8</strong></td><td align="right">93.3</td><td align="right">80.0</td><td align="right">96.3</td></tr><tr><td>BCFLv4</td><td align="right">70.1</td><td align="right"><strong>77.0</strong></td><td align="right">70.8</td><td align="right">70.6</td><td align="right">68.3</td></tr><tr><td>MMLU-Pro</td><td align="right">83.4</td><td align="right"><strong>89.1</strong></td><td align="right">85.8</td><td align="right">80.8</td><td align="right">87.1</td></tr><tr><td>SWE-bench Verified*</td><td align="right">63.2</td><td align="right"><strong>75.6</strong></td><td align="right">72.8</td><td align="right">75.4</td><td align="right">70.8</td></tr></tbody></table>

\*All models evaluated in mini-swe-agent-v2

### Deployment Quickstart

To get started deploying Trinity Large, download the model [here](https://huggingface.co/arcee-ai) and proceed to [quick-deploys](https://docs.arcee.ai/quick-deploys "mention")

### Model Summary

|                                  |                                      |
| -------------------------------- | ------------------------------------ |
| Name                             | Trinity-Large-Thinking               |
| Architecture                     | Sparse MoE (AfmoeForCausalLM)        |
| Parameters                       | 398 Billion Total, 13 Billion Active |
| Experts                          | 256 Experts, 4 Active                |
| Attention Mechanism              | Grouped Query Attention (GQA)        |
| Training Tokens                  | 17 trillion                          |
| License                          | Apache 2.0                           |
| Recommended Inference Parameters | <ul><li>temperature: 0.3</li></ul>   |
