arcee-ai/Trinity-Large-Thinking
TEXT GENERATIONConcurrency Cost:4Model Size:399BQuant:FP8Ctx Length:32kPublished:Apr 1, 2026License:apache-2.0Architecture:Transformer0.1K Open Weights Warm

Trinity-Large-Thinking by Arcee AI is a 398B-parameter sparse Mixture-of-Experts (MoE) model with 13B active parameters per token, optimized for reasoning and agentic workflows. This variant of the Trinity-Large family is post-trained with extended chain-of-thought reasoning and agentic RL, generating explicit ... blocks for its internal thought process. It delivers state-of-the-art performance on agentic benchmarks and is purpose-built for tool calling, multi-step planning, and agentic applications, requiring thinking tokens to be preserved in context.

Loading preview...

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p