arcee-ai/Trinity-Large-Thinking
TEXT GENERATIONConcurrency Cost:4Model Size:399BQuant:FP8Ctx Length:32kPublished:Apr 1, 2026License:apache-2.0Architecture:Transformer0.1K Open Weights Warm
Trinity-Large-Thinking by Arcee AI is a 398B-parameter sparse Mixture-of-Experts (MoE) model with 13B active parameters per token, optimized for reasoning and agentic workflows. This variant of the Trinity-Large family is post-trained with extended chain-of-thought reasoning and agentic RL, generating explicit ... blocks for its internal thought process. It delivers state-of-the-art performance on agentic benchmarks and is purpose-built for tool calling, multi-step planning, and agentic applications, requiring thinking tokens to be preserved in context.
Loading preview...
Popular Sampler Settings
Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.
temperature
–
top_p
–
top_k
–
frequency_penalty
–
presence_penalty
–
repetition_penalty
–
min_p
–