nvidia/Nemotron-Orchestrator-8B
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Nov 25, 2025Architecture:Transformer0.6K Warm

The nvidia/Nemotron-Orchestrator-8B is an 8 billion parameter orchestration model developed by NVIDIA and the University of Hong Kong, designed to manage and coordinate diverse expert models and tools for complex, multi-turn agentic tasks. Trained with Multi-Objective Reinforcement Learning (GRPO), it optimizes for accuracy, latency, and cost efficiency. This model achieves 37.1% on the Humanity's Last Exam (HLE) benchmark, outperforming GPT-5 while being approximately 2.5 times more efficient, making it suitable for robust tool orchestration and versatile reasoning.

Loading preview...

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p