nvidia/Nemotron-Cascade-14B-Thinking
TEXT GENERATIONConcurrency Cost:1Model Size:14BQuant:FP8Ctx Length:32kPublished:Dec 8, 2025License:nvidia-open-model-licenseArchitecture:Transformer0.1K Open Weights Warm

Nemotron-Cascade-14B-Thinking by NVIDIA is a 14 billion parameter general-purpose model, post-trained from Qwen3-14B Base using sequential and domain-wise reinforcement learning. It is specifically designed for 'thinking' mode tasks, achieving best-in-class performance across a wide range of reasoning benchmarks, including competitive programming and mathematical reasoning. The model excels in complex reasoning abilities, surpassing larger models like DeepSeek-R1-0528 (671B) in code benchmarks.

Loading preview...

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p