nvidia/Nemotron-Cascade-14B-Thinking
TEXT GENERATIONConcurrency Cost:1Model Size:14BQuant:FP8Ctx Length:32kPublished:Dec 8, 2025License:nvidia-open-model-licenseArchitecture:Transformer0.1K Open Weights Warm
Nemotron-Cascade-14B-Thinking by NVIDIA is a 14 billion parameter general-purpose model, post-trained from Qwen3-14B Base using sequential and domain-wise reinforcement learning. It is specifically designed for 'thinking' mode tasks, achieving best-in-class performance across a wide range of reasoning benchmarks, including competitive programming and mathematical reasoning. The model excels in complex reasoning abilities, surpassing larger models like DeepSeek-R1-0528 (671B) in code benchmarks.
Loading preview...
Popular Sampler Settings
Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.
temperature
–
top_p
–
top_k
–
frequency_penalty
–
presence_penalty
–
repetition_penalty
–
min_p
–