Nemotron-Cascade-8B is a powerful general-purpose language model developed by NVIDIA, post-trained from Qwen3-8B-Base. It features a unique sequential and domain-wise reinforcement learning (Cascade RL) pipeline, enabling it to operate in both 'thinking' and 'instruct' modes. This 8-billion parameter model achieves best-in-class performance across diverse benchmarks, including knowledge reasoning, alignment, mathematics, and competitive programming, notably matching the LiveCodeBench scores of much larger models like DeepSeek-R1-0528 (671B). Its primary use case is complex reasoning and instruction following, with robust support for long contexts up to 64K tokens via YaRN scaling.
No reviews yet. Be the first to review!