nvidia/OpenCodeReasoning-Nemotron-32B-IOI

Hugging Face
TEXT GENERATIONConcurrency Cost:2Model Size:32.8BQuant:FP8Ctx Length:32kPublished:May 7, 2025License:apache-2.0Architecture:Transformer0.0K Open Weights Warm

OpenCodeReasoning-Nemotron-32B-IOI is a 32 billion parameter large language model developed by NVIDIA, derived from Qwen2.5-32B-Instruct. It is specifically post-trained for reasoning in code generation tasks, supporting a context length of 32K tokens. This model excels at competitive programming problems, particularly in C++ and Python, and is suitable for commercial and non-commercial use.

Loading preview...

Overview

NVIDIA's OpenCodeReasoning-Nemotron-32B-IOI is a 32 billion parameter large language model, building upon the Qwen2.5-32B-Instruct architecture. It is specifically fine-tuned for advanced reasoning in code generation, making it highly effective for competitive programming challenges. The model supports an extensive context length of 32,768 tokens, allowing for complex problem understanding and solution generation.

Key Capabilities

  • Code Reasoning: Post-trained to enhance reasoning abilities for generating correct and efficient code.
  • Competitive Programming: Demonstrates strong performance on benchmarks like LiveCodeBench and CodeContests, particularly for C++ and Python.
  • Large Context Window: Utilizes a 32K token context length for processing detailed problem descriptions and generating comprehensive solutions.
  • Commercial Use: Licensed under Apache 2.0, suitable for both commercial and non-commercial applications.

Performance Highlights

Evaluations show OpenCodeReasoning-Nemotron-32B-IOI achieving a LiveCodeBench (pass@1) score of 61.5 and a CodeContests (pass@1) score of 25.5. It also scored 175.5 on the IOI (Total Score) benchmark, matching or surpassing other models in its class. The model was trained on the OpenCodeReasoning dataset, which includes 736K Python and 356K C++ samples derived from competitive programming questions and DeepSeek-R1 generated responses.

Good For

  • Developers and researchers focused on code generation with strong reasoning requirements.
  • Applications involving competitive programming problem-solving.
  • Tasks requiring the generation of C++ and Python code from complex prompts.

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p