open-r1/OlympicCoder-32B

Warm
Public
32.8B
FP8
131072
License: apache-2.0
Hugging Face
Overview

OlympicCoder-32B: A Specialized Code Generation Model

OlympicCoder-32B, developed by open-r1, is a 32 billion parameter model fine-tuned from Qwen/Qwen2.5-Coder-32B-Instruct. Its primary focus is on competitive programming, demonstrating robust performance on challenging benchmarks.

Key Capabilities & Features

  • Competitive Coding Excellence: Achieves strong results on competitive coding benchmarks, including the 2024 International Olympiad in Informatics (IOI) and LiveCodeBench.
  • Specialized Training: Fine-tuned on a decontaminated version of the Codeforces dataset, with additional post-training on C++ solutions generated by DeepSeek-R1.
  • Chain-of-Thought (CoT) Optimization: Utilizes a modified chat template that pre-fills the assistant's turn with a <think> token to consistently encourage long chain-of-thought reasoning, enhancing problem-solving capabilities.

Evaluation Highlights

The model's performance is rigorously evaluated on:

  • IOI'24: Assesses its ability to solve 6 highly challenging problems from the International Olympiad in Informatics.
  • LiveCodeBench: Evaluated on the v4_v5 subset of livecodebench/code_generation_lite, comprising 268 Python programming problems from platforms like CodeForces and LeetCode. It's important to note that while the model was post-trained on C++ solutions, its LiveCodeBench evaluation is on Python, making it partially out-of-domain.

Ideal Use Cases

  • Competitive Programming: Generating solutions for complex algorithmic problems.
  • Code Generation: Assisting developers with intricate coding challenges, particularly those requiring detailed reasoning.
  • Research: Exploring advanced techniques in code generation and problem-solving with large language models.